Merge remote-tracking branch 'regulator/for-5.8' into regulator-linus
authorMark Brown <broonie@kernel.org>
Mon, 1 Jun 2020 12:01:44 +0000 (13:01 +0100)
committerMark Brown <broonie@kernel.org>
Mon, 1 Jun 2020 12:01:44 +0000 (13:01 +0100)
271 files changed:
.mailmap
Documentation/devicetree/bindings/net/dsa/b53.txt
MAINTAINERS
Makefile
arch/arc/configs/hsdk_defconfig
arch/arc/include/asm/dsp-impl.h
arch/arc/include/asm/entry-arcv2.h
arch/arc/kernel/Makefile
arch/arc/kernel/ptrace.c
arch/arc/kernel/setup.c
arch/arc/kernel/troubleshoot.c
arch/arc/kernel/unwind.c
arch/arc/plat-eznps/Kconfig
arch/arm64/include/asm/uaccess.h
arch/arm64/kernel/ptrace.c
arch/powerpc/Kconfig
arch/powerpc/include/asm/book3s/32/hash.h
arch/powerpc/kernel/head_32.S
arch/powerpc/mm/book3s32/hash_low.S
arch/riscv/kernel/process.c
arch/riscv/mm/init.c
arch/s390/include/asm/pci_io.h
arch/s390/kernel/machine_kexec_file.c
arch/s390/kernel/machine_kexec_reloc.c
arch/s390/mm/hugetlbpage.c
arch/s390/pci/pci_mmio.c
arch/sh/include/uapi/asm/sockios.h
arch/sparc/mm/srmmu.c
arch/um/drivers/vector_user.h
arch/um/include/asm/xor.h
arch/um/kernel/skas/syscall.c
arch/x86/boot/tools/build.c
arch/x86/hyperv/hv_init.c
arch/x86/include/asm/bitops.h
arch/x86/kernel/unwind_orc.c
arch/x86/mm/mmio-mod.c
drivers/acpi/ec.c
drivers/acpi/sleep.c
drivers/base/core.c
drivers/block/null_blk_main.c
drivers/block/null_blk_zoned.c
drivers/bus/mhi/core/init.c
drivers/dax/kmem.c
drivers/dma/dmatest.c
drivers/dma/idxd/device.c
drivers/dma/idxd/irq.c
drivers/dma/owl-dma.c
drivers/dma/tegra210-adma.c
drivers/dma/ti/k3-udma.c
drivers/dma/xilinx/zynqmp_dma.c
drivers/firmware/efi/cper.c
drivers/firmware/efi/earlycon.c
drivers/firmware/efi/efi.c
drivers/firmware/efi/libstub/arm-stub.c
drivers/firmware/efi/libstub/efistub.h
drivers/firmware/efi/libstub/mem.c
drivers/firmware/efi/libstub/tpm.c
drivers/firmware/efi/libstub/x86-stub.c
drivers/firmware/efi/tpm.c
drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer.h
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_init.c
drivers/gpu/drm/amd/display/dc/dcn20/dcn20_init.c
drivers/gpu/drm/amd/display/dc/dcn21/dcn21_init.c
drivers/gpu/drm/amd/display/dc/dml/Makefile
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20.h
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20v2.h
drivers/gpu/drm/amd/display/dc/dml/dcn21/display_rq_dlg_calc_21.h
drivers/gpu/drm/amd/display/dc/dml/display_mode_lib.h
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.h
drivers/gpu/drm/amd/display/dc/dml/display_rq_dlg_helpers.h
drivers/gpu/drm/amd/display/dc/dml/dml1_display_rq_dlg_calc.h
drivers/gpu/drm/amd/display/dc/dml/dml_common_defs.c [deleted file]
drivers/gpu/drm/amd/display/dc/dml/dml_common_defs.h [deleted file]
drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h
drivers/gpu/drm/amd/display/dc/inc/hw_sequencer.h
drivers/gpu/drm/drm_edid.c
drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
drivers/hwtracing/coresight/coresight-cti-platform.c
drivers/i2c/algos/i2c-algo-pca.c
drivers/i2c/busses/i2c-altera.c
drivers/i2c/busses/i2c-at91-master.c
drivers/i2c/i2c-core-base.c
drivers/i2c/i2c-core-of.c
drivers/i2c/muxes/i2c-demux-pinctrl.c
drivers/iio/accel/sca3000.c
drivers/iio/adc/stm32-adc.c
drivers/iio/adc/stm32-dfsdm-adc.c
drivers/iio/adc/ti-ads8344.c
drivers/iio/chemical/atlas-sensor.c
drivers/iio/dac/vf610_dac.c
drivers/iio/imu/st_lsm6dsx/st_lsm6dsx_shub.c
drivers/iommu/amd_iommu.c
drivers/iommu/amd_iommu_init.c
drivers/iommu/iommu.c
drivers/ipack/carriers/tpci200.c
drivers/misc/cardreader/rtsx_pcr.c
drivers/misc/mei/client.c
drivers/mtd/mtdcore.c
drivers/mtd/nand/raw/brcmnand/brcmnand.c
drivers/mtd/nand/spi/core.c
drivers/mtd/ubi/debug.c
drivers/net/can/ifi_canfd/ifi_canfd.c
drivers/net/can/sun4i_can.c
drivers/net/dsa/b53/b53_srab.c
drivers/net/dsa/mt7530.c
drivers/net/dsa/mt7530.h
drivers/net/dsa/ocelot/felix.c
drivers/net/dsa/ocelot/felix.h
drivers/net/dsa/ocelot/felix_vsc9959.c
drivers/net/ethernet/apple/bmac.c
drivers/net/ethernet/freescale/ucc_geth.c
drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c
drivers/net/ethernet/marvell/pxa168_eth.c
drivers/net/ethernet/mellanox/mlx4/fw.c
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
drivers/net/ethernet/mellanox/mlx5/core/en.h
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.h
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls.c
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
drivers/net/ethernet/mellanox/mlx5/core/en_rep.c
drivers/net/ethernet/mellanox/mlx5/core/en_rep.h
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
drivers/net/ethernet/mellanox/mlx5/core/eq.c
drivers/net/ethernet/mellanox/mlx5/core/events.c
drivers/net/ethernet/mellanox/mlx5/core/fs_core.c
drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib.c
drivers/net/ethernet/mellanox/mlx5/core/main.c
drivers/net/ethernet/mellanox/mlxsw/spectrum.c
drivers/net/ethernet/mellanox/mlxsw/switchx2.c
drivers/net/ethernet/mscc/ocelot.c
drivers/net/ethernet/realtek/r8169_main.c
drivers/net/ethernet/sgi/ioc3-eth.c
drivers/net/ethernet/smsc/smsc911x.c
drivers/net/ethernet/stmicro/stmmac/dwmac-ipq806x.c
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
drivers/net/ethernet/sun/cassini.c
drivers/net/ethernet/ti/am65-cpsw-nuss.c
drivers/net/ethernet/ti/cpsw.c
drivers/net/ethernet/ti/cpsw_ale.c
drivers/net/ethernet/ti/cpsw_priv.c
drivers/net/ethernet/ti/netcp_ethss.c
drivers/net/ipa/gsi.c
drivers/net/netdevsim/dev.c
drivers/net/phy/mscc/mscc.h
drivers/net/phy/mscc/mscc_mac.h
drivers/net/phy/mscc/mscc_macsec.c
drivers/net/phy/mscc/mscc_macsec.h
drivers/net/phy/mscc/mscc_main.c
drivers/net/phy/phy_device.c
drivers/net/usb/cdc_ether.c
drivers/net/usb/r8152.c
drivers/net/wireguard/messages.h
drivers/net/wireguard/noise.c
drivers/net/wireguard/noise.h
drivers/net/wireguard/queueing.h
drivers/net/wireguard/receive.c
drivers/net/wireguard/selftest/counter.c
drivers/net/wireguard/send.c
drivers/net/wireless/intel/iwlwifi/pcie/drv.c
drivers/rapidio/devices/rio_mport_cdev.c
drivers/regulator/qcom-rpmh-regulator.c
drivers/scsi/qla2xxx/qla_attr.c
drivers/scsi/scsi_pm.c
drivers/staging/greybus/uart.c
drivers/staging/iio/resolver/ad2s1210.c
drivers/staging/kpc2000/kpc2000/core.c
drivers/staging/wfx/scan.c
drivers/target/target_core_transport.c
drivers/tty/serial/sifive.c
drivers/vdpa/vdpa_sim/vdpa_sim.c
drivers/vhost/vhost.c
fs/afs/fs_probe.c
fs/afs/fsclient.c
fs/afs/vl_probe.c
fs/afs/yfsclient.c
fs/exfat/file.c
fs/exfat/namei.c
fs/exfat/super.c
fs/ext4/ext4.h
fs/ext4/extents.c
fs/ext4/ioctl.c
fs/file.c
fs/io_uring.c
fs/overlayfs/export.c
fs/overlayfs/inode.c
fs/splice.c
fs/ubifs/auth.c
fs/ubifs/file.c
fs/ubifs/replay.c
include/linux/cper.h
include/linux/efi.h
include/linux/i2c-mux.h
include/linux/i2c.h
include/linux/mlx5/driver.h
include/net/act_api.h
include/net/af_rxrpc.h
include/net/ip_fib.h
include/trace/events/rxrpc.h
kernel/bpf/syscall.c
kernel/sched/debug.c
kernel/sched/fair.c
lib/test_printf.c
lib/vsprintf.c
mm/kasan/Makefile
mm/kasan/generic.c
mm/kasan/tags.c
mm/z3fold.c
net/ax25/af_ax25.c
net/core/dev.c
net/core/flow_dissector.c
net/dsa/tag_mtk.c
net/ethtool/netlink.c
net/ethtool/strset.c
net/ipv4/fib_frontend.c
net/ipv4/inet_connection_sock.c
net/ipv4/ipip.c
net/ipv4/ipmr.c
net/ipv4/nexthop.c
net/ipv4/route.c
net/ipv6/ip6_fib.c
net/ipv6/ip6mr.c
net/mptcp/crypto.c
net/mptcp/options.c
net/mptcp/protocol.h
net/mptcp/subflow.c
net/qrtr/qrtr.c
net/rxrpc/Makefile
net/rxrpc/ar-internal.h
net/rxrpc/call_accept.c
net/rxrpc/call_event.c
net/rxrpc/input.c
net/rxrpc/misc.c
net/rxrpc/output.c
net/rxrpc/peer_event.c
net/rxrpc/peer_object.c
net/rxrpc/proc.c
net/rxrpc/rtt.c [new file with mode: 0644]
net/rxrpc/rxkad.c
net/rxrpc/sendmsg.c
net/rxrpc/sysctl.c
net/sctp/sm_sideeffect.c
net/sctp/sm_statefuns.c
net/tipc/udp_media.c
net/tls/tls_sw.c
security/apparmor/apparmorfs.c
security/apparmor/audit.c
security/apparmor/domain.c
security/integrity/evm/evm_crypto.c
security/integrity/evm/evm_main.c
security/integrity/evm/evm_secfs.c
security/integrity/ima/ima_crypto.c
security/integrity/ima/ima_fs.c
security/security.c
sound/core/pcm_lib.c
sound/pci/hda/patch_realtek.c
sound/pci/ice1712/ice1712.c
tools/testing/selftests/bpf/prog_tests/mmap.c
tools/testing/selftests/bpf/progs/test_mmap.c
tools/testing/selftests/drivers/net/mlxsw/qos_mc_aware.sh
tools/testing/selftests/drivers/net/netdevsim/devlink_trap.sh
tools/testing/selftests/vm/.gitignore
tools/testing/selftests/vm/write_to_hugetlbfs.c
tools/testing/selftests/wireguard/qemu/Makefile

index db3754a..4f906b4 100644 (file)
--- a/.mailmap
+++ b/.mailmap
@@ -288,6 +288,8 @@ Vladimir Davydov <vdavydov.dev@gmail.com> <vdavydov@virtuozzo.com>
 Vladimir Davydov <vdavydov.dev@gmail.com> <vdavydov@parallels.com>
 Takashi YOSHII <takashi.yoshii.zj@renesas.com>
 Will Deacon <will@kernel.org> <will.deacon@arm.com>
+Wolfram Sang <wsa@kernel.org> <wsa@the-dreams.de>
+Wolfram Sang <wsa@kernel.org> <w.sang@pengutronix.de>
 Yakir Yang <kuankuan.y@gmail.com> <ykk@rock-chips.com>
 Yusuke Goda <goda.yusuke@renesas.com>
 Gustavo Padovan <gustavo@las.ic.unicamp.br>
index 5201bc1..cfd1afd 100644 (file)
@@ -110,6 +110,9 @@ Ethernet switch connected via MDIO to the host, CPU port wired to eth0:
                        #size-cells = <0>;
 
                        ports {
+                               #address-cells = <1>;
+                               #size-cells = <0>;
+
                                port0@0 {
                                        reg = <0>;
                                        label = "lan1";
index 8a7be35..34bf84d 100644 (file)
@@ -5507,10 +5507,10 @@ F:      drivers/gpu/drm/vboxvideo/
 
 DRM DRIVER FOR VMWARE VIRTUAL GPU
 M:     "VMware Graphics" <linux-graphics-maintainer@vmware.com>
-M:     Thomas Hellstrom <thellstrom@vmware.com>
+M:     Roland Scheidegger <sroland@vmware.com>
 L:     dri-devel@lists.freedesktop.org
 S:     Supported
-T:     git git://people.freedesktop.org/~thomash/linux
+T:     git git://people.freedesktop.org/~sroland/linux
 F:     drivers/gpu/drm/vmwgfx/
 F:     include/uapi/drm/vmwgfx_drm.h
 
@@ -7829,7 +7829,7 @@ T:        git git://linuxtv.org/media_tree.git
 F:     drivers/media/platform/sti/hva
 
 HWPOISON MEMORY FAILURE HANDLING
-M:     Naoya Horiguchi <n-horiguchi@ah.jp.nec.com>
+M:     Naoya Horiguchi <naoya.horiguchi@nec.com>
 L:     linux-mm@kvack.org
 S:     Maintained
 F:     mm/hwpoison-inject.c
@@ -7941,7 +7941,7 @@ F:        Documentation/i2c/busses/i2c-parport.rst
 F:     drivers/i2c/busses/i2c-parport.c
 
 I2C SUBSYSTEM
-M:     Wolfram Sang <wsa@the-dreams.de>
+M:     Wolfram Sang <wsa@kernel.org>
 L:     linux-i2c@vger.kernel.org
 S:     Maintained
 W:     https://i2c.wiki.kernel.org/
@@ -9185,6 +9185,11 @@ L:       kexec@lists.infradead.org
 S:     Maintained
 W:     http://lse.sourceforge.net/kdump/
 F:     Documentation/admin-guide/kdump/
+F:     fs/proc/vmcore.c
+F:     include/linux/crash_core.h
+F:     include/linux/crash_dump.h
+F:     include/uapi/linux/vmcore.h
+F:     kernel/crash_*.c
 
 KEENE FM RADIO TRANSMITTER DRIVER
 M:     Hans Verkuil <hverkuil@xs4all.nl>
@@ -10669,6 +10674,13 @@ L:     netdev@vger.kernel.org
 S:     Maintained
 F:     drivers/net/ethernet/mediatek/
 
+MEDIATEK I2C CONTROLLER DRIVER
+M:     Qii Wang <qii.wang@mediatek.com>
+L:     linux-i2c@vger.kernel.org
+S:     Maintained
+F:     Documentation/devicetree/bindings/i2c/i2c-mt65xx.txt
+F:     drivers/i2c/busses/i2c-mt65xx.c
+
 MEDIATEK JPEG DRIVER
 M:     Rick Chang <rick.chang@mediatek.com>
 M:     Bin Liu <bin.liu@mediatek.com>
index 04f5662..f0d118b 100644 (file)
--- a/Makefile
+++ b/Makefile
@@ -2,7 +2,7 @@
 VERSION = 5
 PATCHLEVEL = 7
 SUBLEVEL = 0
-EXTRAVERSION = -rc6
+EXTRAVERSION = -rc7
 NAME = Kleptomaniac Octopus
 
 # *DOCUMENTATION*
index 0974226..aa00007 100644 (file)
@@ -65,6 +65,7 @@ CONFIG_DRM_UDL=y
 CONFIG_DRM_ETNAVIV=y
 CONFIG_FB=y
 CONFIG_FRAMEBUFFER_CONSOLE=y
+CONFIG_USB=y
 CONFIG_USB_EHCI_HCD=y
 CONFIG_USB_EHCI_HCD_PLATFORM=y
 CONFIG_USB_OHCI_HCD=y
index e1aa212..cd5636d 100644 (file)
 
 /* clobbers r5 register */
 .macro DSP_EARLY_INIT
+#ifdef CONFIG_ISA_ARCV2
        lr      r5, [ARC_AUX_DSP_BUILD]
        bmsk    r5, r5, 7
        breq    r5, 0, 1f
        mov     r5, DSP_CTRL_DISABLED_ALL
        sr      r5, [ARC_AUX_DSP_CTRL]
 1:
+#endif
 .endm
 
 /* clobbers r10, r11 registers pair */
index ae0aa53..0ff4c06 100644 (file)
 
 #ifdef CONFIG_ARC_IRQ_NO_AUTOSAVE
        __RESTORE_REGFILE_HARD
+
+       ; SP points to PC/STAT32: hw restores them despite NO_AUTOSAVE
        add     sp, sp, SZ_PT_REGS - 8
 #else
        add     sp, sp, PT_r0
index 7553967..8c4fc4b 100644 (file)
@@ -3,9 +3,6 @@
 # Copyright (C) 2004, 2007-2010, 2011-2012 Synopsys, Inc. (www.synopsys.com)
 #
 
-# Pass UTS_MACHINE for user_regset definition
-CFLAGS_ptrace.o                += -DUTS_MACHINE='"$(UTS_MACHINE)"'
-
 obj-y  := arcksyms.o setup.o irq.o reset.o ptrace.o process.o devtree.o
 obj-y  += signal.o traps.o sys.o troubleshoot.o stacktrace.o disasm.o
 obj-$(CONFIG_ISA_ARCOMPACT)            += entry-compact.o intc-compact.o
index d5f3fcf..f49a054 100644 (file)
@@ -253,7 +253,7 @@ static const struct user_regset arc_regsets[] = {
 };
 
 static const struct user_regset_view user_arc_view = {
-       .name           = UTS_MACHINE,
+       .name           = "arc",
        .e_machine      = EM_ARC_INUSE,
        .regsets        = arc_regsets,
        .n              = ARRAY_SIZE(arc_regsets)
index b2b1cb6..dad8a65 100644 (file)
@@ -11,6 +11,7 @@
 #include <linux/clocksource.h>
 #include <linux/console.h>
 #include <linux/module.h>
+#include <linux/sizes.h>
 #include <linux/cpu.h>
 #include <linux/of_clk.h>
 #include <linux/of_fdt.h>
@@ -424,12 +425,12 @@ static void arc_chk_core_config(void)
        if ((unsigned int)__arc_dccm_base != cpu->dccm.base_addr)
                panic("Linux built with incorrect DCCM Base address\n");
 
-       if (CONFIG_ARC_DCCM_SZ != cpu->dccm.sz)
+       if (CONFIG_ARC_DCCM_SZ * SZ_1K != cpu->dccm.sz)
                panic("Linux built with incorrect DCCM Size\n");
 #endif
 
 #ifdef CONFIG_ARC_HAS_ICCM
-       if (CONFIG_ARC_ICCM_SZ != cpu->iccm.sz)
+       if (CONFIG_ARC_ICCM_SZ * SZ_1K != cpu->iccm.sz)
                panic("Linux built with incorrect ICCM Size\n");
 #endif
 
index d299950..3393558 100644 (file)
@@ -191,10 +191,9 @@ void show_regs(struct pt_regs *regs)
        if (user_mode(regs))
                show_faulting_vma(regs->ret); /* faulting code, not data */
 
-       pr_info("ECR: 0x%08lx EFA: 0x%08lx ERET: 0x%08lx\n",
-               regs->event, current->thread.fault_address, regs->ret);
-
-       pr_info("STAT32: 0x%08lx", regs->status32);
+       pr_info("ECR: 0x%08lx EFA: 0x%08lx ERET: 0x%08lx\nSTAT: 0x%08lx",
+               regs->event, current->thread.fault_address, regs->ret,
+               regs->status32);
 
 #define STS_BIT(r, bit)        r->status32 & STATUS_##bit##_MASK ? #bit" " : ""
 
@@ -210,11 +209,10 @@ void show_regs(struct pt_regs *regs)
                        (regs->status32 & STATUS_U_MASK) ? "U " : "K ",
                        STS_BIT(regs, DE), STS_BIT(regs, AE));
 #endif
-       pr_cont("  BTA: 0x%08lx\n", regs->bta);
-       pr_info("BLK: %pS\n SP: 0x%08lx  FP: 0x%08lx\n",
-               (void *)regs->blink, regs->sp, regs->fp);
+       pr_cont("  BTA: 0x%08lx\n  SP: 0x%08lx  FP: 0x%08lx BLK: %pS\n",
+               regs->bta, regs->sp, regs->fp, (void *)regs->blink);
        pr_info("LPS: 0x%08lx\tLPE: 0x%08lx\tLPC: 0x%08lx\n",
-              regs->lp_start, regs->lp_end, regs->lp_count);
+               regs->lp_start, regs->lp_end, regs->lp_count);
 
        /* print regs->r0 thru regs->r12
         * Sequential printing was generating horrible code
index 27ea64b..f87758a 100644 (file)
@@ -1178,11 +1178,9 @@ int arc_unwind(struct unwind_frame_info *frame)
 #endif
 
        /* update frame */
-#ifndef CONFIG_AS_CFI_SIGNAL_FRAME
        if (frame->call_frame
            && !UNW_DEFAULT_RA(state.regs[retAddrReg], state.dataAlign))
                frame->call_frame = 0;
-#endif
        cfa = FRAME_REG(state.cfa.reg, unsigned long) + state.cfa.offs;
        startLoc = min_t(unsigned long, UNW_SP(frame), cfa);
        endLoc = max_t(unsigned long, UNW_SP(frame), cfa);
index a931d0a..a645bca 100644 (file)
@@ -6,6 +6,7 @@
 
 menuconfig ARC_PLAT_EZNPS
        bool "\"EZchip\" ARC dev platform"
+       depends on ISA_ARCOMPACT
        select CPU_BIG_ENDIAN
        select CLKSRC_NPS if !PHYS_ADDR_T_64BIT
        select EZNPS_GIC
index 32fc806..bc5c7b0 100644 (file)
@@ -304,7 +304,7 @@ do {                                                                        \
                __p = uaccess_mask_ptr(__p);                            \
                __raw_get_user((x), __p, (err));                        \
        } else {                                                        \
-               (x) = 0; (err) = -EFAULT;                               \
+               (x) = (__force __typeof__(x))0; (err) = -EFAULT;        \
        }                                                               \
 } while (0)
 
index b3d3005..e7b0190 100644 (file)
@@ -1829,10 +1829,11 @@ static void tracehook_report_syscall(struct pt_regs *regs,
 
 int syscall_trace_enter(struct pt_regs *regs)
 {
-       if (test_thread_flag(TIF_SYSCALL_TRACE) ||
-               test_thread_flag(TIF_SYSCALL_EMU)) {
+       unsigned long flags = READ_ONCE(current_thread_info()->flags);
+
+       if (flags & (_TIF_SYSCALL_EMU | _TIF_SYSCALL_TRACE)) {
                tracehook_report_syscall(regs, PTRACE_SYSCALL_ENTER);
-               if (!in_syscall(regs) || test_thread_flag(TIF_SYSCALL_EMU))
+               if (!in_syscall(regs) || (flags & _TIF_SYSCALL_EMU))
                        return -1;
        }
 
index 924c541..d13b532 100644 (file)
@@ -130,7 +130,7 @@ config PPC
        select ARCH_HAS_PTE_SPECIAL
        select ARCH_HAS_MEMBARRIER_CALLBACKS
        select ARCH_HAS_SCALED_CPUTIME          if VIRT_CPU_ACCOUNTING_NATIVE && PPC_BOOK3S_64
-       select ARCH_HAS_STRICT_KERNEL_RWX       if ((PPC_BOOK3S_64 || PPC32) && !HIBERNATION)
+       select ARCH_HAS_STRICT_KERNEL_RWX       if (PPC32 && !HIBERNATION)
        select ARCH_HAS_TICK_BROADCAST          if GENERIC_CLOCKEVENTS_BROADCAST
        select ARCH_HAS_UACCESS_FLUSHCACHE
        select ARCH_HAS_UACCESS_MCSAFE          if PPC64
index 34a7215..2a0a467 100644 (file)
@@ -17,9 +17,9 @@
  * updating the accessed and modified bits in the page table tree.
  */
 
-#define _PAGE_USER     0x001   /* usermode access allowed */
-#define _PAGE_RW       0x002   /* software: user write access allowed */
-#define _PAGE_PRESENT  0x004   /* software: pte contains a translation */
+#define _PAGE_PRESENT  0x001   /* software: pte contains a translation */
+#define _PAGE_HASHPTE  0x002   /* hash_page has made an HPTE for this pte */
+#define _PAGE_USER     0x004   /* usermode access allowed */
 #define _PAGE_GUARDED  0x008   /* G: prohibit speculative access */
 #define _PAGE_COHERENT 0x010   /* M: enforce memory coherence (SMP systems) */
 #define _PAGE_NO_CACHE 0x020   /* I: cache inhibit */
@@ -27,7 +27,7 @@
 #define _PAGE_DIRTY    0x080   /* C: page changed */
 #define _PAGE_ACCESSED 0x100   /* R: page referenced */
 #define _PAGE_EXEC     0x200   /* software: exec allowed */
-#define _PAGE_HASHPTE  0x400   /* hash_page has made an HPTE for this pte */
+#define _PAGE_RW       0x400   /* software: user write access allowed */
 #define _PAGE_SPECIAL  0x800   /* software: Special page */
 
 #ifdef CONFIG_PTE_64BIT
index daaa153..97c8879 100644 (file)
@@ -348,7 +348,7 @@ BEGIN_MMU_FTR_SECTION
        andis.  r0, r5, (DSISR_BAD_FAULT_32S | DSISR_DABRMATCH)@h
 #endif
        bne     handle_page_fault_tramp_2       /* if not, try to put a PTE */
-       rlwinm  r3, r5, 32 - 24, 30, 30         /* DSISR_STORE -> _PAGE_RW */
+       rlwinm  r3, r5, 32 - 15, 21, 21         /* DSISR_STORE -> _PAGE_RW */
        bl      hash_page
        b       handle_page_fault_tramp_1
 FTR_SECTION_ELSE
@@ -497,6 +497,7 @@ InstructionTLBMiss:
        andc.   r1,r1,r0                /* check access & ~permission */
        bne-    InstructionAddressInvalid /* return if access not permitted */
        /* Convert linux-style PTE to low word of PPC-style PTE */
+       rlwimi  r0,r0,32-2,31,31        /* _PAGE_USER -> PP lsb */
        ori     r1, r1, 0xe06           /* clear out reserved bits */
        andc    r1, r0, r1              /* PP = user? 1 : 0 */
 BEGIN_FTR_SECTION
@@ -564,8 +565,9 @@ DataLoadTLBMiss:
         * we would need to update the pte atomically with lwarx/stwcx.
         */
        /* Convert linux-style PTE to low word of PPC-style PTE */
-       rlwinm  r1,r0,0,30,30           /* _PAGE_RW -> PP msb */
-       rlwimi  r0,r0,1,30,30           /* _PAGE_USER -> PP msb */
+       rlwinm  r1,r0,32-9,30,30        /* _PAGE_RW -> PP msb */
+       rlwimi  r0,r0,32-1,30,30        /* _PAGE_USER -> PP msb */
+       rlwimi  r0,r0,32-1,31,31        /* _PAGE_USER -> PP lsb */
        ori     r1,r1,0xe04             /* clear out reserved bits */
        andc    r1,r0,r1                /* PP = user? rw? 1: 3: 0 */
 BEGIN_FTR_SECTION
@@ -643,6 +645,7 @@ DataStoreTLBMiss:
         * we would need to update the pte atomically with lwarx/stwcx.
         */
        /* Convert linux-style PTE to low word of PPC-style PTE */
+       rlwimi  r0,r0,32-2,31,31        /* _PAGE_USER -> PP lsb */
        li      r1,0xe06                /* clear out reserved bits & PP msb */
        andc    r1,r0,r1                /* PP = user? 1: 0 */
 BEGIN_FTR_SECTION
index 6d23608..877d880 100644 (file)
@@ -35,7 +35,7 @@ mmu_hash_lock:
 /*
  * Load a PTE into the hash table, if possible.
  * The address is in r4, and r3 contains an access flag:
- * _PAGE_RW (0x002) if a write.
+ * _PAGE_RW (0x400) if a write.
  * r9 contains the SRR1 value, from which we use the MSR_PR bit.
  * SPRG_THREAD contains the physical address of the current task's thread.
  *
@@ -69,7 +69,7 @@ _GLOBAL(hash_page)
        blt+    112f                    /* assume user more likely */
        lis     r5, (swapper_pg_dir - PAGE_OFFSET)@ha   /* if kernel address, use */
        addi    r5 ,r5 ,(swapper_pg_dir - PAGE_OFFSET)@l        /* kernel page table */
-       rlwimi  r3,r9,32-14,31,31       /* MSR_PR -> _PAGE_USER */
+       rlwimi  r3,r9,32-12,29,29       /* MSR_PR -> _PAGE_USER */
 112:
 #ifndef CONFIG_PTE_64BIT
        rlwimi  r5,r4,12,20,29          /* insert top 10 bits of address */
@@ -94,7 +94,7 @@ _GLOBAL(hash_page)
 #else
        rlwimi  r8,r4,23,20,28          /* compute pte address */
 #endif
-       rlwinm  r0,r3,6,24,24           /* _PAGE_RW access -> _PAGE_DIRTY */
+       rlwinm  r0,r3,32-3,24,24        /* _PAGE_RW access -> _PAGE_DIRTY */
        ori     r0,r0,_PAGE_ACCESSED|_PAGE_HASHPTE
 
        /*
@@ -310,9 +310,11 @@ Hash_msk = (((1 << Hash_bits) - 1) * 64)
 
 _GLOBAL(create_hpte)
        /* Convert linux-style PTE (r5) to low word of PPC-style PTE (r8) */
+       rlwinm  r8,r5,32-9,30,30        /* _PAGE_RW -> PP msb */
        rlwinm  r0,r5,32-6,30,30        /* _PAGE_DIRTY -> PP msb */
-       and     r8,r5,r0                /* writable if _RW & _DIRTY */
-       rlwimi  r5,r5,1,30,30           /* _PAGE_USER -> PP msb */
+       and     r8,r8,r0                /* writable if _RW & _DIRTY */
+       rlwimi  r5,r5,32-1,30,30        /* _PAGE_USER -> PP msb */
+       rlwimi  r5,r5,32-2,31,31        /* _PAGE_USER -> PP lsb */
        ori     r8,r8,0xe04             /* clear out reserved bits */
        andc    r8,r5,r8                /* PP = user? (rw&dirty? 1: 3): 0 */
 BEGIN_FTR_SECTION
@@ -564,7 +566,7 @@ _GLOBAL(flush_hash_pages)
 33:    lwarx   r8,0,r5                 /* fetch the pte flags word */
        andi.   r0,r8,_PAGE_HASHPTE
        beq     8f                      /* done if HASHPTE is already clear */
-       rlwinm  r8,r8,0,~_PAGE_HASHPTE  /* clear HASHPTE bit */
+       rlwinm  r8,r8,0,31,29           /* clear HASHPTE bit */
        stwcx.  r8,0,r5                 /* update the pte */
        bne-    33b
 
index 610c11e..824d117 100644 (file)
@@ -22,7 +22,7 @@
 #include <asm/switch_to.h>
 #include <asm/thread_info.h>
 
-unsigned long gp_in_global __asm__("gp");
+register unsigned long gp_in_global __asm__("gp");
 
 extern asmlinkage void ret_from_fork(void);
 extern asmlinkage void ret_from_kernel_thread(void);
index 27a3341..736de6c 100644 (file)
@@ -47,7 +47,7 @@ static void setup_zero_page(void)
        memset((void *)empty_zero_page, 0, PAGE_SIZE);
 }
 
-#ifdef CONFIG_DEBUG_VM
+#if defined(CONFIG_MMU) && defined(CONFIG_DEBUG_VM)
 static inline void print_mlk(char *name, unsigned long b, unsigned long t)
 {
        pr_notice("%12s : 0x%08lx - 0x%08lx   (%4ld kB)\n", name, b, t,
index cd060b5..e4dc64c 100644 (file)
@@ -8,6 +8,10 @@
 #include <linux/slab.h>
 #include <asm/pci_insn.h>
 
+/* I/O size constraints */
+#define ZPCI_MAX_READ_SIZE     8
+#define ZPCI_MAX_WRITE_SIZE    128
+
 /* I/O Map */
 #define ZPCI_IOMAP_SHIFT               48
 #define ZPCI_IOMAP_ADDR_BASE           0x8000000000000000UL
@@ -140,7 +144,8 @@ static inline int zpci_memcpy_fromio(void *dst,
 
        while (n > 0) {
                size = zpci_get_max_write_size((u64 __force) src,
-                                              (u64) dst, n, 8);
+                                              (u64) dst, n,
+                                              ZPCI_MAX_READ_SIZE);
                rc = zpci_read_single(dst, src, size);
                if (rc)
                        break;
@@ -161,7 +166,8 @@ static inline int zpci_memcpy_toio(volatile void __iomem *dst,
 
        while (n > 0) {
                size = zpci_get_max_write_size((u64 __force) dst,
-                                              (u64) src, n, 128);
+                                              (u64) src, n,
+                                              ZPCI_MAX_WRITE_SIZE);
                if (size > 8) /* main path */
                        rc = zpci_write_block(dst, src, size);
                else
index 8415ae7..f9e4baa 100644 (file)
@@ -151,7 +151,7 @@ static int kexec_file_add_initrd(struct kimage *image,
                buf.mem += crashk_res.start;
        buf.memsz = buf.bufsz;
 
-       data->parm->initrd_start = buf.mem;
+       data->parm->initrd_start = data->memsz;
        data->parm->initrd_size = buf.memsz;
        data->memsz += buf.memsz;
 
index d5035de..b7182ce 100644 (file)
@@ -28,6 +28,7 @@ int arch_kexec_do_relocs(int r_type, void *loc, unsigned long val,
                break;
        case R_390_64:          /* Direct 64 bit.  */
        case R_390_GLOB_DAT:
+       case R_390_JMP_SLOT:
                *(u64 *)loc = val;
                break;
        case R_390_PC16:        /* PC relative 16 bit.  */
index f01dadd..4632d4e 100644 (file)
@@ -159,10 +159,13 @@ void set_huge_pte_at(struct mm_struct *mm, unsigned long addr,
                rste &= ~_SEGMENT_ENTRY_NOEXEC;
 
        /* Set correct table type for 2G hugepages */
-       if ((pte_val(*ptep) & _REGION_ENTRY_TYPE_MASK) == _REGION_ENTRY_TYPE_R3)
-               rste |= _REGION_ENTRY_TYPE_R3 | _REGION3_ENTRY_LARGE;
-       else
+       if ((pte_val(*ptep) & _REGION_ENTRY_TYPE_MASK) == _REGION_ENTRY_TYPE_R3) {
+               if (likely(pte_present(pte)))
+                       rste |= _REGION3_ENTRY_LARGE;
+               rste |= _REGION_ENTRY_TYPE_R3;
+       } else if (likely(pte_present(pte)))
                rste |= _SEGMENT_ENTRY_LARGE;
+
        clear_huge_pte_skeys(mm, rste);
        pte_val(*ptep) = rste;
 }
index 7d42a87..020a2c5 100644 (file)
 #include <linux/mm.h>
 #include <linux/errno.h>
 #include <linux/pci.h>
+#include <asm/pci_io.h>
+#include <asm/pci_debug.h>
+
+static inline void zpci_err_mmio(u8 cc, u8 status, u64 offset)
+{
+       struct {
+               u64 offset;
+               u8 cc;
+               u8 status;
+       } data = {offset, cc, status};
+
+       zpci_err_hex(&data, sizeof(data));
+}
+
+static inline int __pcistb_mio_inuser(
+               void __iomem *ioaddr, const void __user *src,
+               u64 len, u8 *status)
+{
+       int cc = -ENXIO;
+
+       asm volatile (
+               "       sacf 256\n"
+               "0:     .insn   rsy,0xeb00000000d4,%[len],%[ioaddr],%[src]\n"
+               "1:     ipm     %[cc]\n"
+               "       srl     %[cc],28\n"
+               "2:     sacf 768\n"
+               EX_TABLE(0b, 2b) EX_TABLE(1b, 2b)
+               : [cc] "+d" (cc), [len] "+d" (len)
+               : [ioaddr] "a" (ioaddr), [src] "Q" (*((u8 __force *)src))
+               : "cc", "memory");
+       *status = len >> 24 & 0xff;
+       return cc;
+}
+
+static inline int __pcistg_mio_inuser(
+               void __iomem *ioaddr, const void __user *src,
+               u64 ulen, u8 *status)
+{
+       register u64 addr asm("2") = (u64 __force) ioaddr;
+       register u64 len asm("3") = ulen;
+       int cc = -ENXIO;
+       u64 val = 0;
+       u64 cnt = ulen;
+       u8 tmp;
+
+       /*
+        * copy 0 < @len <= 8 bytes from @src into the right most bytes of
+        * a register, then store it to PCI at @ioaddr while in secondary
+        * address space. pcistg then uses the user mappings.
+        */
+       asm volatile (
+               "       sacf    256\n"
+               "0:     llgc    %[tmp],0(%[src])\n"
+               "       sllg    %[val],%[val],8\n"
+               "       aghi    %[src],1\n"
+               "       ogr     %[val],%[tmp]\n"
+               "       brctg   %[cnt],0b\n"
+               "1:     .insn   rre,0xb9d40000,%[val],%[ioaddr]\n"
+               "2:     ipm     %[cc]\n"
+               "       srl     %[cc],28\n"
+               "3:     sacf    768\n"
+               EX_TABLE(0b, 3b) EX_TABLE(1b, 3b) EX_TABLE(2b, 3b)
+               :
+               [src] "+a" (src), [cnt] "+d" (cnt),
+               [val] "+d" (val), [tmp] "=d" (tmp),
+               [len] "+d" (len), [cc] "+d" (cc),
+               [ioaddr] "+a" (addr)
+               :: "cc", "memory");
+       *status = len >> 24 & 0xff;
+
+       /* did we read everything from user memory? */
+       if (!cc && cnt != 0)
+               cc = -EFAULT;
+
+       return cc;
+}
+
+static inline int __memcpy_toio_inuser(void __iomem *dst,
+                                  const void __user *src, size_t n)
+{
+       int size, rc = 0;
+       u8 status = 0;
+       mm_segment_t old_fs;
+
+       if (!src)
+               return -EINVAL;
+
+       old_fs = enable_sacf_uaccess();
+       while (n > 0) {
+               size = zpci_get_max_write_size((u64 __force) dst,
+                                              (u64 __force) src, n,
+                                              ZPCI_MAX_WRITE_SIZE);
+               if (size > 8) /* main path */
+                       rc = __pcistb_mio_inuser(dst, src, size, &status);
+               else
+                       rc = __pcistg_mio_inuser(dst, src, size, &status);
+               if (rc)
+                       break;
+               src += size;
+               dst += size;
+               n -= size;
+       }
+       disable_sacf_uaccess(old_fs);
+       if (rc)
+               zpci_err_mmio(rc, status, (__force u64) dst);
+       return rc;
+}
 
 static long get_pfn(unsigned long user_addr, unsigned long access,
                    unsigned long *pfn)
@@ -46,6 +153,20 @@ SYSCALL_DEFINE3(s390_pci_mmio_write, unsigned long, mmio_addr,
 
        if (length <= 0 || PAGE_SIZE - (mmio_addr & ~PAGE_MASK) < length)
                return -EINVAL;
+
+       /*
+        * Only support read access to MIO capable devices on a MIO enabled
+        * system. Otherwise we would have to check for every address if it is
+        * a special ZPCI_ADDR and we would have to do a get_pfn() which we
+        * don't need for MIO capable devices.
+        */
+       if (static_branch_likely(&have_mio)) {
+               ret = __memcpy_toio_inuser((void  __iomem *) mmio_addr,
+                                       user_buffer,
+                                       length);
+               return ret;
+       }
+
        if (length > 64) {
                buf = kmalloc(length, GFP_KERNEL);
                if (!buf)
@@ -56,7 +177,8 @@ SYSCALL_DEFINE3(s390_pci_mmio_write, unsigned long, mmio_addr,
        ret = get_pfn(mmio_addr, VM_WRITE, &pfn);
        if (ret)
                goto out;
-       io_addr = (void __iomem *)((pfn << PAGE_SHIFT) | (mmio_addr & ~PAGE_MASK));
+       io_addr = (void __iomem *)((pfn << PAGE_SHIFT) |
+                       (mmio_addr & ~PAGE_MASK));
 
        ret = -EFAULT;
        if ((unsigned long) io_addr < ZPCI_IOMAP_ADDR_BASE)
@@ -72,6 +194,78 @@ out:
        return ret;
 }
 
+static inline int __pcilg_mio_inuser(
+               void __user *dst, const void __iomem *ioaddr,
+               u64 ulen, u8 *status)
+{
+       register u64 addr asm("2") = (u64 __force) ioaddr;
+       register u64 len asm("3") = ulen;
+       u64 cnt = ulen;
+       int shift = ulen * 8;
+       int cc = -ENXIO;
+       u64 val, tmp;
+
+       /*
+        * read 0 < @len <= 8 bytes from the PCI memory mapped at @ioaddr (in
+        * user space) into a register using pcilg then store these bytes at
+        * user address @dst
+        */
+       asm volatile (
+               "       sacf    256\n"
+               "0:     .insn   rre,0xb9d60000,%[val],%[ioaddr]\n"
+               "1:     ipm     %[cc]\n"
+               "       srl     %[cc],28\n"
+               "       ltr     %[cc],%[cc]\n"
+               "       jne     4f\n"
+               "2:     ahi     %[shift],-8\n"
+               "       srlg    %[tmp],%[val],0(%[shift])\n"
+               "3:     stc     %[tmp],0(%[dst])\n"
+               "       aghi    %[dst],1\n"
+               "       brctg   %[cnt],2b\n"
+               "4:     sacf    768\n"
+               EX_TABLE(0b, 4b) EX_TABLE(1b, 4b) EX_TABLE(3b, 4b)
+               :
+               [cc] "+d" (cc), [val] "=d" (val), [len] "+d" (len),
+               [dst] "+a" (dst), [cnt] "+d" (cnt), [tmp] "=d" (tmp),
+               [shift] "+d" (shift)
+               :
+               [ioaddr] "a" (addr)
+               : "cc", "memory");
+
+       /* did we write everything to the user space buffer? */
+       if (!cc && cnt != 0)
+               cc = -EFAULT;
+
+       *status = len >> 24 & 0xff;
+       return cc;
+}
+
+static inline int __memcpy_fromio_inuser(void __user *dst,
+                                    const void __iomem *src,
+                                    unsigned long n)
+{
+       int size, rc = 0;
+       u8 status;
+       mm_segment_t old_fs;
+
+       old_fs = enable_sacf_uaccess();
+       while (n > 0) {
+               size = zpci_get_max_write_size((u64 __force) src,
+                                              (u64 __force) dst, n,
+                                              ZPCI_MAX_READ_SIZE);
+               rc = __pcilg_mio_inuser(dst, src, size, &status);
+               if (rc)
+                       break;
+               src += size;
+               dst += size;
+               n -= size;
+       }
+       disable_sacf_uaccess(old_fs);
+       if (rc)
+               zpci_err_mmio(rc, status, (__force u64) dst);
+       return rc;
+}
+
 SYSCALL_DEFINE3(s390_pci_mmio_read, unsigned long, mmio_addr,
                void __user *, user_buffer, size_t, length)
 {
@@ -86,12 +280,27 @@ SYSCALL_DEFINE3(s390_pci_mmio_read, unsigned long, mmio_addr,
 
        if (length <= 0 || PAGE_SIZE - (mmio_addr & ~PAGE_MASK) < length)
                return -EINVAL;
+
+       /*
+        * Only support write access to MIO capable devices on a MIO enabled
+        * system. Otherwise we would have to check for every address if it is
+        * a special ZPCI_ADDR and we would have to do a get_pfn() which we
+        * don't need for MIO capable devices.
+        */
+       if (static_branch_likely(&have_mio)) {
+               ret = __memcpy_fromio_inuser(
+                               user_buffer, (const void __iomem *)mmio_addr,
+                               length);
+               return ret;
+       }
+
        if (length > 64) {
                buf = kmalloc(length, GFP_KERNEL);
                if (!buf)
                        return -ENOMEM;
-       } else
+       } else {
                buf = local_buf;
+       }
 
        ret = get_pfn(mmio_addr, VM_READ, &pfn);
        if (ret)
index 3da5614..ef01ced 100644 (file)
@@ -2,6 +2,8 @@
 #ifndef __ASM_SH_SOCKIOS_H
 #define __ASM_SH_SOCKIOS_H
 
+#include <linux/time_types.h>
+
 /* Socket-level I/O control calls. */
 #define FIOGETOWN      _IOR('f', 123, int)
 #define FIOSETOWN      _IOW('f', 124, int)
index b7c94de..a8c2f26 100644 (file)
@@ -331,9 +331,9 @@ static void __init srmmu_nocache_init(void)
 
        while (vaddr < srmmu_nocache_end) {
                pgd = pgd_offset_k(vaddr);
-               p4d = p4d_offset(__nocache_fix(pgd), vaddr);
-               pud = pud_offset(__nocache_fix(p4d), vaddr);
-               pmd = pmd_offset(__nocache_fix(pgd), vaddr);
+               p4d = p4d_offset(pgd, vaddr);
+               pud = pud_offset(p4d, vaddr);
+               pmd = pmd_offset(__nocache_fix(pud), vaddr);
                pte = pte_offset_kernel(__nocache_fix(pmd), vaddr);
 
                pteval = ((paddr >> 4) | SRMMU_ET_PTE | SRMMU_PRIV);
index 91f35b2..d29d5fd 100644 (file)
@@ -17,7 +17,7 @@
 #define TRANS_TAP_LEN strlen(TRANS_TAP)
 
 #define TRANS_GRE "gre"
-#define TRANS_GRE_LEN strlen(TRANS_RAW)
+#define TRANS_GRE_LEN strlen(TRANS_GRE)
 
 #define TRANS_L2TPV3 "l2tpv3"
 #define TRANS_L2TPV3_LEN strlen(TRANS_L2TPV3)
index 7a3208c..36b33d6 100644 (file)
@@ -1,6 +1,6 @@
 /* SPDX-License-Identifier: GPL-2.0 */
 #include <asm-generic/xor.h>
-#include <shared/timer-internal.h>
+#include <linux/time-internal.h>
 
 /* pick an arbitrary one - measuring isn't possible with inf-cpu */
 #define XOR_SELECT_TEMPLATE(x) \
index 0a12d5a..3d91f89 100644 (file)
@@ -11,6 +11,7 @@
 #include <sysdep/ptrace_user.h>
 #include <sysdep/syscalls.h>
 #include <linux/time-internal.h>
+#include <asm/unistd.h>
 
 void handle_syscall(struct uml_pt_regs *r)
 {
index 8f8c8e3..c8b8c1a 100644 (file)
@@ -59,14 +59,14 @@ u8 buf[SETUP_SECT_MAX*512];
 #define PECOFF_COMPAT_RESERVE 0x0
 #endif
 
-unsigned long efi32_stub_entry;
-unsigned long efi64_stub_entry;
-unsigned long efi_pe_entry;
-unsigned long efi32_pe_entry;
-unsigned long kernel_info;
-unsigned long startup_64;
-unsigned long _ehead;
-unsigned long _end;
+static unsigned long efi32_stub_entry;
+static unsigned long efi64_stub_entry;
+static unsigned long efi_pe_entry;
+static unsigned long efi32_pe_entry;
+static unsigned long kernel_info;
+static unsigned long startup_64;
+static unsigned long _ehead;
+static unsigned long _end;
 
 /*----------------------------------------------------------------------*/
 
index fd51bac..acf76b4 100644 (file)
@@ -226,10 +226,18 @@ static int hv_cpu_die(unsigned int cpu)
 
        rdmsrl(HV_X64_MSR_REENLIGHTENMENT_CONTROL, *((u64 *)&re_ctrl));
        if (re_ctrl.target_vp == hv_vp_index[cpu]) {
-               /* Reassign to some other online CPU */
+               /*
+                * Reassign reenlightenment notifications to some other online
+                * CPU or just disable the feature if there are no online CPUs
+                * left (happens on hibernation).
+                */
                new_cpu = cpumask_any_but(cpu_online_mask, cpu);
 
-               re_ctrl.target_vp = hv_vp_index[new_cpu];
+               if (new_cpu < nr_cpu_ids)
+                       re_ctrl.target_vp = hv_vp_index[new_cpu];
+               else
+                       re_ctrl.enabled = 0;
+
                wrmsrl(HV_X64_MSR_REENLIGHTENMENT_CONTROL, *((u64 *)&re_ctrl));
        }
 
@@ -293,6 +301,13 @@ static void hv_resume(void)
 
        hv_hypercall_pg = hv_hypercall_pg_saved;
        hv_hypercall_pg_saved = NULL;
+
+       /*
+        * Reenlightenment notifications are disabled by hv_cpu_die(0),
+        * reenable them here if hv_reenlightenment_cb was previously set.
+        */
+       if (hv_reenlightenment_cb)
+               set_hv_tscchange_cb(hv_reenlightenment_cb);
 }
 
 /* Note: when the ops are called, only CPU0 is online and IRQs are disabled. */
index 53f246e..0367efd 100644 (file)
@@ -52,9 +52,9 @@ static __always_inline void
 arch_set_bit(long nr, volatile unsigned long *addr)
 {
        if (__builtin_constant_p(nr)) {
-               asm volatile(LOCK_PREFIX "orb %1,%0"
+               asm volatile(LOCK_PREFIX "orb %b1,%0"
                        : CONST_MASK_ADDR(nr, addr)
-                       : "iq" (CONST_MASK(nr) & 0xff)
+                       : "iq" (CONST_MASK(nr))
                        : "memory");
        } else {
                asm volatile(LOCK_PREFIX __ASM_SIZE(bts) " %1,%0"
@@ -72,9 +72,9 @@ static __always_inline void
 arch_clear_bit(long nr, volatile unsigned long *addr)
 {
        if (__builtin_constant_p(nr)) {
-               asm volatile(LOCK_PREFIX "andb %1,%0"
+               asm volatile(LOCK_PREFIX "andb %b1,%0"
                        : CONST_MASK_ADDR(nr, addr)
-                       : "iq" (CONST_MASK(nr) ^ 0xff));
+                       : "iq" (~CONST_MASK(nr)));
        } else {
                asm volatile(LOCK_PREFIX __ASM_SIZE(btr) " %1,%0"
                        : : RLONG_ADDR(addr), "Ir" (nr) : "memory");
@@ -123,9 +123,9 @@ static __always_inline void
 arch_change_bit(long nr, volatile unsigned long *addr)
 {
        if (__builtin_constant_p(nr)) {
-               asm volatile(LOCK_PREFIX "xorb %1,%0"
+               asm volatile(LOCK_PREFIX "xorb %b1,%0"
                        : CONST_MASK_ADDR(nr, addr)
-                       : "iq" ((u8)CONST_MASK(nr)));
+                       : "iq" (CONST_MASK(nr)));
        } else {
                asm volatile(LOCK_PREFIX __ASM_SIZE(btc) " %1,%0"
                        : : RLONG_ADDR(addr), "Ir" (nr) : "memory");
index fa79e42..7f969b2 100644 (file)
@@ -320,12 +320,19 @@ EXPORT_SYMBOL_GPL(unwind_get_return_address);
 
 unsigned long *unwind_get_return_address_ptr(struct unwind_state *state)
 {
+       struct task_struct *task = state->task;
+
        if (unwind_done(state))
                return NULL;
 
        if (state->regs)
                return &state->regs->ip;
 
+       if (task != current && state->sp == task->thread.sp) {
+               struct inactive_task_frame *frame = (void *)task->thread.sp;
+               return &frame->ret_addr;
+       }
+
        if (state->sp)
                return (unsigned long *)state->sp - 1;
 
index 109325d..43fd19b 100644 (file)
@@ -372,7 +372,7 @@ static void enter_uniprocessor(void)
        int cpu;
        int err;
 
-       if (downed_cpus == NULL &&
+       if (!cpumask_available(downed_cpus) &&
            !alloc_cpumask_var(&downed_cpus, GFP_KERNEL)) {
                pr_notice("Failed to allocate mask\n");
                goto out;
@@ -402,7 +402,7 @@ static void leave_uniprocessor(void)
        int cpu;
        int err;
 
-       if (downed_cpus == NULL || cpumask_weight(downed_cpus) == 0)
+       if (!cpumask_available(downed_cpus) || cpumask_weight(downed_cpus) == 0)
                return;
        pr_notice("Re-enabling CPUs...\n");
        for_each_cpu(cpu, downed_cpus) {
index 145ec0b..1af2125 100644 (file)
@@ -2016,9 +2016,13 @@ bool acpi_ec_dispatch_gpe(void)
         * to allow the caller to process events properly after that.
         */
        ret = acpi_dispatch_gpe(NULL, first_ec->gpe);
-       if (ret == ACPI_INTERRUPT_HANDLED)
+       if (ret == ACPI_INTERRUPT_HANDLED) {
                pm_pr_dbg("EC GPE dispatched\n");
 
+               /* Flush the event and query workqueues. */
+               acpi_ec_flush_work();
+       }
+
        return false;
 }
 #endif /* CONFIG_PM_SLEEP */
index 3850704..fd9d4e8 100644 (file)
@@ -980,13 +980,6 @@ static int acpi_s2idle_prepare_late(void)
        return 0;
 }
 
-static void acpi_s2idle_sync(void)
-{
-       /* The EC driver uses special workqueues that need to be flushed. */
-       acpi_ec_flush_work();
-       acpi_os_wait_events_complete(); /* synchronize Notify handling */
-}
-
 static bool acpi_s2idle_wake(void)
 {
        if (!acpi_sci_irq_valid())
@@ -1018,7 +1011,7 @@ static bool acpi_s2idle_wake(void)
                        return true;
 
                /*
-                * Cancel the wakeup and process all pending events in case
+                * Cancel the SCI wakeup and process all pending events in case
                 * there are any wakeup ones in there.
                 *
                 * Note that if any non-EC GPEs are active at this point, the
@@ -1026,8 +1019,7 @@ static bool acpi_s2idle_wake(void)
                 * should be missed by canceling the wakeup here.
                 */
                pm_system_cancel_wakeup();
-
-               acpi_s2idle_sync();
+               acpi_os_wait_events_complete();
 
                /*
                 * The SCI is in the "suspended" state now and it cannot produce
@@ -1060,7 +1052,8 @@ static void acpi_s2idle_restore(void)
         * of GPEs.
         */
        acpi_os_wait_events_complete(); /* synchronize GPE processing */
-       acpi_s2idle_sync();
+       acpi_ec_flush_work(); /* flush the EC driver's workqueues */
+       acpi_os_wait_events_complete(); /* synchronize Notify handling */
 
        s2idle_wakeup = false;
 
index 073045c..0cad34f 100644 (file)
@@ -365,6 +365,7 @@ struct device_link *device_link_add(struct device *consumer,
                                link->flags |= DL_FLAG_STATELESS;
                                goto reorder;
                        } else {
+                               link->flags |= DL_FLAG_STATELESS;
                                goto out;
                        }
                }
@@ -433,12 +434,16 @@ struct device_link *device_link_add(struct device *consumer,
            flags & DL_FLAG_PM_RUNTIME)
                pm_runtime_resume(supplier);
 
+       list_add_tail_rcu(&link->s_node, &supplier->links.consumers);
+       list_add_tail_rcu(&link->c_node, &consumer->links.suppliers);
+
        if (flags & DL_FLAG_SYNC_STATE_ONLY) {
                dev_dbg(consumer,
                        "Linked as a sync state only consumer to %s\n",
                        dev_name(supplier));
                goto out;
        }
+
 reorder:
        /*
         * Move the consumer and all of the devices depending on it to the end
@@ -449,12 +454,9 @@ reorder:
         */
        device_reorder_to_tail(consumer, NULL);
 
-       list_add_tail_rcu(&link->s_node, &supplier->links.consumers);
-       list_add_tail_rcu(&link->c_node, &consumer->links.suppliers);
-
        dev_dbg(consumer, "Linked as a consumer to %s\n", dev_name(supplier));
 
- out:
+out:
        device_pm_unlock();
        device_links_write_unlock();
 
@@ -829,6 +831,13 @@ static void __device_links_supplier_defer_sync(struct device *sup)
                list_add_tail(&sup->links.defer_sync, &deferred_sync);
 }
 
+static void device_link_drop_managed(struct device_link *link)
+{
+       link->flags &= ~DL_FLAG_MANAGED;
+       WRITE_ONCE(link->status, DL_STATE_NONE);
+       kref_put(&link->kref, __device_link_del);
+}
+
 /**
  * device_links_driver_bound - Update device links after probing its driver.
  * @dev: Device to update the links for.
@@ -842,7 +851,7 @@ static void __device_links_supplier_defer_sync(struct device *sup)
  */
 void device_links_driver_bound(struct device *dev)
 {
-       struct device_link *link;
+       struct device_link *link, *ln;
        LIST_HEAD(sync_list);
 
        /*
@@ -882,18 +891,35 @@ void device_links_driver_bound(struct device *dev)
        else
                __device_links_queue_sync_state(dev, &sync_list);
 
-       list_for_each_entry(link, &dev->links.suppliers, c_node) {
+       list_for_each_entry_safe(link, ln, &dev->links.suppliers, c_node) {
+               struct device *supplier;
+
                if (!(link->flags & DL_FLAG_MANAGED))
                        continue;
 
-               WARN_ON(link->status != DL_STATE_CONSUMER_PROBE);
-               WRITE_ONCE(link->status, DL_STATE_ACTIVE);
+               supplier = link->supplier;
+               if (link->flags & DL_FLAG_SYNC_STATE_ONLY) {
+                       /*
+                        * When DL_FLAG_SYNC_STATE_ONLY is set, it means no
+                        * other DL_MANAGED_LINK_FLAGS have been set. So, it's
+                        * save to drop the managed link completely.
+                        */
+                       device_link_drop_managed(link);
+               } else {
+                       WARN_ON(link->status != DL_STATE_CONSUMER_PROBE);
+                       WRITE_ONCE(link->status, DL_STATE_ACTIVE);
+               }
 
+               /*
+                * This needs to be done even for the deleted
+                * DL_FLAG_SYNC_STATE_ONLY device link in case it was the last
+                * device link that was preventing the supplier from getting a
+                * sync_state() call.
+                */
                if (defer_sync_state_count)
-                       __device_links_supplier_defer_sync(link->supplier);
+                       __device_links_supplier_defer_sync(supplier);
                else
-                       __device_links_queue_sync_state(link->supplier,
-                                                       &sync_list);
+                       __device_links_queue_sync_state(supplier, &sync_list);
        }
 
        dev->links.status = DL_DEV_DRIVER_BOUND;
@@ -903,13 +929,6 @@ void device_links_driver_bound(struct device *dev)
        device_links_flush_sync_list(&sync_list, dev);
 }
 
-static void device_link_drop_managed(struct device_link *link)
-{
-       link->flags &= ~DL_FLAG_MANAGED;
-       WRITE_ONCE(link->status, DL_STATE_NONE);
-       kref_put(&link->kref, __device_link_del);
-}
-
 /**
  * __device_links_no_driver - Update links of a device without a driver.
  * @dev: Device without a drvier.
index 8efd877..ce9e336 100644 (file)
@@ -1535,6 +1535,13 @@ static void null_config_discard(struct nullb *nullb)
 {
        if (nullb->dev->discard == false)
                return;
+
+       if (nullb->dev->zoned) {
+               nullb->dev->discard = false;
+               pr_info("discard option is ignored in zoned mode\n");
+               return;
+       }
+
        nullb->q->limits.discard_granularity = nullb->dev->blocksize;
        nullb->q->limits.discard_alignment = nullb->dev->blocksize;
        blk_queue_max_discard_sectors(nullb->q, UINT_MAX >> 9);
index 9e4bcda..ed5458f 100644 (file)
@@ -23,6 +23,10 @@ int null_init_zoned_dev(struct nullb_device *dev, struct request_queue *q)
                pr_err("zone_size must be power-of-two\n");
                return -EINVAL;
        }
+       if (dev->zone_size > dev->size) {
+               pr_err("Zone size larger than device capacity\n");
+               return -EINVAL;
+       }
 
        dev->zone_size_sects = dev->zone_size << ZONE_SIZE_SHIFT;
        dev->nr_zones = dev_size >>
index eb2ab05..1f8c826 100644 (file)
@@ -291,6 +291,7 @@ int mhi_init_dev_ctxt(struct mhi_controller *mhi_cntrl)
        }
 
        /* Setup cmd context */
+       ret = -ENOMEM;
        mhi_ctxt->cmd_ctxt = mhi_alloc_coherent(mhi_cntrl,
                                                sizeof(*mhi_ctxt->cmd_ctxt) *
                                                NR_OF_CMD_RINGS,
@@ -1100,6 +1101,7 @@ static int mhi_driver_probe(struct device *dev)
                }
        }
 
+       ret = -EINVAL;
        if (dl_chan) {
                /*
                 * If channel supports LPM notifications then status_cb should
index 3d0a7e7..1e678bd 100644 (file)
@@ -22,6 +22,7 @@ int dev_dax_kmem_probe(struct device *dev)
        resource_size_t kmem_size;
        resource_size_t kmem_end;
        struct resource *new_res;
+       const char *new_res_name;
        int numa_node;
        int rc;
 
@@ -48,11 +49,16 @@ int dev_dax_kmem_probe(struct device *dev)
        kmem_size &= ~(memory_block_size_bytes() - 1);
        kmem_end = kmem_start + kmem_size;
 
-       /* Region is permanently reserved.  Hot-remove not yet implemented. */
-       new_res = request_mem_region(kmem_start, kmem_size, dev_name(dev));
+       new_res_name = kstrdup(dev_name(dev), GFP_KERNEL);
+       if (!new_res_name)
+               return -ENOMEM;
+
+       /* Region is permanently reserved if hotremove fails. */
+       new_res = request_mem_region(kmem_start, kmem_size, new_res_name);
        if (!new_res) {
                dev_warn(dev, "could not reserve region [%pa-%pa]\n",
                         &kmem_start, &kmem_end);
+               kfree(new_res_name);
                return -EBUSY;
        }
 
@@ -63,12 +69,12 @@ int dev_dax_kmem_probe(struct device *dev)
         * unknown to us that will break add_memory() below.
         */
        new_res->flags = IORESOURCE_SYSTEM_RAM;
-       new_res->name = dev_name(dev);
 
        rc = add_memory(numa_node, new_res->start, resource_size(new_res));
        if (rc) {
                release_resource(new_res);
                kfree(new_res);
+               kfree(new_res_name);
                return rc;
        }
        dev_dax->dax_kmem_res = new_res;
@@ -83,6 +89,7 @@ static int dev_dax_kmem_remove(struct device *dev)
        struct resource *res = dev_dax->dax_kmem_res;
        resource_size_t kmem_start = res->start;
        resource_size_t kmem_size = resource_size(res);
+       const char *res_name = res->name;
        int rc;
 
        /*
@@ -102,6 +109,7 @@ static int dev_dax_kmem_remove(struct device *dev)
        /* Release and free dax resources */
        release_resource(res);
        kfree(res);
+       kfree(res_name);
        dev_dax->dax_kmem_res = NULL;
 
        return 0;
index 364dd34..0425984 100644 (file)
@@ -1166,10 +1166,11 @@ static int dmatest_run_set(const char *val, const struct kernel_param *kp)
                mutex_unlock(&info->lock);
                return ret;
        } else if (dmatest_run) {
-               if (is_threaded_test_pending(info))
-                       start_threaded_tests(info);
-               else
-                       pr_info("Could not start test, no channels configured\n");
+               if (!is_threaded_test_pending(info)) {
+                       pr_info("No channels configured, continue with any\n");
+                       add_threaded_test(info);
+               }
+               start_threaded_tests(info);
        } else {
                stop_threaded_test(info);
        }
index f6f49f0..8d79a87 100644 (file)
@@ -62,6 +62,13 @@ int idxd_unmask_msix_vector(struct idxd_device *idxd, int vec_id)
        perm.ignore = 0;
        iowrite32(perm.bits, idxd->reg_base + offset);
 
+       /*
+        * A readback from the device ensures that any previously generated
+        * completion record writes are visible to software based on PCI
+        * ordering rules.
+        */
+       perm.bits = ioread32(idxd->reg_base + offset);
+
        return 0;
 }
 
index d6fcd2e..6510791 100644 (file)
@@ -173,6 +173,7 @@ static int irq_process_pending_llist(struct idxd_irq_entry *irq_entry,
        struct llist_node *head;
        int queued = 0;
 
+       *processed = 0;
        head = llist_del_all(&irq_entry->pending_llist);
        if (!head)
                return 0;
@@ -197,6 +198,7 @@ static int irq_process_work_list(struct idxd_irq_entry *irq_entry,
        struct list_head *node, *next;
        int queued = 0;
 
+       *processed = 0;
        if (list_empty(&irq_entry->work_list))
                return 0;
 
@@ -218,10 +220,9 @@ static int irq_process_work_list(struct idxd_irq_entry *irq_entry,
        return queued;
 }
 
-irqreturn_t idxd_wq_thread(int irq, void *data)
+static int idxd_desc_process(struct idxd_irq_entry *irq_entry)
 {
-       struct idxd_irq_entry *irq_entry = data;
-       int rc, processed = 0, retry = 0;
+       int rc, processed, total = 0;
 
        /*
         * There are two lists we are processing. The pending_llist is where
@@ -244,15 +245,26 @@ irqreturn_t idxd_wq_thread(int irq, void *data)
         */
        do {
                rc = irq_process_work_list(irq_entry, &processed);
-               if (rc != 0) {
-                       retry++;
+               total += processed;
+               if (rc != 0)
                        continue;
-               }
 
                rc = irq_process_pending_llist(irq_entry, &processed);
-       } while (rc != 0 && retry != 10);
+               total += processed;
+       } while (rc != 0);
+
+       return total;
+}
+
+irqreturn_t idxd_wq_thread(int irq, void *data)
+{
+       struct idxd_irq_entry *irq_entry = data;
+       int processed;
 
+       processed = idxd_desc_process(irq_entry);
        idxd_unmask_msix_vector(irq_entry->idxd, irq_entry->id);
+       /* catch anything unprocessed after unmasking */
+       processed += idxd_desc_process(irq_entry);
 
        if (processed == 0)
                return IRQ_NONE;
index c683051..66ef70b 100644 (file)
@@ -175,13 +175,11 @@ struct owl_dma_txd {
  * @id: physical index to this channel
  * @base: virtual memory base for the dma channel
  * @vchan: the virtual channel currently being served by this physical channel
- * @lock: a lock to use when altering an instance of this struct
  */
 struct owl_dma_pchan {
        u32                     id;
        void __iomem            *base;
        struct owl_dma_vchan    *vchan;
-       spinlock_t              lock;
 };
 
 /**
@@ -437,14 +435,14 @@ static struct owl_dma_pchan *owl_dma_get_pchan(struct owl_dma *od,
        for (i = 0; i < od->nr_pchans; i++) {
                pchan = &od->pchans[i];
 
-               spin_lock_irqsave(&pchan->lock, flags);
+               spin_lock_irqsave(&od->lock, flags);
                if (!pchan->vchan) {
                        pchan->vchan = vchan;
-                       spin_unlock_irqrestore(&pchan->lock, flags);
+                       spin_unlock_irqrestore(&od->lock, flags);
                        break;
                }
 
-               spin_unlock_irqrestore(&pchan->lock, flags);
+               spin_unlock_irqrestore(&od->lock, flags);
        }
 
        return pchan;
index c4ce5df..db58d7e 100644 (file)
@@ -900,7 +900,7 @@ static int tegra_adma_probe(struct platform_device *pdev)
        ret = dma_async_device_register(&tdma->dma_dev);
        if (ret < 0) {
                dev_err(&pdev->dev, "ADMA registration failed: %d\n", ret);
-               goto irq_dispose;
+               goto rpm_put;
        }
 
        ret = of_dma_controller_register(pdev->dev.of_node,
index a9c0251..a90e154 100644 (file)
@@ -2156,7 +2156,8 @@ udma_prep_slave_sg_tr(struct udma_chan *uc, struct scatterlist *sgl,
                d->residue += sg_dma_len(sgent);
        }
 
-       cppi5_tr_csf_set(&tr_req[tr_idx - 1].flags, CPPI5_TR_CSF_EOP);
+       cppi5_tr_csf_set(&tr_req[tr_idx - 1].flags,
+                        CPPI5_TR_CSF_SUPR_EVT | CPPI5_TR_CSF_EOP);
 
        return d;
 }
@@ -2733,7 +2734,8 @@ udma_prep_dma_memcpy(struct dma_chan *chan, dma_addr_t dest, dma_addr_t src,
                tr_req[1].dicnt3 = 1;
        }
 
-       cppi5_tr_csf_set(&tr_req[num_tr - 1].flags, CPPI5_TR_CSF_EOP);
+       cppi5_tr_csf_set(&tr_req[num_tr - 1].flags,
+                        CPPI5_TR_CSF_SUPR_EVT | CPPI5_TR_CSF_EOP);
 
        if (uc->config.metadata_size)
                d->vd.tx.metadata_ops = &metadata_ops;
index d47749a..ff25369 100644 (file)
@@ -434,6 +434,7 @@ static void zynqmp_dma_free_descriptor(struct zynqmp_dma_chan *chan,
        struct zynqmp_dma_desc_sw *child, *next;
 
        chan->desc_free_cnt++;
+       list_del(&sdesc->node);
        list_add_tail(&sdesc->node, &chan->free_list);
        list_for_each_entry_safe(child, next, &sdesc->tx_list, node) {
                chan->desc_free_cnt++;
@@ -608,8 +609,6 @@ static void zynqmp_dma_chan_desc_cleanup(struct zynqmp_dma_chan *chan)
                dma_async_tx_callback callback;
                void *callback_param;
 
-               list_del(&desc->node);
-
                callback = desc->async_tx.callback;
                callback_param = desc->async_tx.callback_param;
                if (callback) {
index 9d25129..f564e15 100644 (file)
@@ -407,6 +407,58 @@ static void cper_print_pcie(const char *pfx, const struct cper_sec_pcie *pcie,
        }
 }
 
+static const char * const fw_err_rec_type_strs[] = {
+       "IPF SAL Error Record",
+       "SOC Firmware Error Record Type1 (Legacy CrashLog Support)",
+       "SOC Firmware Error Record Type2",
+};
+
+static void cper_print_fw_err(const char *pfx,
+                             struct acpi_hest_generic_data *gdata,
+                             const struct cper_sec_fw_err_rec_ref *fw_err)
+{
+       void *buf = acpi_hest_get_payload(gdata);
+       u32 offset, length = gdata->error_data_length;
+
+       printk("%s""Firmware Error Record Type: %s\n", pfx,
+              fw_err->record_type < ARRAY_SIZE(fw_err_rec_type_strs) ?
+              fw_err_rec_type_strs[fw_err->record_type] : "unknown");
+       printk("%s""Revision: %d\n", pfx, fw_err->revision);
+
+       /* Record Type based on UEFI 2.7 */
+       if (fw_err->revision == 0) {
+               printk("%s""Record Identifier: %08llx\n", pfx,
+                      fw_err->record_identifier);
+       } else if (fw_err->revision == 2) {
+               printk("%s""Record Identifier: %pUl\n", pfx,
+                      &fw_err->record_identifier_guid);
+       }
+
+       /*
+        * The FW error record may contain trailing data beyond the
+        * structure defined by the specification. As the fields
+        * defined (and hence the offset of any trailing data) vary
+        * with the revision, set the offset to account for this
+        * variation.
+        */
+       if (fw_err->revision == 0) {
+               /* record_identifier_guid not defined */
+               offset = offsetof(struct cper_sec_fw_err_rec_ref,
+                                 record_identifier_guid);
+       } else if (fw_err->revision == 1) {
+               /* record_identifier not defined */
+               offset = offsetof(struct cper_sec_fw_err_rec_ref,
+                                 record_identifier);
+       } else {
+               offset = sizeof(*fw_err);
+       }
+
+       buf += offset;
+       length -= offset;
+
+       print_hex_dump(pfx, "", DUMP_PREFIX_OFFSET, 16, 4, buf, length, true);
+}
+
 static void cper_print_tstamp(const char *pfx,
                                   struct acpi_hest_generic_data_v300 *gdata)
 {
@@ -494,6 +546,16 @@ cper_estatus_print_section(const char *pfx, struct acpi_hest_generic_data *gdata
                else
                        goto err_section_too_small;
 #endif
+       } else if (guid_equal(sec_type, &CPER_SEC_FW_ERR_REC_REF)) {
+               struct cper_sec_fw_err_rec_ref *fw_err = acpi_hest_get_payload(gdata);
+
+               printk("%ssection_type: Firmware Error Record Reference\n",
+                      newpfx);
+               /* The minimal FW Error Record contains 16 bytes */
+               if (gdata->error_data_length >= SZ_16)
+                       cper_print_fw_err(newpfx, gdata, fw_err);
+               else
+                       goto err_section_too_small;
        } else {
                const void *err = acpi_hest_get_payload(gdata);
 
index 5d4f847..a52236e 100644 (file)
@@ -114,14 +114,16 @@ static void efi_earlycon_write_char(u32 *dst, unsigned char c, unsigned int h)
        const u32 color_black = 0x00000000;
        const u32 color_white = 0x00ffffff;
        const u8 *src;
-       u8 s8;
-       int m;
+       int m, n, bytes;
+       u8 x;
 
-       src = font->data + c * font->height;
-       s8 = *(src + h);
+       bytes = BITS_TO_BYTES(font->width);
+       src = font->data + c * font->height * bytes + h * bytes;
 
-       for (m = 0; m < 8; m++) {
-               if ((s8 >> (7 - m)) & 1)
+       for (m = 0; m < font->width; m++) {
+               n = m % 8;
+               x = *(src + m / 8);
+               if ((x >> (7 - n)) & 1)
                        *dst = color_white;
                else
                        *dst = color_black;
index 911a2bd..4e30552 100644 (file)
@@ -130,11 +130,8 @@ static ssize_t systab_show(struct kobject *kobj,
        if (efi.smbios != EFI_INVALID_TABLE_ADDR)
                str += sprintf(str, "SMBIOS=0x%lx\n", efi.smbios);
 
-       if (IS_ENABLED(CONFIG_IA64) || IS_ENABLED(CONFIG_X86)) {
-               extern char *efi_systab_show_arch(char *str);
-
+       if (IS_ENABLED(CONFIG_IA64) || IS_ENABLED(CONFIG_X86))
                str = efi_systab_show_arch(str);
-       }
 
        return str - buf;
 }
index 99a5cde..48161b1 100644 (file)
@@ -60,7 +60,11 @@ static struct screen_info *setup_graphics(void)
                si = alloc_screen_info();
                if (!si)
                        return NULL;
-               efi_setup_gop(si, &gop_proto, size);
+               status = efi_setup_gop(si, &gop_proto, size);
+               if (status != EFI_SUCCESS) {
+                       free_screen_info(si);
+                       return NULL;
+               }
        }
        return si;
 }
index 67d2694..6294399 100644 (file)
@@ -92,6 +92,19 @@ extern __pure efi_system_table_t  *efi_system_table(void);
 #define EFI_LOCATE_BY_REGISTER_NOTIFY          1
 #define EFI_LOCATE_BY_PROTOCOL                 2
 
+/*
+ * An efi_boot_memmap is used by efi_get_memory_map() to return the
+ * EFI memory map in a dynamically allocated buffer.
+ *
+ * The buffer allocated for the EFI memory map includes extra room for
+ * a minimum of EFI_MMAP_NR_SLACK_SLOTS additional EFI memory descriptors.
+ * This facilitates the reuse of the EFI memory map buffer when a second
+ * call to ExitBootServices() is needed because of intervening changes to
+ * the EFI memory map. Other related structures, e.g. x86 e820ext, need
+ * to factor in this headroom requirement as well.
+ */
+#define EFI_MMAP_NR_SLACK_SLOTS        8
+
 struct efi_boot_memmap {
        efi_memory_desc_t       **map;
        unsigned long           *map_size;
index 869a79c..09f4fa0 100644 (file)
@@ -5,8 +5,6 @@
 
 #include "efistub.h"
 
-#define EFI_MMAP_NR_SLACK_SLOTS        8
-
 static inline bool mmap_has_headroom(unsigned long buff_size,
                                     unsigned long map_size,
                                     unsigned long desc_size)
index 1d59e10..e9a6846 100644 (file)
@@ -54,7 +54,7 @@ void efi_retrieve_tpm2_eventlog(void)
        efi_status_t status;
        efi_physical_addr_t log_location = 0, log_last_entry = 0;
        struct linux_efi_tpm_eventlog *log_tbl = NULL;
-       struct efi_tcg2_final_events_table *final_events_table;
+       struct efi_tcg2_final_events_table *final_events_table = NULL;
        unsigned long first_entry_addr, last_entry_addr;
        size_t log_size, last_entry_size;
        efi_bool_t truncated;
@@ -127,7 +127,8 @@ void efi_retrieve_tpm2_eventlog(void)
         * Figure out whether any events have already been logged to the
         * final events structure, and if so how much space they take up
         */
-       final_events_table = get_efi_config_table(LINUX_EFI_TPM_FINAL_LOG_GUID);
+       if (version == EFI_TCG2_EVENT_LOG_FORMAT_TCG_2)
+               final_events_table = get_efi_config_table(LINUX_EFI_TPM_FINAL_LOG_GUID);
        if (final_events_table && final_events_table->nr_events) {
                struct tcg_pcr_event2_head *header;
                int offset;
index 05ccb22..f0339b5 100644 (file)
@@ -606,24 +606,18 @@ static efi_status_t allocate_e820(struct boot_params *params,
                                  struct setup_data **e820ext,
                                  u32 *e820ext_size)
 {
-       unsigned long map_size, desc_size, buff_size;
-       struct efi_boot_memmap boot_map;
-       efi_memory_desc_t *map;
+       unsigned long map_size, desc_size, map_key;
        efi_status_t status;
-       __u32 nr_desc;
+       __u32 nr_desc, desc_version;
 
-       boot_map.map            = &map;
-       boot_map.map_size       = &map_size;
-       boot_map.desc_size      = &desc_size;
-       boot_map.desc_ver       = NULL;
-       boot_map.key_ptr        = NULL;
-       boot_map.buff_size      = &buff_size;
+       /* Only need the size of the mem map and size of each mem descriptor */
+       map_size = 0;
+       status = efi_bs_call(get_memory_map, &map_size, NULL, &map_key,
+                            &desc_size, &desc_version);
+       if (status != EFI_BUFFER_TOO_SMALL)
+               return (status != EFI_SUCCESS) ? status : EFI_UNSUPPORTED;
 
-       status = efi_get_memory_map(&boot_map);
-       if (status != EFI_SUCCESS)
-               return status;
-
-       nr_desc = buff_size / desc_size;
+       nr_desc = map_size / desc_size + EFI_MMAP_NR_SLACK_SLOTS;
 
        if (nr_desc > ARRAY_SIZE(params->e820_table)) {
                u32 nr_e820ext = nr_desc - ARRAY_SIZE(params->e820_table);
index 55b031d..c1955d3 100644 (file)
@@ -62,8 +62,11 @@ int __init efi_tpm_eventlog_init(void)
        tbl_size = sizeof(*log_tbl) + log_tbl->size;
        memblock_reserve(efi.tpm_log, tbl_size);
 
-       if (efi.tpm_final_log == EFI_INVALID_TABLE_ADDR)
+       if (efi.tpm_final_log == EFI_INVALID_TABLE_ADDR ||
+           log_tbl->version != EFI_TCG2_EVENT_LOG_FORMAT_TCG_2) {
+               pr_warn(FW_BUG "TPM Final Events table missing or invalid\n");
                goto out;
+       }
 
        final_tbl = early_memremap(efi.tpm_final_log, sizeof(*final_tbl));
 
index 27a7d2a..caa090d 100644 (file)
@@ -220,6 +220,30 @@ static enum dpcd_training_patterns
        return dpcd_tr_pattern;
 }
 
+static uint8_t dc_dp_initialize_scrambling_data_symbols(
+       struct dc_link *link,
+       enum dc_dp_training_pattern pattern)
+{
+       uint8_t disable_scrabled_data_symbols = 0;
+
+       switch (pattern) {
+       case DP_TRAINING_PATTERN_SEQUENCE_1:
+       case DP_TRAINING_PATTERN_SEQUENCE_2:
+       case DP_TRAINING_PATTERN_SEQUENCE_3:
+               disable_scrabled_data_symbols = 1;
+               break;
+       case DP_TRAINING_PATTERN_SEQUENCE_4:
+               disable_scrabled_data_symbols = 0;
+               break;
+       default:
+               ASSERT(0);
+               DC_LOG_HW_LINK_TRAINING("%s: Invalid HW Training pattern: %d\n",
+                       __func__, pattern);
+               break;
+       }
+       return disable_scrabled_data_symbols;
+}
+
 static inline bool is_repeater(struct dc_link *link, uint32_t offset)
 {
        return (!link->is_lttpr_mode_transparent && offset != 0);
@@ -252,6 +276,9 @@ static void dpcd_set_lt_pattern_and_lane_settings(
        dpcd_pattern.v1_4.TRAINING_PATTERN_SET =
                dc_dp_training_pattern_to_dpcd_training_pattern(link, pattern);
 
+       dpcd_pattern.v1_4.SCRAMBLING_DISABLE =
+               dc_dp_initialize_scrambling_data_symbols(link, pattern);
+
        dpcd_lt_buffer[DP_TRAINING_PATTERN_SET - DP_TRAINING_PATTERN_SET]
                = dpcd_pattern.raw;
 
index 085c1a3..82fc3d5 100644 (file)
@@ -1625,12 +1625,79 @@ void dcn10_pipe_control_lock(
                hws->funcs.verify_allow_pstate_change_high(dc);
 }
 
+/**
+ * delay_cursor_until_vupdate() - Delay cursor update if too close to VUPDATE.
+ *
+ * Software keepout workaround to prevent cursor update locking from stalling
+ * out cursor updates indefinitely or from old values from being retained in
+ * the case where the viewport changes in the same frame as the cursor.
+ *
+ * The idea is to calculate the remaining time from VPOS to VUPDATE. If it's
+ * too close to VUPDATE, then stall out until VUPDATE finishes.
+ *
+ * TODO: Optimize cursor programming to be once per frame before VUPDATE
+ *       to avoid the need for this workaround.
+ */
+static void delay_cursor_until_vupdate(struct dc *dc, struct pipe_ctx *pipe_ctx)
+{
+       struct dc_stream_state *stream = pipe_ctx->stream;
+       struct crtc_position position;
+       uint32_t vupdate_start, vupdate_end;
+       unsigned int lines_to_vupdate, us_to_vupdate, vpos;
+       unsigned int us_per_line, us_vupdate;
+
+       if (!dc->hwss.calc_vupdate_position || !dc->hwss.get_position)
+               return;
+
+       if (!pipe_ctx->stream_res.stream_enc || !pipe_ctx->stream_res.tg)
+               return;
+
+       dc->hwss.calc_vupdate_position(dc, pipe_ctx, &vupdate_start,
+                                      &vupdate_end);
+
+       dc->hwss.get_position(&pipe_ctx, 1, &position);
+       vpos = position.vertical_count;
+
+       /* Avoid wraparound calculation issues */
+       vupdate_start += stream->timing.v_total;
+       vupdate_end += stream->timing.v_total;
+       vpos += stream->timing.v_total;
+
+       if (vpos <= vupdate_start) {
+               /* VPOS is in VACTIVE or back porch. */
+               lines_to_vupdate = vupdate_start - vpos;
+       } else if (vpos > vupdate_end) {
+               /* VPOS is in the front porch. */
+               return;
+       } else {
+               /* VPOS is in VUPDATE. */
+               lines_to_vupdate = 0;
+       }
+
+       /* Calculate time until VUPDATE in microseconds. */
+       us_per_line =
+               stream->timing.h_total * 10000u / stream->timing.pix_clk_100hz;
+       us_to_vupdate = lines_to_vupdate * us_per_line;
+
+       /* 70 us is a conservative estimate of cursor update time*/
+       if (us_to_vupdate > 70)
+               return;
+
+       /* Stall out until the cursor update completes. */
+       us_vupdate = (vupdate_end - vupdate_start + 1) * us_per_line;
+       udelay(us_to_vupdate + us_vupdate);
+}
+
 void dcn10_cursor_lock(struct dc *dc, struct pipe_ctx *pipe, bool lock)
 {
        /* cursor lock is per MPCC tree, so only need to lock one pipe per stream */
        if (!pipe || pipe->top_pipe)
                return;
 
+       /* Prevent cursor lock from stalling out cursor updates. */
+       if (lock)
+               delay_cursor_until_vupdate(dc, pipe);
+
        dc->res_pool->mpc->funcs->cursor_lock(dc->res_pool->mpc,
                        pipe->stream_res.opp->inst, lock);
 }
@@ -3236,7 +3303,7 @@ int dcn10_get_vupdate_offset_from_vsync(struct pipe_ctx *pipe_ctx)
        return vertical_line_start;
 }
 
-static void dcn10_calc_vupdate_position(
+void dcn10_calc_vupdate_position(
                struct dc *dc,
                struct pipe_ctx *pipe_ctx,
                uint32_t *start_line,
index af51424..42b6e01 100644 (file)
@@ -34,6 +34,11 @@ struct dc;
 void dcn10_hw_sequencer_construct(struct dc *dc);
 
 int dcn10_get_vupdate_offset_from_vsync(struct pipe_ctx *pipe_ctx);
+void dcn10_calc_vupdate_position(
+               struct dc *dc,
+               struct pipe_ctx *pipe_ctx,
+               uint32_t *start_line,
+               uint32_t *end_line);
 void dcn10_setup_vupdate_interrupt(struct dc *dc, struct pipe_ctx *pipe_ctx);
 enum dc_status dcn10_enable_stream_timing(
                struct pipe_ctx *pipe_ctx,
index 700509b..9e8e326 100644 (file)
@@ -72,6 +72,7 @@ static const struct hw_sequencer_funcs dcn10_funcs = {
        .set_clock = dcn10_set_clock,
        .get_clock = dcn10_get_clock,
        .get_vupdate_offset_from_vsync = dcn10_get_vupdate_offset_from_vsync,
+       .calc_vupdate_position = dcn10_calc_vupdate_position,
 };
 
 static const struct hwseq_private_funcs dcn10_private_funcs = {
index 6a21228..8334bbd 100644 (file)
@@ -83,6 +83,7 @@ static const struct hw_sequencer_funcs dcn20_funcs = {
        .init_vm_ctx = dcn20_init_vm_ctx,
        .set_flip_control_gsl = dcn20_set_flip_control_gsl,
        .get_vupdate_offset_from_vsync = dcn10_get_vupdate_offset_from_vsync,
+       .calc_vupdate_position = dcn10_calc_vupdate_position,
 };
 
 static const struct hwseq_private_funcs dcn20_private_funcs = {
index 707ce0f..4dd6341 100644 (file)
@@ -86,6 +86,7 @@ static const struct hw_sequencer_funcs dcn21_funcs = {
        .optimize_pwr_state = dcn21_optimize_pwr_state,
        .exit_optimized_pwr_state = dcn21_exit_optimized_pwr_state,
        .get_vupdate_offset_from_vsync = dcn10_get_vupdate_offset_from_vsync,
+       .calc_vupdate_position = dcn10_calc_vupdate_position,
        .set_cursor_position = dcn10_set_cursor_position,
        .set_cursor_attribute = dcn10_set_cursor_attribute,
        .set_cursor_sdr_white_level = dcn10_set_cursor_sdr_white_level,
index 7ee8b84..e34c337 100644 (file)
@@ -63,10 +63,8 @@ CFLAGS_$(AMDDALPATH)/dc/dml/dcn21/display_rq_dlg_calc_21.o := $(dml_ccflags)
 endif
 CFLAGS_$(AMDDALPATH)/dc/dml/dml1_display_rq_dlg_calc.o := $(dml_ccflags)
 CFLAGS_$(AMDDALPATH)/dc/dml/display_rq_dlg_helpers.o := $(dml_ccflags)
-CFLAGS_$(AMDDALPATH)/dc/dml/dml_common_defs.o := $(dml_ccflags)
 
 DML = display_mode_lib.o display_rq_dlg_helpers.o dml1_display_rq_dlg_calc.o \
-       dml_common_defs.o
 
 ifdef CONFIG_DRM_AMD_DC_DCN
 DML += display_mode_vba.o dcn20/display_rq_dlg_calc_20.o dcn20/display_mode_vba_20.o
index 8c86b63..1e557dd 100644 (file)
@@ -26,7 +26,6 @@
 #ifndef __DML20_DISPLAY_RQ_DLG_CALC_H__
 #define __DML20_DISPLAY_RQ_DLG_CALC_H__
 
-#include "../dml_common_defs.h"
 #include "../display_rq_dlg_helpers.h"
 
 struct display_mode_lib;
index 0378406..0d53e87 100644 (file)
@@ -26,7 +26,6 @@
 #ifndef __DML20V2_DISPLAY_RQ_DLG_CALC_H__
 #define __DML20V2_DISPLAY_RQ_DLG_CALC_H__
 
-#include "../dml_common_defs.h"
 #include "../display_rq_dlg_helpers.h"
 
 struct display_mode_lib;
index 83e95f8..e8f7785 100644 (file)
@@ -26,7 +26,7 @@
 #ifndef __DML21_DISPLAY_RQ_DLG_CALC_H__
 #define __DML21_DISPLAY_RQ_DLG_CALC_H__
 
-#include "../dml_common_defs.h"
+#include "dm_services.h"
 #include "../display_rq_dlg_helpers.h"
 
 struct display_mode_lib;
index cf2758c..c77c3d8 100644 (file)
 #ifndef __DISPLAY_MODE_LIB_H__
 #define __DISPLAY_MODE_LIB_H__
 
-
-#include "dml_common_defs.h"
+#include "dm_services.h"
+#include "dc_features.h"
+#include "display_mode_structs.h"
+#include "display_mode_enums.h"
 #include "display_mode_vba.h"
 
 enum dml_project {
index 5d82fc5..3a73417 100644 (file)
@@ -27,8 +27,6 @@
 #ifndef __DML2_DISPLAY_MODE_VBA_H__
 #define __DML2_DISPLAY_MODE_VBA_H__
 
-#include "dml_common_defs.h"
-
 struct display_mode_lib;
 
 void ModeSupportAndSystemConfiguration(struct display_mode_lib *mode_lib);
index 1f24db8..2555ef0 100644 (file)
@@ -26,7 +26,6 @@
 #ifndef __DISPLAY_RQ_DLG_HELPERS_H__
 #define __DISPLAY_RQ_DLG_HELPERS_H__
 
-#include "dml_common_defs.h"
 #include "display_mode_lib.h"
 
 /* Function: Printer functions
index 3041649..9c06913 100644 (file)
@@ -26,8 +26,6 @@
 #ifndef __DISPLAY_RQ_DLG_CALC_H__
 #define __DISPLAY_RQ_DLG_CALC_H__
 
-#include "dml_common_defs.h"
-
 struct display_mode_lib;
 
 #include "display_rq_dlg_helpers.h"
diff --git a/drivers/gpu/drm/amd/display/dc/dml/dml_common_defs.c b/drivers/gpu/drm/amd/display/dc/dml/dml_common_defs.c
deleted file mode 100644 (file)
index 723af0b..0000000
+++ /dev/null
@@ -1,43 +0,0 @@
-/*
- * Copyright 2017 Advanced Micro Devices, Inc.
- *
- * Permission is hereby granted, free of charge, to any person obtaining a
- * copy of this software and associated documentation files (the "Software"),
- * to deal in the Software without restriction, including without limitation
- * the rights to use, copy, modify, merge, publish, distribute, sublicense,
- * and/or sell copies of the Software, and to permit persons to whom the
- * Software is furnished to do so, subject to the following conditions:
- *
- * The above copyright notice and this permission notice shall be included in
- * all copies or substantial portions of the Software.
- *
- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
- * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
- * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
- * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
- * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
- * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
- * OTHER DEALINGS IN THE SOFTWARE.
- *
- * Authors: AMD
- *
- */
-
-#include "dml_common_defs.h"
-#include "dcn_calc_math.h"
-
-#include "dml_inline_defs.h"
-
-double dml_round(double a)
-{
-       double round_pt = 0.5;
-       double ceil = dml_ceil(a, 1);
-       double floor = dml_floor(a, 1);
-
-       if (a - floor >= round_pt)
-               return ceil;
-       else
-               return floor;
-}
-
-
diff --git a/drivers/gpu/drm/amd/display/dc/dml/dml_common_defs.h b/drivers/gpu/drm/amd/display/dc/dml/dml_common_defs.h
deleted file mode 100644 (file)
index f78cbae..0000000
+++ /dev/null
@@ -1,37 +0,0 @@
-/*
- * Copyright 2017 Advanced Micro Devices, Inc.
- *
- * Permission is hereby granted, free of charge, to any person obtaining a
- * copy of this software and associated documentation files (the "Software"),
- * to deal in the Software without restriction, including without limitation
- * the rights to use, copy, modify, merge, publish, distribute, sublicense,
- * and/or sell copies of the Software, and to permit persons to whom the
- * Software is furnished to do so, subject to the following conditions:
- *
- * The above copyright notice and this permission notice shall be included in
- * all copies or substantial portions of the Software.
- *
- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
- * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
- * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
- * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
- * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
- * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
- * OTHER DEALINGS IN THE SOFTWARE.
- *
- * Authors: AMD
- *
- */
-
-#ifndef __DC_COMMON_DEFS_H__
-#define __DC_COMMON_DEFS_H__
-
-#include "dm_services.h"
-#include "dc_features.h"
-#include "display_mode_structs.h"
-#include "display_mode_enums.h"
-
-
-double dml_round(double a);
-
-#endif /* __DC_COMMON_DEFS_H__ */
index ded71ea..02e06c9 100644 (file)
@@ -26,7 +26,6 @@
 #ifndef __DML_INLINE_DEFS_H__
 #define __DML_INLINE_DEFS_H__
 
-#include "dml_common_defs.h"
 #include "dcn_calc_math.h"
 #include "dml_logger.h"
 
@@ -75,6 +74,18 @@ static inline double dml_floor(double a, double granularity)
        return (double) dcn_bw_floor2(a, granularity);
 }
 
+static inline double dml_round(double a)
+{
+       double round_pt = 0.5;
+       double ceil = dml_ceil(a, 1);
+       double floor = dml_floor(a, 1);
+
+       if (a - floor >= round_pt)
+               return ceil;
+       else
+               return floor;
+}
+
 static inline int dml_log2(double x)
 {
        return dml_round((double)dcn_bw_log(x, 2));
@@ -112,7 +123,7 @@ static inline double dml_log(double x, double base)
 
 static inline unsigned int dml_round_to_multiple(unsigned int num,
                                                 unsigned int multiple,
-                                                bool up)
+                                                unsigned char up)
 {
        unsigned int remainder;
 
index e57467d..08307f3 100644 (file)
@@ -92,6 +92,11 @@ struct hw_sequencer_funcs {
        void (*get_position)(struct pipe_ctx **pipe_ctx, int num_pipes,
                        struct crtc_position *position);
        int (*get_vupdate_offset_from_vsync)(struct pipe_ctx *pipe_ctx);
+       void (*calc_vupdate_position)(
+                       struct dc *dc,
+                       struct pipe_ctx *pipe_ctx,
+                       uint32_t *start_line,
+                       uint32_t *end_line);
        void (*enable_per_frame_crtc_position_reset)(struct dc *dc,
                        int group_size, struct pipe_ctx *grouped_pipes[]);
        void (*enable_timing_synchronization)(struct dc *dc,
index 4ede08a..d96e3ce 100644 (file)
@@ -191,10 +191,11 @@ static const struct edid_quirk {
        { "HVR", 0xaa01, EDID_QUIRK_NON_DESKTOP },
        { "HVR", 0xaa02, EDID_QUIRK_NON_DESKTOP },
 
-       /* Oculus Rift DK1, DK2, and CV1 VR Headsets */
+       /* Oculus Rift DK1, DK2, CV1 and Rift S VR Headsets */
        { "OVR", 0x0001, EDID_QUIRK_NON_DESKTOP },
        { "OVR", 0x0003, EDID_QUIRK_NON_DESKTOP },
        { "OVR", 0x0004, EDID_QUIRK_NON_DESKTOP },
+       { "OVR", 0x0012, EDID_QUIRK_NON_DESKTOP },
 
        /* Windows Mixed Reality Headsets */
        { "ACR", 0x7fce, EDID_QUIRK_NON_DESKTOP },
index 3b0afa1..54def34 100644 (file)
@@ -238,8 +238,10 @@ static int submit_pin_objects(struct etnaviv_gem_submit *submit)
                }
 
                if ((submit->flags & ETNA_SUBMIT_SOFTPIN) &&
-                    submit->bos[i].va != mapping->iova)
+                    submit->bos[i].va != mapping->iova) {
+                       etnaviv_gem_mapping_unreference(mapping);
                        return -EINVAL;
+               }
 
                atomic_inc(&etnaviv_obj->gpu_active);
 
index e6795ba..75f9db8 100644 (file)
@@ -453,7 +453,7 @@ static const struct etnaviv_pm_domain *pm_domain(const struct etnaviv_gpu *gpu,
                if (!(gpu->identity.features & meta->feature))
                        continue;
 
-               if (meta->nr_domains < (index - offset)) {
+               if (index - offset >= meta->nr_domains) {
                        offset += meta->nr_domains;
                        continue;
                }
index 8cdcd6e..3596f39 100644 (file)
@@ -850,7 +850,7 @@ extern void vmw_bo_bo_free(struct ttm_buffer_object *bo);
 extern int vmw_bo_init(struct vmw_private *dev_priv,
                       struct vmw_buffer_object *vmw_bo,
                       size_t size, struct ttm_placement *placement,
-                      bool interuptable,
+                      bool interruptible,
                       void (*bo_free)(struct ttm_buffer_object *bo));
 extern int vmw_user_bo_verify_access(struct ttm_buffer_object *bo,
                                     struct ttm_object_file *tfile);
index 178a6cd..0f8d293 100644 (file)
@@ -515,7 +515,7 @@ bool vmw_fence_obj_signaled(struct vmw_fence_obj *fence)
        struct vmw_fence_manager *fman = fman_from_fence(fence);
 
        if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->base.flags))
-               return 1;
+               return true;
 
        vmw_fences_update(fman);
 
index 7ef51fa..126f93c 100644 (file)
@@ -1651,7 +1651,7 @@ vmw_gb_surface_reference_internal(struct drm_device *dev,
        struct vmw_surface_metadata *metadata;
        struct ttm_base_object *base;
        uint32_t backup_handle;
-       int ret = -EINVAL;
+       int ret;
 
        ret = vmw_surface_handle_reference(dev_priv, file_priv, req->sid,
                                           req->handle_type, &base);
index b44d831..2fdaeec 100644 (file)
@@ -120,7 +120,7 @@ static int cti_plat_create_v8_etm_connection(struct device *dev,
 
        /* Can optionally have an etm node - return if not  */
        cs_fwnode = fwnode_find_reference(root_fwnode, CTI_DT_CSDEV_ASSOC, 0);
-       if (IS_ERR_OR_NULL(cs_fwnode))
+       if (IS_ERR(cs_fwnode))
                return 0;
 
        /* allocate memory */
@@ -393,7 +393,7 @@ static int cti_plat_create_connection(struct device *dev,
                /* associated device ? */
                cs_fwnode = fwnode_find_reference(fwnode,
                                                  CTI_DT_CSDEV_ASSOC, 0);
-               if (!IS_ERR_OR_NULL(cs_fwnode)) {
+               if (!IS_ERR(cs_fwnode)) {
                        assoc_name = cti_plat_get_csdev_or_node_name(cs_fwnode,
                                                                     &csdev);
                        fwnode_handle_put(cs_fwnode);
index dff4e17..7f10312 100644 (file)
@@ -542,7 +542,7 @@ int i2c_pca_add_numbered_bus(struct i2c_adapter *adap)
 EXPORT_SYMBOL(i2c_pca_add_numbered_bus);
 
 MODULE_AUTHOR("Ian Campbell <icampbell@arcom.com>, "
-       "Wolfram Sang <w.sang@pengutronix.de>");
+       "Wolfram Sang <kernel@pengutronix.de>");
 MODULE_DESCRIPTION("I2C-Bus PCA9564/PCA9665 algorithm");
 MODULE_LICENSE("GPL");
 
index f5c00f9..16ddc26 100644 (file)
@@ -70,6 +70,7 @@
  * @isr_mask: cached copy of local ISR enables.
  * @isr_status: cached copy of local ISR status.
  * @lock: spinlock for IRQ synchronization.
+ * @isr_mutex: mutex for IRQ thread.
  */
 struct altr_i2c_dev {
        void __iomem *base;
@@ -86,6 +87,7 @@ struct altr_i2c_dev {
        u32 isr_mask;
        u32 isr_status;
        spinlock_t lock;        /* IRQ synchronization */
+       struct mutex isr_mutex;
 };
 
 static void
@@ -245,10 +247,11 @@ static irqreturn_t altr_i2c_isr(int irq, void *_dev)
        struct altr_i2c_dev *idev = _dev;
        u32 status = idev->isr_status;
 
+       mutex_lock(&idev->isr_mutex);
        if (!idev->msg) {
                dev_warn(idev->dev, "unexpected interrupt\n");
                altr_i2c_int_clear(idev, ALTR_I2C_ALL_IRQ);
-               return IRQ_HANDLED;
+               goto out;
        }
        read = (idev->msg->flags & I2C_M_RD) != 0;
 
@@ -301,6 +304,8 @@ static irqreturn_t altr_i2c_isr(int irq, void *_dev)
                complete(&idev->msg_complete);
                dev_dbg(idev->dev, "Message Complete\n");
        }
+out:
+       mutex_unlock(&idev->isr_mutex);
 
        return IRQ_HANDLED;
 }
@@ -312,6 +317,7 @@ static int altr_i2c_xfer_msg(struct altr_i2c_dev *idev, struct i2c_msg *msg)
        u32 value;
        u8 addr = i2c_8bit_addr_from_msg(msg);
 
+       mutex_lock(&idev->isr_mutex);
        idev->msg = msg;
        idev->msg_len = msg->len;
        idev->buf = msg->buf;
@@ -336,6 +342,7 @@ static int altr_i2c_xfer_msg(struct altr_i2c_dev *idev, struct i2c_msg *msg)
                altr_i2c_int_enable(idev, imask, true);
                altr_i2c_fill_tx_fifo(idev);
        }
+       mutex_unlock(&idev->isr_mutex);
 
        time_left = wait_for_completion_timeout(&idev->msg_complete,
                                                ALTR_I2C_XFER_TIMEOUT);
@@ -409,6 +416,7 @@ static int altr_i2c_probe(struct platform_device *pdev)
        idev->dev = &pdev->dev;
        init_completion(&idev->msg_complete);
        spin_lock_init(&idev->lock);
+       mutex_init(&idev->isr_mutex);
 
        ret = device_property_read_u32(idev->dev, "fifo-size",
                                       &idev->fifo_size);
index 0aba51a..37b96ac 100644 (file)
@@ -845,6 +845,18 @@ static int at91_init_twi_recovery_info(struct platform_device *pdev,
                                                         PINCTRL_STATE_DEFAULT);
        dev->pinctrl_pins_gpio = pinctrl_lookup_state(dev->pinctrl,
                                                      "gpio");
+       if (IS_ERR(dev->pinctrl_pins_default) ||
+           IS_ERR(dev->pinctrl_pins_gpio)) {
+               dev_info(&pdev->dev, "pinctrl states incomplete for recovery\n");
+               return -EINVAL;
+       }
+
+       /*
+        * pins will be taken as GPIO, so we might as well inform pinctrl about
+        * this and move the state to GPIO
+        */
+       pinctrl_select_state(dev->pinctrl, dev->pinctrl_pins_gpio);
+
        rinfo->sda_gpiod = devm_gpiod_get(&pdev->dev, "sda", GPIOD_IN);
        if (PTR_ERR(rinfo->sda_gpiod) == -EPROBE_DEFER)
                return -EPROBE_DEFER;
@@ -855,9 +867,7 @@ static int at91_init_twi_recovery_info(struct platform_device *pdev,
                return -EPROBE_DEFER;
 
        if (IS_ERR(rinfo->sda_gpiod) ||
-           IS_ERR(rinfo->scl_gpiod) ||
-           IS_ERR(dev->pinctrl_pins_default) ||
-           IS_ERR(dev->pinctrl_pins_gpio)) {
+           IS_ERR(rinfo->scl_gpiod)) {
                dev_info(&pdev->dev, "recovery information incomplete\n");
                if (!IS_ERR(rinfo->sda_gpiod)) {
                        gpiod_put(rinfo->sda_gpiod);
@@ -867,9 +877,13 @@ static int at91_init_twi_recovery_info(struct platform_device *pdev,
                        gpiod_put(rinfo->scl_gpiod);
                        rinfo->scl_gpiod = NULL;
                }
+               pinctrl_select_state(dev->pinctrl, dev->pinctrl_pins_default);
                return -EINVAL;
        }
 
+       /* change the state of the pins back to their default state */
+       pinctrl_select_state(dev->pinctrl, dev->pinctrl_pins_default);
+
        dev_info(&pdev->dev, "using scl, sda for recovery\n");
 
        rinfo->prepare_recovery = at91_prepare_twi_recovery;
index a669127..1f1442d 100644 (file)
@@ -7,7 +7,7 @@
  *   Mux support by Rodolfo Giometti <giometti@enneenne.com> and
  *   Michael Lawnick <michael.lawnick.ext@nsn.com>
  *
- * Copyright (C) 2013-2017 Wolfram Sang <wsa@the-dreams.de>
+ * Copyright (C) 2013-2017 Wolfram Sang <wsa@kernel.org>
  */
 
 #define pr_fmt(fmt) "i2c-core: " fmt
@@ -338,8 +338,10 @@ static int i2c_device_probe(struct device *dev)
                } else if (ACPI_COMPANION(dev)) {
                        irq = i2c_acpi_get_irq(client);
                }
-               if (irq == -EPROBE_DEFER)
-                       return irq;
+               if (irq == -EPROBE_DEFER) {
+                       status = irq;
+                       goto put_sync_adapter;
+               }
 
                if (irq < 0)
                        irq = 0;
@@ -353,15 +355,19 @@ static int i2c_device_probe(struct device *dev)
         */
        if (!driver->id_table &&
            !i2c_acpi_match_device(dev->driver->acpi_match_table, client) &&
-           !i2c_of_match_device(dev->driver->of_match_table, client))
-               return -ENODEV;
+           !i2c_of_match_device(dev->driver->of_match_table, client)) {
+               status = -ENODEV;
+               goto put_sync_adapter;
+       }
 
        if (client->flags & I2C_CLIENT_WAKE) {
                int wakeirq;
 
                wakeirq = of_irq_get_byname(dev->of_node, "wakeup");
-               if (wakeirq == -EPROBE_DEFER)
-                       return wakeirq;
+               if (wakeirq == -EPROBE_DEFER) {
+                       status = wakeirq;
+                       goto put_sync_adapter;
+               }
 
                device_init_wakeup(&client->dev, true);
 
@@ -408,6 +414,10 @@ err_detach_pm_domain:
 err_clear_wakeup_irq:
        dev_pm_clear_wake_irq(&client->dev);
        device_init_wakeup(&client->dev, false);
+put_sync_adapter:
+       if (client->flags & I2C_CLIENT_HOST_NOTIFY)
+               pm_runtime_put_sync(&client->adapter->dev);
+
        return status;
 }
 
index 6787c1f..3ed74aa 100644 (file)
@@ -5,7 +5,7 @@
  * Copyright (C) 2008 Jochen Friedrich <jochen@scram.de>
  * based on a previous patch from Jon Smirl <jonsmirl@gmail.com>
  *
- * Copyright (C) 2013, 2018 Wolfram Sang <wsa@the-dreams.de>
+ * Copyright (C) 2013, 2018 Wolfram Sang <wsa@kernel.org>
  */
 
 #include <dt-bindings/i2c/i2c.h>
index 0e16490..5365199 100644 (file)
@@ -272,6 +272,7 @@ static int i2c_demux_pinctrl_probe(struct platform_device *pdev)
 err_rollback_available:
        device_remove_file(&pdev->dev, &dev_attr_available_masters);
 err_rollback:
+       i2c_demux_deactivate_master(priv);
        for (j = 0; j < i; j++) {
                of_node_put(priv->chan[j].parent_np);
                of_changeset_destroy(&priv->chan[j].chgset);
index 66d768d..6e42907 100644 (file)
@@ -980,7 +980,7 @@ static int sca3000_read_data(struct sca3000_state *st,
        st->tx[0] = SCA3000_READ_REG(reg_address_high);
        ret = spi_sync_transfer(st->us, xfer, ARRAY_SIZE(xfer));
        if (ret) {
-               dev_err(get_device(&st->us->dev), "problem reading register");
+               dev_err(&st->us->dev, "problem reading register\n");
                return ret;
        }
 
index ae622ee..dfc3a30 100644 (file)
@@ -1812,18 +1812,18 @@ static int stm32_adc_chan_of_init(struct iio_dev *indio_dev)
        return 0;
 }
 
-static int stm32_adc_dma_request(struct iio_dev *indio_dev)
+static int stm32_adc_dma_request(struct device *dev, struct iio_dev *indio_dev)
 {
        struct stm32_adc *adc = iio_priv(indio_dev);
        struct dma_slave_config config;
        int ret;
 
-       adc->dma_chan = dma_request_chan(&indio_dev->dev, "rx");
+       adc->dma_chan = dma_request_chan(dev, "rx");
        if (IS_ERR(adc->dma_chan)) {
                ret = PTR_ERR(adc->dma_chan);
                if (ret != -ENODEV) {
                        if (ret != -EPROBE_DEFER)
-                               dev_err(&indio_dev->dev,
+                               dev_err(dev,
                                        "DMA channel request failed with %d\n",
                                        ret);
                        return ret;
@@ -1930,7 +1930,7 @@ static int stm32_adc_probe(struct platform_device *pdev)
        if (ret < 0)
                return ret;
 
-       ret = stm32_adc_dma_request(indio_dev);
+       ret = stm32_adc_dma_request(dev, indio_dev);
        if (ret < 0)
                return ret;
 
index 76a60d9..506bf51 100644 (file)
@@ -62,7 +62,7 @@ enum sd_converter_type {
 
 struct stm32_dfsdm_dev_data {
        int type;
-       int (*init)(struct iio_dev *indio_dev);
+       int (*init)(struct device *dev, struct iio_dev *indio_dev);
        unsigned int num_channels;
        const struct regmap_config *regmap_cfg;
 };
@@ -1365,11 +1365,12 @@ static void stm32_dfsdm_dma_release(struct iio_dev *indio_dev)
        }
 }
 
-static int stm32_dfsdm_dma_request(struct iio_dev *indio_dev)
+static int stm32_dfsdm_dma_request(struct device *dev,
+                                  struct iio_dev *indio_dev)
 {
        struct stm32_dfsdm_adc *adc = iio_priv(indio_dev);
 
-       adc->dma_chan = dma_request_chan(&indio_dev->dev, "rx");
+       adc->dma_chan = dma_request_chan(dev, "rx");
        if (IS_ERR(adc->dma_chan)) {
                int ret = PTR_ERR(adc->dma_chan);
 
@@ -1425,7 +1426,7 @@ static int stm32_dfsdm_adc_chan_init_one(struct iio_dev *indio_dev,
                                          &adc->dfsdm->ch_list[ch->channel]);
 }
 
-static int stm32_dfsdm_audio_init(struct iio_dev *indio_dev)
+static int stm32_dfsdm_audio_init(struct device *dev, struct iio_dev *indio_dev)
 {
        struct iio_chan_spec *ch;
        struct stm32_dfsdm_adc *adc = iio_priv(indio_dev);
@@ -1452,10 +1453,10 @@ static int stm32_dfsdm_audio_init(struct iio_dev *indio_dev)
        indio_dev->num_channels = 1;
        indio_dev->channels = ch;
 
-       return stm32_dfsdm_dma_request(indio_dev);
+       return stm32_dfsdm_dma_request(dev, indio_dev);
 }
 
-static int stm32_dfsdm_adc_init(struct iio_dev *indio_dev)
+static int stm32_dfsdm_adc_init(struct device *dev, struct iio_dev *indio_dev)
 {
        struct iio_chan_spec *ch;
        struct stm32_dfsdm_adc *adc = iio_priv(indio_dev);
@@ -1499,17 +1500,17 @@ static int stm32_dfsdm_adc_init(struct iio_dev *indio_dev)
        init_completion(&adc->completion);
 
        /* Optionally request DMA */
-       ret = stm32_dfsdm_dma_request(indio_dev);
+       ret = stm32_dfsdm_dma_request(dev, indio_dev);
        if (ret) {
                if (ret != -ENODEV) {
                        if (ret != -EPROBE_DEFER)
-                               dev_err(&indio_dev->dev,
+                               dev_err(dev,
                                        "DMA channel request failed with %d\n",
                                        ret);
                        return ret;
                }
 
-               dev_dbg(&indio_dev->dev, "No DMA support\n");
+               dev_dbg(dev, "No DMA support\n");
                return 0;
        }
 
@@ -1622,7 +1623,7 @@ static int stm32_dfsdm_adc_probe(struct platform_device *pdev)
                adc->dfsdm->fl_list[adc->fl_id].sync_mode = val;
 
        adc->dev_data = dev_data;
-       ret = dev_data->init(iio);
+       ret = dev_data->init(dev, iio);
        if (ret < 0)
                return ret;
 
index abe4b56..8a87920 100644 (file)
@@ -32,16 +32,17 @@ struct ads8344 {
        u8 rx_buf[3];
 };
 
-#define ADS8344_VOLTAGE_CHANNEL(chan, si)                              \
+#define ADS8344_VOLTAGE_CHANNEL(chan, addr)                            \
        {                                                               \
                .type = IIO_VOLTAGE,                                    \
                .indexed = 1,                                           \
                .channel = chan,                                        \
                .info_mask_separate = BIT(IIO_CHAN_INFO_RAW),           \
                .info_mask_shared_by_type = BIT(IIO_CHAN_INFO_SCALE),   \
+               .address = addr,                                        \
        }
 
-#define ADS8344_VOLTAGE_CHANNEL_DIFF(chan1, chan2, si)                 \
+#define ADS8344_VOLTAGE_CHANNEL_DIFF(chan1, chan2, addr)               \
        {                                                               \
                .type = IIO_VOLTAGE,                                    \
                .indexed = 1,                                           \
@@ -50,6 +51,7 @@ struct ads8344 {
                .differential = 1,                                      \
                .info_mask_separate = BIT(IIO_CHAN_INFO_RAW),           \
                .info_mask_shared_by_type = BIT(IIO_CHAN_INFO_SCALE),   \
+               .address = addr,                                        \
        }
 
 static const struct iio_chan_spec ads8344_channels[] = {
@@ -105,7 +107,7 @@ static int ads8344_read_raw(struct iio_dev *iio,
        switch (mask) {
        case IIO_CHAN_INFO_RAW:
                mutex_lock(&adc->lock);
-               *value = ads8344_adc_conversion(adc, channel->scan_index,
+               *value = ads8344_adc_conversion(adc, channel->address,
                                                channel->differential);
                mutex_unlock(&adc->lock);
                if (*value < 0)
index 82d4705..7b199ce 100644 (file)
@@ -194,7 +194,19 @@ static const struct iio_chan_spec atlas_orp_channels[] = {
 };
 
 static const struct iio_chan_spec atlas_do_channels[] = {
-       ATLAS_CONCENTRATION_CHANNEL(0, ATLAS_REG_DO_DATA),
+       {
+               .type = IIO_CONCENTRATION,
+               .address = ATLAS_REG_DO_DATA,
+               .info_mask_separate =
+                       BIT(IIO_CHAN_INFO_RAW) | BIT(IIO_CHAN_INFO_SCALE),
+               .scan_index = 0,
+               .scan_type = {
+                       .sign = 'u',
+                       .realbits = 32,
+                       .storagebits = 32,
+                       .endianness = IIO_BE,
+               },
+       },
        IIO_CHAN_SOFT_TIMESTAMP(1),
        {
                .type = IIO_TEMP,
index 71f8a5c..7f1e931 100644 (file)
@@ -223,6 +223,7 @@ static int vf610_dac_probe(struct platform_device *pdev)
        return 0;
 
 error_iio_device_register:
+       vf610_dac_exit(info);
        clk_disable_unprepare(info->clk);
 
        return ret;
index 64ef07a..1cf9819 100644 (file)
@@ -544,8 +544,10 @@ st_lsm6dsx_shub_write_raw(struct iio_dev *iio_dev,
 
                        ref_sensor = iio_priv(hw->iio_devs[ST_LSM6DSX_ID_ACC]);
                        odr = st_lsm6dsx_check_odr(ref_sensor, val, &odr_val);
-                       if (odr < 0)
-                               return odr;
+                       if (odr < 0) {
+                               err = odr;
+                               goto release;
+                       }
 
                        sensor->ext_info.slv_odr = val;
                        sensor->odr = odr;
@@ -557,6 +559,7 @@ st_lsm6dsx_shub_write_raw(struct iio_dev *iio_dev,
                break;
        }
 
+release:
        iio_device_release_direct_mode(iio_dev);
 
        return err;
index 1dc3718..2883ac3 100644 (file)
@@ -127,7 +127,8 @@ static inline int get_acpihid_device_id(struct device *dev,
                return -ENODEV;
 
        list_for_each_entry(p, &acpihid_map, list) {
-               if (acpi_dev_hid_uid_match(adev, p->hid, p->uid)) {
+               if (acpi_dev_hid_uid_match(adev, p->hid,
+                                          p->uid[0] ? p->uid : NULL)) {
                        if (entry)
                                *entry = p;
                        return p->devid;
index 2b9a67e..5b81fd1 100644 (file)
@@ -1329,8 +1329,8 @@ static int __init init_iommu_from_acpi(struct amd_iommu *iommu,
                }
                case IVHD_DEV_ACPI_HID: {
                        u16 devid;
-                       u8 hid[ACPIHID_HID_LEN] = {0};
-                       u8 uid[ACPIHID_UID_LEN] = {0};
+                       u8 hid[ACPIHID_HID_LEN];
+                       u8 uid[ACPIHID_UID_LEN];
                        int ret;
 
                        if (h->type != 0x40) {
@@ -1347,6 +1347,7 @@ static int __init init_iommu_from_acpi(struct amd_iommu *iommu,
                                break;
                        }
 
+                       uid[0] = '\0';
                        switch (e->uidf) {
                        case UID_NOT_PRESENT:
 
@@ -1361,8 +1362,8 @@ static int __init init_iommu_from_acpi(struct amd_iommu *iommu,
                                break;
                        case UID_IS_CHARACTER:
 
-                               memcpy(uid, (u8 *)(&e->uid), ACPIHID_UID_LEN - 1);
-                               uid[ACPIHID_UID_LEN - 1] = '\0';
+                               memcpy(uid, &e->uid, e->uidl);
+                               uid[e->uidl] = '\0';
 
                                break;
                        default:
index 7b37542..1faa08c 100644 (file)
@@ -693,6 +693,15 @@ out:
        return ret;
 }
 
+static bool iommu_is_attach_deferred(struct iommu_domain *domain,
+                                    struct device *dev)
+{
+       if (domain->ops->is_attach_deferred)
+               return domain->ops->is_attach_deferred(domain, dev);
+
+       return false;
+}
+
 /**
  * iommu_group_add_device - add a device to an iommu group
  * @group: the group into which to add the device (reference should be held)
@@ -747,7 +756,7 @@ rename:
 
        mutex_lock(&group->mutex);
        list_add_tail(&device->list, &group->devices);
-       if (group->domain)
+       if (group->domain  && !iommu_is_attach_deferred(group->domain, dev))
                ret = __iommu_attach_device(group->domain, dev);
        mutex_unlock(&group->mutex);
        if (ret)
@@ -1653,9 +1662,6 @@ static int __iommu_attach_device(struct iommu_domain *domain,
                                 struct device *dev)
 {
        int ret;
-       if ((domain->ops->is_attach_deferred != NULL) &&
-           domain->ops->is_attach_deferred(domain, dev))
-               return 0;
 
        if (unlikely(domain->ops->attach_dev == NULL))
                return -ENODEV;
@@ -1727,8 +1733,7 @@ EXPORT_SYMBOL_GPL(iommu_sva_unbind_gpasid);
 static void __iommu_detach_device(struct iommu_domain *domain,
                                  struct device *dev)
 {
-       if ((domain->ops->is_attach_deferred != NULL) &&
-           domain->ops->is_attach_deferred(domain, dev))
+       if (iommu_is_attach_deferred(domain, dev))
                return;
 
        if (unlikely(domain->ops->detach_dev == NULL))
index 23445eb..ec71063 100644 (file)
@@ -306,6 +306,7 @@ static int tpci200_register(struct tpci200_board *tpci200)
                        "(bn 0x%X, sn 0x%X) failed to map driver user space!",
                        tpci200->info->pdev->bus->number,
                        tpci200->info->pdev->devfn);
+               res = -ENOMEM;
                goto out_release_mem8_space;
        }
 
index 06038b3..55da642 100644 (file)
@@ -142,6 +142,9 @@ static void rtsx_comm_pm_full_on(struct rtsx_pcr *pcr)
 
        rtsx_disable_aspm(pcr);
 
+       /* Fixes DMA transfer timout issue after disabling ASPM on RTS5260 */
+       msleep(1);
+
        if (option->ltr_enabled)
                rtsx_set_ltr_latency(pcr, option->ltr_active_latency);
 
index 204d807..b32c825 100644 (file)
@@ -266,6 +266,7 @@ void mei_me_cl_rm_by_uuid(struct mei_device *dev, const uuid_le *uuid)
        down_write(&dev->me_clients_rwsem);
        me_cl = __mei_me_cl_by_uuid(dev, uuid);
        __mei_me_cl_del(dev, me_cl);
+       mei_me_cl_put(me_cl);
        up_write(&dev->me_clients_rwsem);
 }
 
@@ -287,6 +288,7 @@ void mei_me_cl_rm_by_uuid_id(struct mei_device *dev, const uuid_le *uuid, u8 id)
        down_write(&dev->me_clients_rwsem);
        me_cl = __mei_me_cl_by_uuid_id(dev, uuid, id);
        __mei_me_cl_del(dev, me_cl);
+       mei_me_cl_put(me_cl);
        up_write(&dev->me_clients_rwsem);
 }
 
index 2916674..29d4100 100644 (file)
@@ -555,7 +555,7 @@ static int mtd_nvmem_add(struct mtd_info *mtd)
 
        config.id = -1;
        config.dev = &mtd->dev;
-       config.name = mtd->name;
+       config.name = dev_name(&mtd->dev);
        config.owner = THIS_MODULE;
        config.reg_read = mtd_nvmem_reg_read;
        config.size = mtd->size;
index e4e3cee..8f9ffb4 100644 (file)
@@ -2728,9 +2728,8 @@ static int brcmnand_resume(struct device *dev)
                flash_dma_writel(ctrl, FLASH_DMA_ERROR_STATUS, 0);
        }
 
-       if (has_edu(ctrl))
+       if (has_edu(ctrl)) {
                ctrl->edu_config = edu_readl(ctrl, EDU_CONFIG);
-       else {
                edu_writel(ctrl, EDU_CONFIG, ctrl->edu_config);
                edu_readl(ctrl, EDU_CONFIG);
                brcmnand_edu_init(ctrl);
index b6bb358..e2c382f 100644 (file)
@@ -1089,6 +1089,10 @@ static int spinand_init(struct spinand_device *spinand)
 
        mtd->oobavail = ret;
 
+       /* Propagate ECC information to mtd_info */
+       mtd->ecc_strength = nand->eccreq.strength;
+       mtd->ecc_step_size = nand->eccreq.step_size;
+
        return 0;
 
 err_cleanup_nanddev:
index 54646c2..ac2bdba 100644 (file)
@@ -393,9 +393,6 @@ static void *eraseblk_count_seq_start(struct seq_file *s, loff_t *pos)
 {
        struct ubi_device *ubi = s->private;
 
-       if (*pos == 0)
-               return SEQ_START_TOKEN;
-
        if (*pos < ubi->peb_count)
                return pos;
 
@@ -409,8 +406,6 @@ static void *eraseblk_count_seq_next(struct seq_file *s, void *v, loff_t *pos)
 {
        struct ubi_device *ubi = s->private;
 
-       if (v == SEQ_START_TOKEN)
-               return pos;
        (*pos)++;
 
        if (*pos < ubi->peb_count)
@@ -432,11 +427,8 @@ static int eraseblk_count_seq_show(struct seq_file *s, void *iter)
        int err;
 
        /* If this is the start, print a header */
-       if (iter == SEQ_START_TOKEN) {
-               seq_puts(s,
-                        "physical_block_number\terase_count\tblock_status\tread_status\n");
-               return 0;
-       }
+       if (*block_number == 0)
+               seq_puts(s, "physical_block_number\terase_count\n");
 
        err = ubi_io_is_bad(ubi, *block_number);
        if (err)
index 04d59be..74503ca 100644 (file)
@@ -947,8 +947,11 @@ static int ifi_canfd_plat_probe(struct platform_device *pdev)
        u32 id, rev;
 
        addr = devm_platform_ioremap_resource(pdev, 0);
+       if (IS_ERR(addr))
+               return PTR_ERR(addr);
+
        irq = platform_get_irq(pdev, 0);
-       if (IS_ERR(addr) || irq < 0)
+       if (irq < 0)
                return -EINVAL;
 
        id = readl(addr + IFI_CANFD_IP_ID);
index e3ba8ab..e2c6cf4 100644 (file)
@@ -792,7 +792,7 @@ static int sun4ican_probe(struct platform_device *pdev)
 
        addr = devm_platform_ioremap_resource(pdev, 0);
        if (IS_ERR(addr)) {
-               err = -EBUSY;
+               err = PTR_ERR(addr);
                goto exit;
        }
 
index 0a1be52..38cd828 100644 (file)
@@ -609,7 +609,7 @@ static int b53_srab_probe(struct platform_device *pdev)
 
        priv->regs = devm_platform_ioremap_resource(pdev, 0);
        if (IS_ERR(priv->regs))
-               return -ENOMEM;
+               return PTR_ERR(priv->regs);
 
        dev = b53_switch_alloc(&pdev->dev, &b53_srab_ops, priv);
        if (!dev)
index 5c444cd..34e4aad 100644 (file)
@@ -628,11 +628,8 @@ mt7530_cpu_port_enable(struct mt7530_priv *priv,
        mt7530_write(priv, MT7530_PVC_P(port),
                     PORT_SPEC_TAG);
 
-       /* Disable auto learning on the cpu port */
-       mt7530_set(priv, MT7530_PSC_P(port), SA_DIS);
-
-       /* Unknown unicast frame fordwarding to the cpu port */
-       mt7530_set(priv, MT7530_MFC, UNU_FFP(BIT(port)));
+       /* Unknown multicast frame forwarding to the cpu port */
+       mt7530_rmw(priv, MT7530_MFC, UNM_FFP_MASK, UNM_FFP(BIT(port)));
 
        /* Set CPU port number */
        if (priv->id == ID_MT7621)
@@ -1294,8 +1291,6 @@ mt7530_setup(struct dsa_switch *ds)
        /* Enable and reset MIB counters */
        mt7530_mib_reset(ds);
 
-       mt7530_clear(priv, MT7530_MFC, UNU_FFP_MASK);
-
        for (i = 0; i < MT7530_NUM_PORTS; i++) {
                /* Disable forwarding by default on all ports */
                mt7530_rmw(priv, MT7530_PCR_P(i), PCR_MATRIX_MASK,
index 979bb63..82af4d2 100644 (file)
@@ -31,6 +31,7 @@ enum {
 #define MT7530_MFC                     0x10
 #define  BC_FFP(x)                     (((x) & 0xff) << 24)
 #define  UNM_FFP(x)                    (((x) & 0xff) << 16)
+#define  UNM_FFP_MASK                  UNM_FFP(~0)
 #define  UNU_FFP(x)                    (((x) & 0xff) << 8)
 #define  UNU_FFP_MASK                  UNU_FFP(~0)
 #define  CPU_EN                                BIT(7)
index e2c6bf0..e8aae64 100644 (file)
@@ -388,6 +388,7 @@ static int felix_init_structs(struct felix *felix, int num_phys_ports)
        struct ocelot *ocelot = &felix->ocelot;
        phy_interface_t *port_phy_modes;
        resource_size_t switch_base;
+       struct resource res;
        int port, i, err;
 
        ocelot->num_phys_ports = num_phys_ports;
@@ -422,17 +423,16 @@ static int felix_init_structs(struct felix *felix, int num_phys_ports)
 
        for (i = 0; i < TARGET_MAX; i++) {
                struct regmap *target;
-               struct resource *res;
 
                if (!felix->info->target_io_res[i].name)
                        continue;
 
-               res = &felix->info->target_io_res[i];
-               res->flags = IORESOURCE_MEM;
-               res->start += switch_base;
-               res->end += switch_base;
+               memcpy(&res, &felix->info->target_io_res[i], sizeof(res));
+               res.flags = IORESOURCE_MEM;
+               res.start += switch_base;
+               res.end += switch_base;
 
-               target = ocelot_regmap_init(ocelot, res);
+               target = ocelot_regmap_init(ocelot, &res);
                if (IS_ERR(target)) {
                        dev_err(ocelot->dev,
                                "Failed to map device memory space\n");
@@ -453,7 +453,6 @@ static int felix_init_structs(struct felix *felix, int num_phys_ports)
        for (port = 0; port < num_phys_ports; port++) {
                struct ocelot_port *ocelot_port;
                void __iomem *port_regs;
-               struct resource *res;
 
                ocelot_port = devm_kzalloc(ocelot->dev,
                                           sizeof(struct ocelot_port),
@@ -465,12 +464,12 @@ static int felix_init_structs(struct felix *felix, int num_phys_ports)
                        return -ENOMEM;
                }
 
-               res = &felix->info->port_io_res[port];
-               res->flags = IORESOURCE_MEM;
-               res->start += switch_base;
-               res->end += switch_base;
+               memcpy(&res, &felix->info->port_io_res[port], sizeof(res));
+               res.flags = IORESOURCE_MEM;
+               res.start += switch_base;
+               res.end += switch_base;
 
-               port_regs = devm_ioremap_resource(ocelot->dev, res);
+               port_regs = devm_ioremap_resource(ocelot->dev, &res);
                if (IS_ERR(port_regs)) {
                        dev_err(ocelot->dev,
                                "failed to map registers for port %d\n", port);
index 9af1065..730a8a9 100644 (file)
@@ -8,9 +8,9 @@
 
 /* Platform-specific information */
 struct felix_info {
-       struct resource                 *target_io_res;
-       struct resource                 *port_io_res;
-       struct resource                 *imdio_res;
+       const struct resource           *target_io_res;
+       const struct resource           *port_io_res;
+       const struct resource           *imdio_res;
        const struct reg_field          *regfields;
        const u32 *const                *map;
        const struct ocelot_ops         *ops;
index 8bf395f..5211f05 100644 (file)
@@ -333,10 +333,8 @@ static const u32 *vsc9959_regmap[] = {
        [GCB]   = vsc9959_gcb_regmap,
 };
 
-/* Addresses are relative to the PCI device's base address and
- * will be fixed up at ioremap time.
- */
-static struct resource vsc9959_target_io_res[] = {
+/* Addresses are relative to the PCI device's base address */
+static const struct resource vsc9959_target_io_res[] = {
        [ANA] = {
                .start  = 0x0280000,
                .end    = 0x028ffff,
@@ -379,7 +377,7 @@ static struct resource vsc9959_target_io_res[] = {
        },
 };
 
-static struct resource vsc9959_port_io_res[] = {
+static const struct resource vsc9959_port_io_res[] = {
        {
                .start  = 0x0100000,
                .end    = 0x010ffff,
@@ -415,7 +413,7 @@ static struct resource vsc9959_port_io_res[] = {
 /* Port MAC 0 Internal MDIO bus through which the SerDes acting as an
  * SGMII/QSGMII MAC PCS can be found.
  */
-static struct resource vsc9959_imdio_res = {
+static const struct resource vsc9959_imdio_res = {
        .start          = 0x8030,
        .end            = 0x8040,
        .name           = "imdio",
@@ -1111,7 +1109,7 @@ static int vsc9959_mdio_bus_alloc(struct ocelot *ocelot)
        struct device *dev = ocelot->dev;
        resource_size_t imdio_base;
        void __iomem *imdio_regs;
-       struct resource *res;
+       struct resource res;
        struct enetc_hw *hw;
        struct mii_bus *bus;
        int port;
@@ -1128,12 +1126,12 @@ static int vsc9959_mdio_bus_alloc(struct ocelot *ocelot)
        imdio_base = pci_resource_start(felix->pdev,
                                        felix->info->imdio_pci_bar);
 
-       res = felix->info->imdio_res;
-       res->flags = IORESOURCE_MEM;
-       res->start += imdio_base;
-       res->end += imdio_base;
+       memcpy(&res, felix->info->imdio_res, sizeof(res));
+       res.flags = IORESOURCE_MEM;
+       res.start += imdio_base;
+       res.end += imdio_base;
 
-       imdio_regs = devm_ioremap_resource(dev, res);
+       imdio_regs = devm_ioremap_resource(dev, &res);
        if (IS_ERR(imdio_regs)) {
                dev_err(dev, "failed to map internal MDIO registers\n");
                return PTR_ERR(imdio_regs);
index a58185b..3e3711b 100644 (file)
@@ -1182,7 +1182,7 @@ bmac_get_station_address(struct net_device *dev, unsigned char *ea)
        int i;
        unsigned short data;
 
-       for (i = 0; i < 6; i++)
+       for (i = 0; i < 3; i++)
                {
                        reset_and_select_srom(dev);
                        data = read_srom(dev, i + EnetAddressOffset/2, SROMAddressBits);
index 6e5f6dd..552e755 100644 (file)
@@ -42,6 +42,7 @@
 #include <soc/fsl/qe/ucc.h>
 #include <soc/fsl/qe/ucc_fast.h>
 #include <asm/machdep.h>
+#include <net/sch_generic.h>
 
 #include "ucc_geth.h"
 
@@ -1548,11 +1549,8 @@ static int ugeth_disable(struct ucc_geth_private *ugeth, enum comm_dir mode)
 
 static void ugeth_quiesce(struct ucc_geth_private *ugeth)
 {
-       /* Prevent any further xmits, plus detach the device. */
-       netif_device_detach(ugeth->ndev);
-
-       /* Wait for any current xmits to finish. */
-       netif_tx_disable(ugeth->ndev);
+       /* Prevent any further xmits */
+       netif_tx_stop_all_queues(ugeth->ndev);
 
        /* Disable the interrupt to avoid NAPI rescheduling. */
        disable_irq(ugeth->ug_info->uf_info.irq);
@@ -1565,7 +1563,10 @@ static void ugeth_activate(struct ucc_geth_private *ugeth)
 {
        napi_enable(&ugeth->napi);
        enable_irq(ugeth->ug_info->uf_info.irq);
-       netif_device_attach(ugeth->ndev);
+
+       /* allow to xmit again  */
+       netif_tx_wake_all_queues(ugeth->ndev);
+       __netdev_watchdog_up(ugeth->ndev);
 }
 
 /* Called every time the controller might need to be made
index 7352244..d4a4e24 100644 (file)
@@ -1070,7 +1070,7 @@ void mvpp2_cls_oversize_rxq_set(struct mvpp2_port *port)
                    (port->first_rxq >> MVPP2_CLS_OVERSIZE_RXQ_LOW_BITS));
 
        val = mvpp2_read(port->priv, MVPP2_CLS_SWFWD_PCTRL_REG);
-       val |= MVPP2_CLS_SWFWD_PCTRL_MASK(port->id);
+       val &= ~MVPP2_CLS_SWFWD_PCTRL_MASK(port->id);
        mvpp2_write(port->priv, MVPP2_CLS_SWFWD_PCTRL_REG, val);
 }
 
index 7a0d785..17243bb 100644 (file)
@@ -1418,7 +1418,7 @@ static int pxa168_eth_probe(struct platform_device *pdev)
 
        pep->base = devm_platform_ioremap_resource(pdev, 0);
        if (IS_ERR(pep->base)) {
-               err = -ENOMEM;
+               err = PTR_ERR(pep->base);
                goto err_netdev;
        }
 
index 6e501af..f6ff962 100644 (file)
@@ -2734,7 +2734,7 @@ void mlx4_opreq_action(struct work_struct *work)
                if (err) {
                        mlx4_err(dev, "Failed to retrieve required operation: %d\n",
                                 err);
-                       return;
+                       goto out;
                }
                MLX4_GET(modifier, outbox, GET_OP_REQ_MODIFIER_OFFSET);
                MLX4_GET(token, outbox, GET_OP_REQ_TOKEN_OFFSET);
index cede5bd..7a77fe4 100644 (file)
@@ -848,6 +848,14 @@ static void free_msg(struct mlx5_core_dev *dev, struct mlx5_cmd_msg *msg);
 static void mlx5_free_cmd_msg(struct mlx5_core_dev *dev,
                              struct mlx5_cmd_msg *msg);
 
+static bool opcode_allowed(struct mlx5_cmd *cmd, u16 opcode)
+{
+       if (cmd->allowed_opcode == CMD_ALLOWED_OPCODE_ALL)
+               return true;
+
+       return cmd->allowed_opcode == opcode;
+}
+
 static void cmd_work_handler(struct work_struct *work)
 {
        struct mlx5_cmd_work_ent *ent = container_of(work, struct mlx5_cmd_work_ent, work);
@@ -861,6 +869,7 @@ static void cmd_work_handler(struct work_struct *work)
        int alloc_ret;
        int cmd_mode;
 
+       complete(&ent->handling);
        sem = ent->page_queue ? &cmd->pages_sem : &cmd->sem;
        down(sem);
        if (!ent->page_queue) {
@@ -913,7 +922,9 @@ static void cmd_work_handler(struct work_struct *work)
 
        /* Skip sending command to fw if internal error */
        if (pci_channel_offline(dev->pdev) ||
-           dev->state == MLX5_DEVICE_STATE_INTERNAL_ERROR) {
+           dev->state == MLX5_DEVICE_STATE_INTERNAL_ERROR ||
+           cmd->state != MLX5_CMDIF_STATE_UP ||
+           !opcode_allowed(&dev->cmd, ent->op)) {
                u8 status = 0;
                u32 drv_synd;
 
@@ -978,6 +989,11 @@ static int wait_func(struct mlx5_core_dev *dev, struct mlx5_cmd_work_ent *ent)
        struct mlx5_cmd *cmd = &dev->cmd;
        int err;
 
+       if (!wait_for_completion_timeout(&ent->handling, timeout) &&
+           cancel_work_sync(&ent->work)) {
+               ent->ret = -ECANCELED;
+               goto out_err;
+       }
        if (cmd->mode == CMD_MODE_POLLING || ent->polling) {
                wait_for_completion(&ent->done);
        } else if (!wait_for_completion_timeout(&ent->done, timeout)) {
@@ -985,12 +1001,17 @@ static int wait_func(struct mlx5_core_dev *dev, struct mlx5_cmd_work_ent *ent)
                mlx5_cmd_comp_handler(dev, 1UL << ent->idx, true);
        }
 
+out_err:
        err = ent->ret;
 
        if (err == -ETIMEDOUT) {
                mlx5_core_warn(dev, "%s(0x%x) timeout. Will cause a leak of a command resource\n",
                               mlx5_command_str(msg_to_opcode(ent->in)),
                               msg_to_opcode(ent->in));
+       } else if (err == -ECANCELED) {
+               mlx5_core_warn(dev, "%s(0x%x) canceled on out of queue timeout.\n",
+                              mlx5_command_str(msg_to_opcode(ent->in)),
+                              msg_to_opcode(ent->in));
        }
        mlx5_core_dbg(dev, "err %d, delivery status %s(%d)\n",
                      err, deliv_status_to_str(ent->status), ent->status);
@@ -1026,6 +1047,7 @@ static int mlx5_cmd_invoke(struct mlx5_core_dev *dev, struct mlx5_cmd_msg *in,
        ent->token = token;
        ent->polling = force_polling;
 
+       init_completion(&ent->handling);
        if (!callback)
                init_completion(&ent->done);
 
@@ -1045,6 +1067,8 @@ static int mlx5_cmd_invoke(struct mlx5_core_dev *dev, struct mlx5_cmd_msg *in,
        err = wait_func(dev, ent);
        if (err == -ETIMEDOUT)
                goto out;
+       if (err == -ECANCELED)
+               goto out_free;
 
        ds = ent->ts2 - ent->ts1;
        op = MLX5_GET(mbox_in, in->first.data, opcode);
@@ -1391,6 +1415,22 @@ static void create_debugfs_files(struct mlx5_core_dev *dev)
        mlx5_cmdif_debugfs_init(dev);
 }
 
+void mlx5_cmd_allowed_opcode(struct mlx5_core_dev *dev, u16 opcode)
+{
+       struct mlx5_cmd *cmd = &dev->cmd;
+       int i;
+
+       for (i = 0; i < cmd->max_reg_cmds; i++)
+               down(&cmd->sem);
+       down(&cmd->pages_sem);
+
+       cmd->allowed_opcode = opcode;
+
+       up(&cmd->pages_sem);
+       for (i = 0; i < cmd->max_reg_cmds; i++)
+               up(&cmd->sem);
+}
+
 static void mlx5_cmd_change_mod(struct mlx5_core_dev *dev, int mode)
 {
        struct mlx5_cmd *cmd = &dev->cmd;
@@ -1667,12 +1707,14 @@ static int cmd_exec(struct mlx5_core_dev *dev, void *in, int in_size, void *out,
        int err;
        u8 status = 0;
        u32 drv_synd;
+       u16 opcode;
        u8 token;
 
+       opcode = MLX5_GET(mbox_in, in, opcode);
        if (pci_channel_offline(dev->pdev) ||
-           dev->state == MLX5_DEVICE_STATE_INTERNAL_ERROR) {
-               u16 opcode = MLX5_GET(mbox_in, in, opcode);
-
+           dev->state == MLX5_DEVICE_STATE_INTERNAL_ERROR ||
+           dev->cmd.state != MLX5_CMDIF_STATE_UP ||
+           !opcode_allowed(&dev->cmd, opcode)) {
                err = mlx5_internal_err_ret_value(dev, opcode, &drv_synd, &status);
                MLX5_SET(mbox_out, out, status, status);
                MLX5_SET(mbox_out, out, syndrome, drv_synd);
@@ -1937,6 +1979,7 @@ int mlx5_cmd_init(struct mlx5_core_dev *dev)
                goto err_free_page;
        }
 
+       cmd->state = MLX5_CMDIF_STATE_DOWN;
        cmd->checksum_disabled = 1;
        cmd->max_reg_cmds = (1 << cmd->log_sz) - 1;
        cmd->bitmask = (1UL << cmd->max_reg_cmds) - 1;
@@ -1974,6 +2017,7 @@ int mlx5_cmd_init(struct mlx5_core_dev *dev)
        mlx5_core_dbg(dev, "descriptor at dma 0x%llx\n", (unsigned long long)(cmd->dma));
 
        cmd->mode = CMD_MODE_POLLING;
+       cmd->allowed_opcode = CMD_ALLOWED_OPCODE_ALL;
 
        create_msg_cache(dev);
 
@@ -2013,3 +2057,10 @@ void mlx5_cmd_cleanup(struct mlx5_core_dev *dev)
        dma_pool_destroy(cmd->pool);
 }
 EXPORT_SYMBOL(mlx5_cmd_cleanup);
+
+void mlx5_cmd_set_state(struct mlx5_core_dev *dev,
+                       enum mlx5_cmdif_state cmdif_state)
+{
+       dev->cmd.state = cmdif_state;
+}
+EXPORT_SYMBOL(mlx5_cmd_set_state);
index 23701c0..5974540 100644 (file)
@@ -1121,7 +1121,7 @@ void mlx5e_close_drop_rq(struct mlx5e_rq *drop_rq);
 int mlx5e_create_indirect_rqt(struct mlx5e_priv *priv);
 
 int mlx5e_create_indirect_tirs(struct mlx5e_priv *priv, bool inner_ttc);
-void mlx5e_destroy_indirect_tirs(struct mlx5e_priv *priv, bool inner_ttc);
+void mlx5e_destroy_indirect_tirs(struct mlx5e_priv *priv);
 
 int mlx5e_create_direct_rqts(struct mlx5e_priv *priv, struct mlx5e_tir *tirs);
 void mlx5e_destroy_direct_rqts(struct mlx5e_priv *priv, struct mlx5e_tir *tirs);
index a172c5e..4eb305a 100644 (file)
@@ -699,6 +699,7 @@ mlx5_tc_ct_parse_match(struct mlx5e_priv *priv,
                       struct netlink_ext_ack *extack)
 {
        struct mlx5_tc_ct_priv *ct_priv = mlx5_tc_ct_get_ct_priv(priv);
+       struct flow_rule *rule = flow_cls_offload_flow_rule(f);
        struct flow_dissector_key_ct *mask, *key;
        bool trk, est, untrk, unest, new;
        u32 ctstate = 0, ctstate_mask = 0;
@@ -706,7 +707,7 @@ mlx5_tc_ct_parse_match(struct mlx5e_priv *priv,
        u16 ct_state, ct_state_mask;
        struct flow_match_ct match;
 
-       if (!flow_rule_match_key(f->rule, FLOW_DISSECTOR_KEY_CT))
+       if (!flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_CT))
                return 0;
 
        if (!ct_priv) {
@@ -715,7 +716,7 @@ mlx5_tc_ct_parse_match(struct mlx5e_priv *priv,
                return -EOPNOTSUPP;
        }
 
-       flow_rule_match_ct(f->rule, &match);
+       flow_rule_match_ct(rule, &match);
 
        key = match.key;
        mask = match.mask;
index 091d305..626f6c0 100644 (file)
@@ -130,7 +130,9 @@ mlx5_tc_ct_parse_match(struct mlx5e_priv *priv,
                       struct flow_cls_offload *f,
                       struct netlink_ext_ack *extack)
 {
-       if (!flow_rule_match_key(f->rule, FLOW_DISSECTOR_KEY_CT))
+       struct flow_rule *rule = flow_cls_offload_flow_rule(f);
+
+       if (!flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_CT))
                return 0;
 
        NL_SET_ERR_MSG_MOD(extack, "mlx5 tc ct offload isn't enabled.");
index 46725cd..7d1985f 100644 (file)
@@ -69,8 +69,8 @@ static void mlx5e_ktls_del(struct net_device *netdev,
        struct mlx5e_ktls_offload_context_tx *tx_priv =
                mlx5e_get_ktls_tx_priv_ctx(tls_ctx);
 
-       mlx5_ktls_destroy_key(priv->mdev, tx_priv->key_id);
        mlx5e_destroy_tis(priv->mdev, tx_priv->tisn);
+       mlx5_ktls_destroy_key(priv->mdev, tx_priv->key_id);
        kvfree(tx_priv);
 }
 
index b314adf..c6b8304 100644 (file)
@@ -2717,7 +2717,8 @@ void mlx5e_modify_tirs_hash(struct mlx5e_priv *priv, void *in, int inlen)
                mlx5_core_modify_tir(mdev, priv->indir_tir[tt].tirn, in, inlen);
        }
 
-       if (!mlx5e_tunnel_inner_ft_supported(priv->mdev))
+       /* Verify inner tirs resources allocated */
+       if (!priv->inner_indir_tir[0].tirn)
                return;
 
        for (tt = 0; tt < MLX5E_NUM_INDIR_TIRS; tt++) {
@@ -3408,14 +3409,15 @@ out:
        return err;
 }
 
-void mlx5e_destroy_indirect_tirs(struct mlx5e_priv *priv, bool inner_ttc)
+void mlx5e_destroy_indirect_tirs(struct mlx5e_priv *priv)
 {
        int i;
 
        for (i = 0; i < MLX5E_NUM_INDIR_TIRS; i++)
                mlx5e_destroy_tir(priv->mdev, &priv->indir_tir[i]);
 
-       if (!inner_ttc || !mlx5e_tunnel_inner_ft_supported(priv->mdev))
+       /* Verify inner tirs resources allocated */
+       if (!priv->inner_indir_tir[0].tirn)
                return;
 
        for (i = 0; i < MLX5E_NUM_INDIR_TIRS; i++)
@@ -5123,7 +5125,7 @@ err_destroy_xsk_rqts:
 err_destroy_direct_tirs:
        mlx5e_destroy_direct_tirs(priv, priv->direct_tir);
 err_destroy_indirect_tirs:
-       mlx5e_destroy_indirect_tirs(priv, true);
+       mlx5e_destroy_indirect_tirs(priv);
 err_destroy_direct_rqts:
        mlx5e_destroy_direct_rqts(priv, priv->direct_tir);
 err_destroy_indirect_rqts:
@@ -5142,7 +5144,7 @@ static void mlx5e_cleanup_nic_rx(struct mlx5e_priv *priv)
        mlx5e_destroy_direct_tirs(priv, priv->xsk_tir);
        mlx5e_destroy_direct_rqts(priv, priv->xsk_tir);
        mlx5e_destroy_direct_tirs(priv, priv->direct_tir);
-       mlx5e_destroy_indirect_tirs(priv, true);
+       mlx5e_destroy_indirect_tirs(priv);
        mlx5e_destroy_direct_rqts(priv, priv->direct_tir);
        mlx5e_destroy_rqt(priv, &priv->indir_rqt);
        mlx5e_close_drop_rq(&priv->drop_rq);
index f372e94..4a8e0df 100644 (file)
@@ -1484,13 +1484,9 @@ bool mlx5e_eswitch_uplink_rep(struct net_device *netdev)
        return netdev->netdev_ops == &mlx5e_netdev_ops_uplink_rep;
 }
 
-bool mlx5e_eswitch_rep(struct net_device *netdev)
+bool mlx5e_eswitch_vf_rep(struct net_device *netdev)
 {
-       if (netdev->netdev_ops == &mlx5e_netdev_ops_rep ||
-           netdev->netdev_ops == &mlx5e_netdev_ops_uplink_rep)
-               return true;
-
-       return false;
+       return netdev->netdev_ops == &mlx5e_netdev_ops_rep;
 }
 
 static void mlx5e_build_rep_params(struct net_device *netdev)
@@ -1747,7 +1743,7 @@ err_destroy_ttc_table:
 err_destroy_direct_tirs:
        mlx5e_destroy_direct_tirs(priv, priv->direct_tir);
 err_destroy_indirect_tirs:
-       mlx5e_destroy_indirect_tirs(priv, false);
+       mlx5e_destroy_indirect_tirs(priv);
 err_destroy_direct_rqts:
        mlx5e_destroy_direct_rqts(priv, priv->direct_tir);
 err_destroy_indirect_rqts:
@@ -1765,7 +1761,7 @@ static void mlx5e_cleanup_rep_rx(struct mlx5e_priv *priv)
        mlx5e_destroy_rep_root_ft(priv);
        mlx5e_destroy_ttc_table(priv, &priv->fs.ttc);
        mlx5e_destroy_direct_tirs(priv, priv->direct_tir);
-       mlx5e_destroy_indirect_tirs(priv, false);
+       mlx5e_destroy_indirect_tirs(priv);
        mlx5e_destroy_direct_rqts(priv, priv->direct_tir);
        mlx5e_destroy_rqt(priv, &priv->indir_rqt);
        mlx5e_close_drop_rq(&priv->drop_rq);
index 6a23379..612b5cf 100644 (file)
@@ -210,8 +210,13 @@ void mlx5e_rep_encap_entry_detach(struct mlx5e_priv *priv,
 
 void mlx5e_rep_queue_neigh_stats_work(struct mlx5e_priv *priv);
 
-bool mlx5e_eswitch_rep(struct net_device *netdev);
+bool mlx5e_eswitch_vf_rep(struct net_device *netdev);
 bool mlx5e_eswitch_uplink_rep(struct net_device *netdev);
+static inline bool mlx5e_eswitch_rep(struct net_device *netdev)
+{
+       return mlx5e_eswitch_vf_rep(netdev) ||
+              mlx5e_eswitch_uplink_rep(netdev);
+}
 
 #else /* CONFIG_MLX5_ESWITCH */
 static inline bool mlx5e_is_uplink_rep(struct mlx5e_priv *priv) { return false; }
index a574c58..5bcf95f 100644 (file)
@@ -3073,6 +3073,11 @@ static bool actions_match_supported(struct mlx5e_priv *priv,
        return true;
 }
 
+static bool same_port_devs(struct mlx5e_priv *priv, struct mlx5e_priv *peer_priv)
+{
+       return priv->mdev == peer_priv->mdev;
+}
+
 static bool same_hw_devs(struct mlx5e_priv *priv, struct mlx5e_priv *peer_priv)
 {
        struct mlx5_core_dev *fmdev, *pmdev;
@@ -3291,7 +3296,7 @@ static inline int hash_encap_info(struct encap_key *key)
 }
 
 
-static bool is_merged_eswitch_dev(struct mlx5e_priv *priv,
+static bool is_merged_eswitch_vfs(struct mlx5e_priv *priv,
                                  struct net_device *peer_netdev)
 {
        struct mlx5e_priv *peer_priv;
@@ -3299,13 +3304,11 @@ static bool is_merged_eswitch_dev(struct mlx5e_priv *priv,
        peer_priv = netdev_priv(peer_netdev);
 
        return (MLX5_CAP_ESW(priv->mdev, merged_eswitch) &&
-               mlx5e_eswitch_rep(priv->netdev) &&
-               mlx5e_eswitch_rep(peer_netdev) &&
+               mlx5e_eswitch_vf_rep(priv->netdev) &&
+               mlx5e_eswitch_vf_rep(peer_netdev) &&
                same_hw_devs(priv, peer_priv));
 }
 
-
-
 bool mlx5e_encap_take(struct mlx5e_encap_entry *e)
 {
        return refcount_inc_not_zero(&e->refcnt);
@@ -3575,14 +3578,37 @@ static int add_vlan_pop_action(struct mlx5e_priv *priv,
        return err;
 }
 
+static bool same_hw_reps(struct mlx5e_priv *priv,
+                        struct net_device *peer_netdev)
+{
+       struct mlx5e_priv *peer_priv;
+
+       peer_priv = netdev_priv(peer_netdev);
+
+       return mlx5e_eswitch_rep(priv->netdev) &&
+              mlx5e_eswitch_rep(peer_netdev) &&
+              same_hw_devs(priv, peer_priv);
+}
+
+static bool is_lag_dev(struct mlx5e_priv *priv,
+                      struct net_device *peer_netdev)
+{
+       return ((mlx5_lag_is_sriov(priv->mdev) ||
+                mlx5_lag_is_multipath(priv->mdev)) &&
+                same_hw_reps(priv, peer_netdev));
+}
+
 bool mlx5e_is_valid_eswitch_fwd_dev(struct mlx5e_priv *priv,
                                    struct net_device *out_dev)
 {
-       if (is_merged_eswitch_dev(priv, out_dev))
+       if (is_merged_eswitch_vfs(priv, out_dev))
+               return true;
+
+       if (is_lag_dev(priv, out_dev))
                return true;
 
        return mlx5e_eswitch_rep(out_dev) &&
-              same_hw_devs(priv, netdev_priv(out_dev));
+              same_port_devs(priv, netdev_priv(out_dev));
 }
 
 static bool is_duplicated_output_device(struct net_device *dev,
index fd6b2a1..119a5c6 100644 (file)
@@ -537,10 +537,9 @@ bool mlx5e_poll_tx_cq(struct mlx5e_cq *cq, int napi_budget)
 void mlx5e_free_txqsq_descs(struct mlx5e_txqsq *sq)
 {
        struct mlx5e_tx_wqe_info *wi;
+       u32 dma_fifo_cc, nbytes = 0;
+       u16 ci, sqcc, npkts = 0;
        struct sk_buff *skb;
-       u32 dma_fifo_cc;
-       u16 sqcc;
-       u16 ci;
        int i;
 
        sqcc = sq->cc;
@@ -565,11 +564,15 @@ void mlx5e_free_txqsq_descs(struct mlx5e_txqsq *sq)
                }
 
                dev_kfree_skb_any(skb);
+               npkts++;
+               nbytes += wi->num_bytes;
                sqcc += wi->num_wqebbs;
        }
 
        sq->dma_fifo_cc = dma_fifo_cc;
        sq->cc = sqcc;
+
+       netdev_tx_completed_queue(sq->txq, npkts, nbytes);
 }
 
 #ifdef CONFIG_MLX5_CORE_IPOIB
index cccea3a..ce6c621 100644 (file)
@@ -611,11 +611,13 @@ static int create_async_eqs(struct mlx5_core_dev *dev)
                .nent = MLX5_NUM_CMD_EQE,
                .mask[0] = 1ull << MLX5_EVENT_TYPE_CMD,
        };
+       mlx5_cmd_allowed_opcode(dev, MLX5_CMD_OP_CREATE_EQ);
        err = setup_async_eq(dev, &table->cmd_eq, &param, "cmd");
        if (err)
                goto err1;
 
        mlx5_cmd_use_events(dev);
+       mlx5_cmd_allowed_opcode(dev, CMD_ALLOWED_OPCODE_ALL);
 
        param = (struct mlx5_eq_param) {
                .irq_index = 0,
@@ -645,6 +647,7 @@ err2:
        mlx5_cmd_use_polling(dev);
        cleanup_async_eq(dev, &table->cmd_eq, "cmd");
 err1:
+       mlx5_cmd_allowed_opcode(dev, CMD_ALLOWED_OPCODE_ALL);
        mlx5_eq_notifier_unregister(dev, &table->cq_err_nb);
        return err;
 }
index 8bcf342..3ce17c3 100644 (file)
@@ -346,8 +346,10 @@ int mlx5_events_init(struct mlx5_core_dev *dev)
        events->dev = dev;
        dev->priv.events = events;
        events->wq = create_singlethread_workqueue("mlx5_events");
-       if (!events->wq)
+       if (!events->wq) {
+               kfree(events);
                return -ENOMEM;
+       }
        INIT_WORK(&events->pcie_core_work, mlx5_pcie_event);
 
        return 0;
index d5defe0..9620c86 100644 (file)
@@ -344,17 +344,12 @@ static void tree_put_node(struct fs_node *node, bool locked)
                if (node->del_hw_func)
                        node->del_hw_func(node);
                if (parent_node) {
-                       /* Only root namespace doesn't have parent and we just
-                        * need to free its node.
-                        */
                        down_write_ref_node(parent_node, locked);
                        list_del_init(&node->list);
-                       if (node->del_sw_func)
-                               node->del_sw_func(node);
-                       up_write_ref_node(parent_node, locked);
-               } else {
-                       kfree(node);
                }
+               node->del_sw_func(node);
+               if (parent_node)
+                       up_write_ref_node(parent_node, locked);
                node = NULL;
        }
        if (!node && parent_node)
@@ -468,8 +463,10 @@ static void del_sw_flow_table(struct fs_node *node)
        fs_get_obj(ft, node);
 
        rhltable_destroy(&ft->fgs_hash);
-       fs_get_obj(prio, ft->node.parent);
-       prio->num_ft--;
+       if (ft->node.parent) {
+               fs_get_obj(prio, ft->node.parent);
+               prio->num_ft--;
+       }
        kfree(ft);
 }
 
@@ -2351,6 +2348,17 @@ static int init_root_tree(struct mlx5_flow_steering *steering,
        return 0;
 }
 
+static void del_sw_root_ns(struct fs_node *node)
+{
+       struct mlx5_flow_root_namespace *root_ns;
+       struct mlx5_flow_namespace *ns;
+
+       fs_get_obj(ns, node);
+       root_ns = container_of(ns, struct mlx5_flow_root_namespace, ns);
+       mutex_destroy(&root_ns->chain_lock);
+       kfree(node);
+}
+
 static struct mlx5_flow_root_namespace
 *create_root_ns(struct mlx5_flow_steering *steering,
                enum fs_flow_table_type table_type)
@@ -2377,7 +2385,7 @@ static struct mlx5_flow_root_namespace
        ns = &root_ns->ns;
        fs_init_namespace(ns);
        mutex_init(&root_ns->chain_lock);
-       tree_init_node(&ns->node, NULL, NULL);
+       tree_init_node(&ns->node, NULL, del_sw_root_ns);
        tree_add_node(&ns->node, NULL);
 
        return root_ns;
index 673aaa8..505cf6e 100644 (file)
@@ -396,7 +396,7 @@ static int mlx5i_init_rx(struct mlx5e_priv *priv)
 err_destroy_direct_tirs:
        mlx5e_destroy_direct_tirs(priv, priv->direct_tir);
 err_destroy_indirect_tirs:
-       mlx5e_destroy_indirect_tirs(priv, true);
+       mlx5e_destroy_indirect_tirs(priv);
 err_destroy_direct_rqts:
        mlx5e_destroy_direct_rqts(priv, priv->direct_tir);
 err_destroy_indirect_rqts:
@@ -412,7 +412,7 @@ static void mlx5i_cleanup_rx(struct mlx5e_priv *priv)
 {
        mlx5i_destroy_flow_steering(priv);
        mlx5e_destroy_direct_tirs(priv, priv->direct_tir);
-       mlx5e_destroy_indirect_tirs(priv, true);
+       mlx5e_destroy_indirect_tirs(priv);
        mlx5e_destroy_direct_rqts(priv, priv->direct_tir);
        mlx5e_destroy_rqt(priv, &priv->indir_rqt);
        mlx5e_close_drop_rq(&priv->drop_rq);
index 7af4210..c1618b8 100644 (file)
@@ -965,6 +965,8 @@ static int mlx5_function_setup(struct mlx5_core_dev *dev, bool boot)
                goto err_cmd_cleanup;
        }
 
+       mlx5_cmd_set_state(dev, MLX5_CMDIF_STATE_UP);
+
        err = mlx5_core_enable_hca(dev, 0);
        if (err) {
                mlx5_core_err(dev, "enable hca failed\n");
@@ -1026,6 +1028,7 @@ reclaim_boot_pages:
 err_disable_hca:
        mlx5_core_disable_hca(dev, 0);
 err_cmd_cleanup:
+       mlx5_cmd_set_state(dev, MLX5_CMDIF_STATE_DOWN);
        mlx5_cmd_cleanup(dev);
 
        return err;
@@ -1043,6 +1046,7 @@ static int mlx5_function_teardown(struct mlx5_core_dev *dev, bool boot)
        }
        mlx5_reclaim_startup_pages(dev);
        mlx5_core_disable_hca(dev, 0);
+       mlx5_cmd_set_state(dev, MLX5_CMDIF_STATE_DOWN);
        mlx5_cmd_cleanup(dev);
 
        return 0;
@@ -1191,7 +1195,7 @@ int mlx5_load_one(struct mlx5_core_dev *dev, bool boot)
 
        err = mlx5_function_setup(dev, boot);
        if (err)
-               goto out;
+               goto err_function;
 
        if (boot) {
                err = mlx5_init_once(dev);
@@ -1229,6 +1233,7 @@ err_load:
                mlx5_cleanup_once(dev);
 function_teardown:
        mlx5_function_teardown(dev, boot);
+err_function:
        dev->state = MLX5_DEVICE_STATE_INTERNAL_ERROR;
        mutex_unlock(&dev->intf_state_mutex);
 
index 24ca8d5..6b39978 100644 (file)
@@ -3986,6 +3986,7 @@ static void mlxsw_sp_ports_remove(struct mlxsw_sp *mlxsw_sp)
                        mlxsw_sp_port_remove(mlxsw_sp, i);
        mlxsw_sp_cpu_port_remove(mlxsw_sp);
        kfree(mlxsw_sp->ports);
+       mlxsw_sp->ports = NULL;
 }
 
 static int mlxsw_sp_ports_create(struct mlxsw_sp *mlxsw_sp)
@@ -4022,6 +4023,7 @@ err_port_create:
        mlxsw_sp_cpu_port_remove(mlxsw_sp);
 err_cpu_port_create:
        kfree(mlxsw_sp->ports);
+       mlxsw_sp->ports = NULL;
        return err;
 }
 
@@ -4143,6 +4145,14 @@ static int mlxsw_sp_local_ports_offset(struct mlxsw_core *mlxsw_core,
        return mlxsw_core_res_get(mlxsw_core, local_ports_in_x_res_id);
 }
 
+static struct mlxsw_sp_port *
+mlxsw_sp_port_get_by_local_port(struct mlxsw_sp *mlxsw_sp, u8 local_port)
+{
+       if (mlxsw_sp->ports && mlxsw_sp->ports[local_port])
+               return mlxsw_sp->ports[local_port];
+       return NULL;
+}
+
 static int mlxsw_sp_port_split(struct mlxsw_core *mlxsw_core, u8 local_port,
                               unsigned int count,
                               struct netlink_ext_ack *extack)
@@ -4156,7 +4166,7 @@ static int mlxsw_sp_port_split(struct mlxsw_core *mlxsw_core, u8 local_port,
        int i;
        int err;
 
-       mlxsw_sp_port = mlxsw_sp->ports[local_port];
+       mlxsw_sp_port = mlxsw_sp_port_get_by_local_port(mlxsw_sp, local_port);
        if (!mlxsw_sp_port) {
                dev_err(mlxsw_sp->bus_info->dev, "Port number \"%d\" does not exist\n",
                        local_port);
@@ -4251,7 +4261,7 @@ static int mlxsw_sp_port_unsplit(struct mlxsw_core *mlxsw_core, u8 local_port,
        int offset;
        int i;
 
-       mlxsw_sp_port = mlxsw_sp->ports[local_port];
+       mlxsw_sp_port = mlxsw_sp_port_get_by_local_port(mlxsw_sp, local_port);
        if (!mlxsw_sp_port) {
                dev_err(mlxsw_sp->bus_info->dev, "Port number \"%d\" does not exist\n",
                        local_port);
index 9053582..2503f61 100644 (file)
@@ -1259,6 +1259,7 @@ static void mlxsw_sx_ports_remove(struct mlxsw_sx *mlxsw_sx)
                if (mlxsw_sx_port_created(mlxsw_sx, i))
                        mlxsw_sx_port_remove(mlxsw_sx, i);
        kfree(mlxsw_sx->ports);
+       mlxsw_sx->ports = NULL;
 }
 
 static int mlxsw_sx_ports_create(struct mlxsw_sx *mlxsw_sx)
@@ -1293,6 +1294,7 @@ err_port_module_info_get:
                if (mlxsw_sx_port_created(mlxsw_sx, i))
                        mlxsw_sx_port_remove(mlxsw_sx, i);
        kfree(mlxsw_sx->ports);
+       mlxsw_sx->ports = NULL;
        return err;
 }
 
@@ -1376,6 +1378,12 @@ static int mlxsw_sx_port_type_set(struct mlxsw_core *mlxsw_core, u8 local_port,
        u8 module, width;
        int err;
 
+       if (!mlxsw_sx->ports || !mlxsw_sx->ports[local_port]) {
+               dev_err(mlxsw_sx->bus_info->dev, "Port number \"%d\" does not exist\n",
+                       local_port);
+               return -EINVAL;
+       }
+
        if (new_type == DEVLINK_PORT_TYPE_AUTO)
                return -EOPNOTSUPP;
 
index 02350c3..efb3965 100644 (file)
@@ -1467,7 +1467,7 @@ static void ocelot_port_attr_ageing_set(struct ocelot *ocelot, int port,
                                        unsigned long ageing_clock_t)
 {
        unsigned long ageing_jiffies = clock_t_to_jiffies(ageing_clock_t);
-       u32 ageing_time = jiffies_to_msecs(ageing_jiffies) / 1000;
+       u32 ageing_time = jiffies_to_msecs(ageing_jiffies);
 
        ocelot_set_ageing_time(ocelot, ageing_time);
 }
index 78e15cc..c51b48d 100644 (file)
@@ -1050,6 +1050,13 @@ static u16 rtl_ephy_read(struct rtl8169_private *tp, int reg_addr)
                RTL_R32(tp, EPHYAR) & EPHYAR_DATA_MASK : ~0;
 }
 
+static void r8168fp_adjust_ocp_cmd(struct rtl8169_private *tp, u32 *cmd, int type)
+{
+       /* based on RTL8168FP_OOBMAC_BASE in vendor driver */
+       if (tp->mac_version == RTL_GIGA_MAC_VER_52 && type == ERIAR_OOB)
+               *cmd |= 0x7f0 << 18;
+}
+
 DECLARE_RTL_COND(rtl_eriar_cond)
 {
        return RTL_R32(tp, ERIAR) & ERIAR_FLAG;
@@ -1058,9 +1065,12 @@ DECLARE_RTL_COND(rtl_eriar_cond)
 static void _rtl_eri_write(struct rtl8169_private *tp, int addr, u32 mask,
                           u32 val, int type)
 {
+       u32 cmd = ERIAR_WRITE_CMD | type | mask | addr;
+
        BUG_ON((addr & 3) || (mask == 0));
        RTL_W32(tp, ERIDR, val);
-       RTL_W32(tp, ERIAR, ERIAR_WRITE_CMD | type | mask | addr);
+       r8168fp_adjust_ocp_cmd(tp, &cmd, type);
+       RTL_W32(tp, ERIAR, cmd);
 
        rtl_udelay_loop_wait_low(tp, &rtl_eriar_cond, 100, 100);
 }
@@ -1073,7 +1083,10 @@ static void rtl_eri_write(struct rtl8169_private *tp, int addr, u32 mask,
 
 static u32 _rtl_eri_read(struct rtl8169_private *tp, int addr, int type)
 {
-       RTL_W32(tp, ERIAR, ERIAR_READ_CMD | type | ERIAR_MASK_1111 | addr);
+       u32 cmd = ERIAR_READ_CMD | type | ERIAR_MASK_1111 | addr;
+
+       r8168fp_adjust_ocp_cmd(tp, &cmd, type);
+       RTL_W32(tp, ERIAR, cmd);
 
        return rtl_udelay_loop_wait_high(tp, &rtl_eriar_cond, 100, 100) ?
                RTL_R32(tp, ERIDR) : ~0;
index 7305e8e..6646eba 100644 (file)
@@ -848,14 +848,14 @@ static int ioc3eth_probe(struct platform_device *pdev)
        ip = netdev_priv(dev);
        ip->dma_dev = pdev->dev.parent;
        ip->regs = devm_platform_ioremap_resource(pdev, 0);
-       if (!ip->regs) {
-               err = -ENOMEM;
+       if (IS_ERR(ip->regs)) {
+               err = PTR_ERR(ip->regs);
                goto out_free;
        }
 
        ip->ssram = devm_platform_ioremap_resource(pdev, 1);
-       if (!ip->ssram) {
-               err = -ENOMEM;
+       if (IS_ERR(ip->ssram)) {
+               err = PTR_ERR(ip->ssram);
                goto out_free;
        }
 
index 49a6a91..fc168f8 100644 (file)
@@ -2493,20 +2493,20 @@ static int smsc911x_drv_probe(struct platform_device *pdev)
 
        retval = smsc911x_init(dev);
        if (retval < 0)
-               goto out_disable_resources;
+               goto out_init_fail;
 
        netif_carrier_off(dev);
 
        retval = smsc911x_mii_init(pdev, dev);
        if (retval) {
                SMSC_WARN(pdata, probe, "Error %i initialising mii", retval);
-               goto out_disable_resources;
+               goto out_init_fail;
        }
 
        retval = register_netdev(dev);
        if (retval) {
                SMSC_WARN(pdata, probe, "Error %i registering device", retval);
-               goto out_disable_resources;
+               goto out_init_fail;
        } else {
                SMSC_TRACE(pdata, probe,
                           "Network interface: \"%s\"", dev->name);
@@ -2547,9 +2547,10 @@ static int smsc911x_drv_probe(struct platform_device *pdev)
 
        return 0;
 
-out_disable_resources:
+out_init_fail:
        pm_runtime_put(&pdev->dev);
        pm_runtime_disable(&pdev->dev);
+out_disable_resources:
        (void)smsc911x_disable_resources(pdev);
 out_enable_resources_fail:
        smsc911x_free_resources(pdev);
index 6ae13dc..02102c7 100644 (file)
@@ -319,6 +319,19 @@ static int ipq806x_gmac_probe(struct platform_device *pdev)
        /* Enable PTP clock */
        regmap_read(gmac->nss_common, NSS_COMMON_CLK_GATE, &val);
        val |= NSS_COMMON_CLK_GATE_PTP_EN(gmac->id);
+       switch (gmac->phy_mode) {
+       case PHY_INTERFACE_MODE_RGMII:
+               val |= NSS_COMMON_CLK_GATE_RGMII_RX_EN(gmac->id) |
+                       NSS_COMMON_CLK_GATE_RGMII_TX_EN(gmac->id);
+               break;
+       case PHY_INTERFACE_MODE_SGMII:
+               val |= NSS_COMMON_CLK_GATE_GMII_RX_EN(gmac->id) |
+                               NSS_COMMON_CLK_GATE_GMII_TX_EN(gmac->id);
+               break;
+       default:
+               /* We don't get here; the switch above will have errored out */
+               unreachable();
+       }
        regmap_write(gmac->nss_common, NSS_COMMON_CLK_GATE, val);
 
        if (gmac->phy_mode == PHY_INTERFACE_MODE_SGMII) {
index a999d6b..1f319c9 100644 (file)
@@ -5190,8 +5190,6 @@ int stmmac_resume(struct device *dev)
                        return ret;
        }
 
-       netif_device_attach(ndev);
-
        mutex_lock(&priv->lock);
 
        stmmac_reset_queues_param(priv);
@@ -5218,6 +5216,8 @@ int stmmac_resume(struct device *dev)
 
        phylink_mac_change(priv->phylink, true);
 
+       netif_device_attach(ndev);
+
        return 0;
 }
 EXPORT_SYMBOL_GPL(stmmac_resume);
index e6d1aa8..f1c8615 100644 (file)
@@ -4963,7 +4963,7 @@ static int cas_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
                                          cas_cacheline_size)) {
                        dev_err(&pdev->dev, "Could not set PCI cache "
                               "line size\n");
-                       goto err_write_cacheline;
+                       goto err_out_free_res;
                }
        }
 #endif
@@ -5136,7 +5136,6 @@ err_out_iounmap:
 err_out_free_res:
        pci_release_regions(pdev);
 
-err_write_cacheline:
        /* Try to restore it in case the error occurred after we
         * set it.
         */
index 2517ffb..88f52a2 100644 (file)
@@ -1895,8 +1895,9 @@ static int am65_cpsw_nuss_probe(struct platform_device *pdev)
        ale_params.nu_switch_ale = true;
 
        common->ale = cpsw_ale_create(&ale_params);
-       if (!common->ale) {
+       if (IS_ERR(common->ale)) {
                dev_err(dev, "error initializing ale engine\n");
+               ret = PTR_ERR(common->ale);
                goto err_of_clear;
        }
 
index c2c5bf8..ffeb863 100644 (file)
@@ -1753,11 +1753,15 @@ static int cpsw_suspend(struct device *dev)
        struct cpsw_common *cpsw = dev_get_drvdata(dev);
        int i;
 
+       rtnl_lock();
+
        for (i = 0; i < cpsw->data.slaves; i++)
                if (cpsw->slaves[i].ndev)
                        if (netif_running(cpsw->slaves[i].ndev))
                                cpsw_ndo_stop(cpsw->slaves[i].ndev);
 
+       rtnl_unlock();
+
        /* Select sleep pin state */
        pinctrl_pm_select_sleep_state(dev);
 
index 0374e69..8dc6be1 100644 (file)
@@ -955,7 +955,7 @@ struct cpsw_ale *cpsw_ale_create(struct cpsw_ale_params *params)
 
        ale = devm_kzalloc(params->dev, sizeof(*ale), GFP_KERNEL);
        if (!ale)
-               return NULL;
+               return ERR_PTR(-ENOMEM);
 
        ale->p0_untag_vid_mask =
                devm_kmalloc_array(params->dev, BITS_TO_LONGS(VLAN_N_VID),
index 97a058c..d0b6c41 100644 (file)
@@ -490,9 +490,9 @@ int cpsw_init_common(struct cpsw_common *cpsw, void __iomem *ss_regs,
        ale_params.ale_ports            = CPSW_ALE_PORTS_NUM;
 
        cpsw->ale = cpsw_ale_create(&ale_params);
-       if (!cpsw->ale) {
+       if (IS_ERR(cpsw->ale)) {
                dev_err(dev, "error initializing ale engine\n");
-               return -ENODEV;
+               return PTR_ERR(cpsw->ale);
        }
 
        dma_params.dev          = dev;
index fb36115..fdbae73 100644 (file)
@@ -3704,9 +3704,9 @@ static int gbe_probe(struct netcp_device *netcp_device, struct device *dev,
                ale_params.nu_switch_ale = true;
        }
        gbe_dev->ale = cpsw_ale_create(&ale_params);
-       if (!gbe_dev->ale) {
+       if (IS_ERR(gbe_dev->ale)) {
                dev_err(gbe_dev->dev, "error initializing ale engine\n");
-               ret = -ENODEV;
+               ret = PTR_ERR(gbe_dev->ale);
                goto free_sec_ports;
        } else {
                dev_dbg(gbe_dev->dev, "Created a gbe ale engine\n");
index b671bea..8d9ca1c 100644 (file)
@@ -1392,6 +1392,7 @@ static int gsi_channel_poll(struct napi_struct *napi, int budget)
        while (count < budget) {
                struct gsi_trans *trans;
 
+               count++;
                trans = gsi_channel_poll_one(channel);
                if (!trans)
                        break;
index 68668a2..dc3ff0e 100644 (file)
@@ -858,8 +858,7 @@ nsim_dev_devlink_trap_policer_counter_get(struct devlink *devlink,
                return -EINVAL;
 
        cnt = &nsim_dev->trap_data->trap_policers_cnt_arr[policer->id - 1];
-       *p_drops = *cnt;
-       *cnt += jiffies % 64;
+       *p_drops = (*cnt)++;
 
        return 0;
 }
index 030bf8b..414e3b3 100644 (file)
@@ -354,6 +354,8 @@ struct vsc8531_private {
        u64 *stats;
        int nstats;
        bool pkg_init;
+       /* PHY address within the package. */
+       u8 addr;
        /* For multiple port PHYs; the MDIO address of the base PHY in the
         * package.
         */
index fcb5ba5..59b6837 100644 (file)
 #define MSCC_MAC_PAUSE_CFG_STATE_PAUSE_STATE                   BIT(0)
 #define MSCC_MAC_PAUSE_CFG_STATE_MAC_TX_PAUSE_GEN              BIT(4)
 
-#define MSCC_PROC_0_IP_1588_TOP_CFG_STAT_MODE_CTL                      0x2
-#define MSCC_PROC_0_IP_1588_TOP_CFG_STAT_MODE_CTL_PROTOCOL_MODE(x)     (x)
-#define MSCC_PROC_0_IP_1588_TOP_CFG_STAT_MODE_CTL_PROTOCOL_MODE_M      GENMASK(2, 0)
+#define MSCC_PROC_IP_1588_TOP_CFG_STAT_MODE_CTL                        0x2
+#define MSCC_PROC_IP_1588_TOP_CFG_STAT_MODE_CTL_PROTOCOL_MODE(x)       (x)
+#define MSCC_PROC_IP_1588_TOP_CFG_STAT_MODE_CTL_PROTOCOL_MODE_M        GENMASK(2, 0)
 
 #endif /* _MSCC_PHY_LINE_MAC_H_ */
index e99e2cd..b4d3dc4 100644 (file)
@@ -316,6 +316,8 @@ static void vsc8584_macsec_mac_init(struct phy_device *phydev,
 /* Must be called with mdio_lock taken */
 static int __vsc8584_macsec_init(struct phy_device *phydev)
 {
+       struct vsc8531_private *priv = phydev->priv;
+       enum macsec_bank proc_bank;
        u32 val;
 
        vsc8584_macsec_block_init(phydev, MACSEC_INGR);
@@ -351,12 +353,14 @@ static int __vsc8584_macsec_init(struct phy_device *phydev)
        val |= MSCC_FCBUF_ENA_CFG_TX_ENA | MSCC_FCBUF_ENA_CFG_RX_ENA;
        vsc8584_macsec_phy_write(phydev, FC_BUFFER, MSCC_FCBUF_ENA_CFG, val);
 
-       val = vsc8584_macsec_phy_read(phydev, IP_1588,
-                                     MSCC_PROC_0_IP_1588_TOP_CFG_STAT_MODE_CTL);
-       val &= ~MSCC_PROC_0_IP_1588_TOP_CFG_STAT_MODE_CTL_PROTOCOL_MODE_M;
-       val |= MSCC_PROC_0_IP_1588_TOP_CFG_STAT_MODE_CTL_PROTOCOL_MODE(4);
-       vsc8584_macsec_phy_write(phydev, IP_1588,
-                                MSCC_PROC_0_IP_1588_TOP_CFG_STAT_MODE_CTL, val);
+       proc_bank = (priv->addr < 2) ? PROC_0 : PROC_2;
+
+       val = vsc8584_macsec_phy_read(phydev, proc_bank,
+                                     MSCC_PROC_IP_1588_TOP_CFG_STAT_MODE_CTL);
+       val &= ~MSCC_PROC_IP_1588_TOP_CFG_STAT_MODE_CTL_PROTOCOL_MODE_M;
+       val |= MSCC_PROC_IP_1588_TOP_CFG_STAT_MODE_CTL_PROTOCOL_MODE(4);
+       vsc8584_macsec_phy_write(phydev, proc_bank,
+                                MSCC_PROC_IP_1588_TOP_CFG_STAT_MODE_CTL, val);
 
        return 0;
 }
index d078394..d751f29 100644 (file)
@@ -64,7 +64,8 @@ enum macsec_bank {
        FC_BUFFER   = 0x04,
        HOST_MAC    = 0x05,
        LINE_MAC    = 0x06,
-       IP_1588     = 0x0e,
+       PROC_0      = 0x0e,
+       PROC_2      = 0x0f,
        MACSEC_INGR = 0x38,
        MACSEC_EGR  = 0x3c,
 };
index acddef7..c8aa6d9 100644 (file)
@@ -1347,6 +1347,8 @@ static int vsc8584_config_init(struct phy_device *phydev)
        else
                vsc8531->base_addr = phydev->mdio.addr - addr;
 
+       vsc8531->addr = addr;
+
        /* Some parts of the init sequence are identical for every PHY in the
         * package. Some parts are modifying the GPIO register bank which is a
         * set of registers that are affecting all PHYs, a few resetting the
@@ -1771,6 +1773,8 @@ static int vsc8514_config_init(struct phy_device *phydev)
        else
                vsc8531->base_addr = phydev->mdio.addr - addr;
 
+       vsc8531->addr = addr;
+
        /* Some parts of the init sequence are identical for every PHY in the
         * package. Some parts are modifying the GPIO register bank which is a
         * set of registers that are affecting all PHYs, a few resetting the
index ac27841..697c74d 100644 (file)
@@ -1233,7 +1233,7 @@ int phy_sfp_probe(struct phy_device *phydev,
                  const struct sfp_upstream_ops *ops)
 {
        struct sfp_bus *bus;
-       int ret;
+       int ret = 0;
 
        if (phydev->mdio.dev.fwnode) {
                bus = sfp_bus_find_fwnode(phydev->mdio.dev.fwnode);
@@ -1245,7 +1245,7 @@ int phy_sfp_probe(struct phy_device *phydev,
                ret = sfp_bus_add_upstream(bus, phydev, ops);
                sfp_bus_put(bus);
        }
-       return 0;
+       return ret;
 }
 EXPORT_SYMBOL(phy_sfp_probe);
 
index 0cdb2ce..a657943 100644 (file)
@@ -815,14 +815,21 @@ static const struct usb_device_id products[] = {
        .driver_info = 0,
 },
 
-/* Microsoft Surface 3 dock (based on Realtek RTL8153) */
+/* Microsoft Surface Ethernet Adapter (based on Realtek RTL8153) */
 {
        USB_DEVICE_AND_INTERFACE_INFO(MICROSOFT_VENDOR_ID, 0x07c6, USB_CLASS_COMM,
                        USB_CDC_SUBCLASS_ETHERNET, USB_CDC_PROTO_NONE),
        .driver_info = 0,
 },
 
-       /* TP-LINK UE300 USB 3.0 Ethernet Adapters (based on Realtek RTL8153) */
+/* Microsoft Surface Ethernet Adapter (based on Realtek RTL8153B) */
+{
+       USB_DEVICE_AND_INTERFACE_INFO(MICROSOFT_VENDOR_ID, 0x0927, USB_CLASS_COMM,
+                       USB_CDC_SUBCLASS_ETHERNET, USB_CDC_PROTO_NONE),
+       .driver_info = 0,
+},
+
+/* TP-LINK UE300 USB 3.0 Ethernet Adapters (based on Realtek RTL8153) */
 {
        USB_DEVICE_AND_INTERFACE_INFO(TPLINK_VENDOR_ID, 0x0601, USB_CLASS_COMM,
                        USB_CDC_SUBCLASS_ETHERNET, USB_CDC_PROTO_NONE),
index 8f8d988..c8c873a 100644 (file)
@@ -6880,6 +6880,7 @@ static const struct usb_device_id rtl8152_table[] = {
        {REALTEK_USB_DEVICE(VENDOR_ID_REALTEK, 0x8153)},
        {REALTEK_USB_DEVICE(VENDOR_ID_MICROSOFT, 0x07ab)},
        {REALTEK_USB_DEVICE(VENDOR_ID_MICROSOFT, 0x07c6)},
+       {REALTEK_USB_DEVICE(VENDOR_ID_MICROSOFT, 0x0927)},
        {REALTEK_USB_DEVICE(VENDOR_ID_SAMSUNG, 0xa101)},
        {REALTEK_USB_DEVICE(VENDOR_ID_LENOVO,  0x304f)},
        {REALTEK_USB_DEVICE(VENDOR_ID_LENOVO,  0x3062)},
index b8a7b9c..208da72 100644 (file)
@@ -32,7 +32,7 @@ enum cookie_values {
 };
 
 enum counter_values {
-       COUNTER_BITS_TOTAL = 2048,
+       COUNTER_BITS_TOTAL = 8192,
        COUNTER_REDUNDANT_BITS = BITS_PER_LONG,
        COUNTER_WINDOW_SIZE = COUNTER_BITS_TOTAL - COUNTER_REDUNDANT_BITS
 };
index 708dc61..6264336 100644 (file)
@@ -104,6 +104,7 @@ static struct noise_keypair *keypair_create(struct wg_peer *peer)
 
        if (unlikely(!keypair))
                return NULL;
+       spin_lock_init(&keypair->receiving_counter.lock);
        keypair->internal_id = atomic64_inc_return(&keypair_counter);
        keypair->entry.type = INDEX_HASHTABLE_KEYPAIR;
        keypair->entry.peer = peer;
@@ -358,25 +359,16 @@ out:
        memzero_explicit(output, BLAKE2S_HASH_SIZE + 1);
 }
 
-static void symmetric_key_init(struct noise_symmetric_key *key)
-{
-       spin_lock_init(&key->counter.receive.lock);
-       atomic64_set(&key->counter.counter, 0);
-       memset(key->counter.receive.backtrack, 0,
-              sizeof(key->counter.receive.backtrack));
-       key->birthdate = ktime_get_coarse_boottime_ns();
-       key->is_valid = true;
-}
-
 static void derive_keys(struct noise_symmetric_key *first_dst,
                        struct noise_symmetric_key *second_dst,
                        const u8 chaining_key[NOISE_HASH_LEN])
 {
+       u64 birthdate = ktime_get_coarse_boottime_ns();
        kdf(first_dst->key, second_dst->key, NULL, NULL,
            NOISE_SYMMETRIC_KEY_LEN, NOISE_SYMMETRIC_KEY_LEN, 0, 0,
            chaining_key);
-       symmetric_key_init(first_dst);
-       symmetric_key_init(second_dst);
+       first_dst->birthdate = second_dst->birthdate = birthdate;
+       first_dst->is_valid = second_dst->is_valid = true;
 }
 
 static bool __must_check mix_dh(u8 chaining_key[NOISE_HASH_LEN],
@@ -715,6 +707,7 @@ wg_noise_handshake_consume_response(struct message_handshake_response *src,
        u8 e[NOISE_PUBLIC_KEY_LEN];
        u8 ephemeral_private[NOISE_PUBLIC_KEY_LEN];
        u8 static_private[NOISE_PUBLIC_KEY_LEN];
+       u8 preshared_key[NOISE_SYMMETRIC_KEY_LEN];
 
        down_read(&wg->static_identity.lock);
 
@@ -733,6 +726,8 @@ wg_noise_handshake_consume_response(struct message_handshake_response *src,
        memcpy(chaining_key, handshake->chaining_key, NOISE_HASH_LEN);
        memcpy(ephemeral_private, handshake->ephemeral_private,
               NOISE_PUBLIC_KEY_LEN);
+       memcpy(preshared_key, handshake->preshared_key,
+              NOISE_SYMMETRIC_KEY_LEN);
        up_read(&handshake->lock);
 
        if (state != HANDSHAKE_CREATED_INITIATION)
@@ -750,7 +745,7 @@ wg_noise_handshake_consume_response(struct message_handshake_response *src,
                goto fail;
 
        /* psk */
-       mix_psk(chaining_key, hash, key, handshake->preshared_key);
+       mix_psk(chaining_key, hash, key, preshared_key);
 
        /* {} */
        if (!message_decrypt(NULL, src->encrypted_nothing,
@@ -783,6 +778,7 @@ out:
        memzero_explicit(chaining_key, NOISE_HASH_LEN);
        memzero_explicit(ephemeral_private, NOISE_PUBLIC_KEY_LEN);
        memzero_explicit(static_private, NOISE_PUBLIC_KEY_LEN);
+       memzero_explicit(preshared_key, NOISE_SYMMETRIC_KEY_LEN);
        up_read(&wg->static_identity.lock);
        return ret_peer;
 }
index f532d59..c527253 100644 (file)
 #include <linux/mutex.h>
 #include <linux/kref.h>
 
-union noise_counter {
-       struct {
-               u64 counter;
-               unsigned long backtrack[COUNTER_BITS_TOTAL / BITS_PER_LONG];
-               spinlock_t lock;
-       } receive;
-       atomic64_t counter;
+struct noise_replay_counter {
+       u64 counter;
+       spinlock_t lock;
+       unsigned long backtrack[COUNTER_BITS_TOTAL / BITS_PER_LONG];
 };
 
 struct noise_symmetric_key {
        u8 key[NOISE_SYMMETRIC_KEY_LEN];
-       union noise_counter counter;
        u64 birthdate;
        bool is_valid;
 };
@@ -34,7 +30,9 @@ struct noise_symmetric_key {
 struct noise_keypair {
        struct index_hashtable_entry entry;
        struct noise_symmetric_key sending;
+       atomic64_t sending_counter;
        struct noise_symmetric_key receiving;
+       struct noise_replay_counter receiving_counter;
        __le32 remote_index;
        bool i_am_the_initiator;
        struct kref refcount;
index 3432232..c58df43 100644 (file)
@@ -87,12 +87,20 @@ static inline bool wg_check_packet_protocol(struct sk_buff *skb)
        return real_protocol && skb->protocol == real_protocol;
 }
 
-static inline void wg_reset_packet(struct sk_buff *skb)
+static inline void wg_reset_packet(struct sk_buff *skb, bool encapsulating)
 {
+       u8 l4_hash = skb->l4_hash;
+       u8 sw_hash = skb->sw_hash;
+       u32 hash = skb->hash;
        skb_scrub_packet(skb, true);
        memset(&skb->headers_start, 0,
               offsetof(struct sk_buff, headers_end) -
                       offsetof(struct sk_buff, headers_start));
+       if (encapsulating) {
+               skb->l4_hash = l4_hash;
+               skb->sw_hash = sw_hash;
+               skb->hash = hash;
+       }
        skb->queue_mapping = 0;
        skb->nohdr = 0;
        skb->peeked = 0;
index 3bb5b9a..9143814 100644 (file)
@@ -245,20 +245,20 @@ static void keep_key_fresh(struct wg_peer *peer)
        }
 }
 
-static bool decrypt_packet(struct sk_buff *skb, struct noise_symmetric_key *key)
+static bool decrypt_packet(struct sk_buff *skb, struct noise_keypair *keypair)
 {
        struct scatterlist sg[MAX_SKB_FRAGS + 8];
        struct sk_buff *trailer;
        unsigned int offset;
        int num_frags;
 
-       if (unlikely(!key))
+       if (unlikely(!keypair))
                return false;
 
-       if (unlikely(!READ_ONCE(key->is_valid) ||
-                 wg_birthdate_has_expired(key->birthdate, REJECT_AFTER_TIME) ||
-                 key->counter.receive.counter >= REJECT_AFTER_MESSAGES)) {
-               WRITE_ONCE(key->is_valid, false);
+       if (unlikely(!READ_ONCE(keypair->receiving.is_valid) ||
+                 wg_birthdate_has_expired(keypair->receiving.birthdate, REJECT_AFTER_TIME) ||
+                 keypair->receiving_counter.counter >= REJECT_AFTER_MESSAGES)) {
+               WRITE_ONCE(keypair->receiving.is_valid, false);
                return false;
        }
 
@@ -283,7 +283,7 @@ static bool decrypt_packet(struct sk_buff *skb, struct noise_symmetric_key *key)
 
        if (!chacha20poly1305_decrypt_sg_inplace(sg, skb->len, NULL, 0,
                                                 PACKET_CB(skb)->nonce,
-                                                key->key))
+                                                keypair->receiving.key))
                return false;
 
        /* Another ugly situation of pushing and pulling the header so as to
@@ -298,41 +298,41 @@ static bool decrypt_packet(struct sk_buff *skb, struct noise_symmetric_key *key)
 }
 
 /* This is RFC6479, a replay detection bitmap algorithm that avoids bitshifts */
-static bool counter_validate(union noise_counter *counter, u64 their_counter)
+static bool counter_validate(struct noise_replay_counter *counter, u64 their_counter)
 {
        unsigned long index, index_current, top, i;
        bool ret = false;
 
-       spin_lock_bh(&counter->receive.lock);
+       spin_lock_bh(&counter->lock);
 
-       if (unlikely(counter->receive.counter >= REJECT_AFTER_MESSAGES + 1 ||
+       if (unlikely(counter->counter >= REJECT_AFTER_MESSAGES + 1 ||
                     their_counter >= REJECT_AFTER_MESSAGES))
                goto out;
 
        ++their_counter;
 
        if (unlikely((COUNTER_WINDOW_SIZE + their_counter) <
-                    counter->receive.counter))
+                    counter->counter))
                goto out;
 
        index = their_counter >> ilog2(BITS_PER_LONG);
 
-       if (likely(their_counter > counter->receive.counter)) {
-               index_current = counter->receive.counter >> ilog2(BITS_PER_LONG);
+       if (likely(their_counter > counter->counter)) {
+               index_current = counter->counter >> ilog2(BITS_PER_LONG);
                top = min_t(unsigned long, index - index_current,
                            COUNTER_BITS_TOTAL / BITS_PER_LONG);
                for (i = 1; i <= top; ++i)
-                       counter->receive.backtrack[(i + index_current) &
+                       counter->backtrack[(i + index_current) &
                                ((COUNTER_BITS_TOTAL / BITS_PER_LONG) - 1)] = 0;
-               counter->receive.counter = their_counter;
+               counter->counter = their_counter;
        }
 
        index &= (COUNTER_BITS_TOTAL / BITS_PER_LONG) - 1;
        ret = !test_and_set_bit(their_counter & (BITS_PER_LONG - 1),
-                               &counter->receive.backtrack[index]);
+                               &counter->backtrack[index]);
 
 out:
-       spin_unlock_bh(&counter->receive.lock);
+       spin_unlock_bh(&counter->lock);
        return ret;
 }
 
@@ -472,19 +472,19 @@ int wg_packet_rx_poll(struct napi_struct *napi, int budget)
                if (unlikely(state != PACKET_STATE_CRYPTED))
                        goto next;
 
-               if (unlikely(!counter_validate(&keypair->receiving.counter,
+               if (unlikely(!counter_validate(&keypair->receiving_counter,
                                               PACKET_CB(skb)->nonce))) {
                        net_dbg_ratelimited("%s: Packet has invalid nonce %llu (max %llu)\n",
                                            peer->device->dev->name,
                                            PACKET_CB(skb)->nonce,
-                                           keypair->receiving.counter.receive.counter);
+                                           keypair->receiving_counter.counter);
                        goto next;
                }
 
                if (unlikely(wg_socket_endpoint_from_skb(&endpoint, skb)))
                        goto next;
 
-               wg_reset_packet(skb);
+               wg_reset_packet(skb, false);
                wg_packet_consume_data_done(peer, skb, &endpoint);
                free = false;
 
@@ -511,8 +511,8 @@ void wg_packet_decrypt_worker(struct work_struct *work)
        struct sk_buff *skb;
 
        while ((skb = ptr_ring_consume_bh(&queue->ring)) != NULL) {
-               enum packet_state state = likely(decrypt_packet(skb,
-                               &PACKET_CB(skb)->keypair->receiving)) ?
+               enum packet_state state =
+                       likely(decrypt_packet(skb, PACKET_CB(skb)->keypair)) ?
                                PACKET_STATE_CRYPTED : PACKET_STATE_DEAD;
                wg_queue_enqueue_per_peer_napi(skb, state);
                if (need_resched())
index f4fbb90..ec3c156 100644 (file)
@@ -6,18 +6,24 @@
 #ifdef DEBUG
 bool __init wg_packet_counter_selftest(void)
 {
+       struct noise_replay_counter *counter;
        unsigned int test_num = 0, i;
-       union noise_counter counter;
        bool success = true;
 
-#define T_INIT do {                                               \
-               memset(&counter, 0, sizeof(union noise_counter)); \
-               spin_lock_init(&counter.receive.lock);            \
+       counter = kmalloc(sizeof(*counter), GFP_KERNEL);
+       if (unlikely(!counter)) {
+               pr_err("nonce counter self-test malloc: FAIL\n");
+               return false;
+       }
+
+#define T_INIT do {                                    \
+               memset(counter, 0, sizeof(*counter));  \
+               spin_lock_init(&counter->lock);        \
        } while (0)
 #define T_LIM (COUNTER_WINDOW_SIZE + 1)
 #define T(n, v) do {                                                  \
                ++test_num;                                           \
-               if (counter_validate(&counter, n) != (v)) {           \
+               if (counter_validate(counter, n) != (v)) {            \
                        pr_err("nonce counter self-test %u: FAIL\n",  \
                               test_num);                             \
                        success = false;                              \
@@ -99,6 +105,7 @@ bool __init wg_packet_counter_selftest(void)
 
        if (success)
                pr_info("nonce counter self-tests: pass\n");
+       kfree(counter);
        return success;
 }
 #endif
index 6687db6..f74b934 100644 (file)
@@ -129,7 +129,7 @@ static void keep_key_fresh(struct wg_peer *peer)
        rcu_read_lock_bh();
        keypair = rcu_dereference_bh(peer->keypairs.current_keypair);
        send = keypair && READ_ONCE(keypair->sending.is_valid) &&
-              (atomic64_read(&keypair->sending.counter.counter) > REKEY_AFTER_MESSAGES ||
+              (atomic64_read(&keypair->sending_counter) > REKEY_AFTER_MESSAGES ||
                (keypair->i_am_the_initiator &&
                 wg_birthdate_has_expired(keypair->sending.birthdate, REKEY_AFTER_TIME)));
        rcu_read_unlock_bh();
@@ -167,6 +167,11 @@ static bool encrypt_packet(struct sk_buff *skb, struct noise_keypair *keypair)
        struct sk_buff *trailer;
        int num_frags;
 
+       /* Force hash calculation before encryption so that flow analysis is
+        * consistent over the inner packet.
+        */
+       skb_get_hash(skb);
+
        /* Calculate lengths. */
        padding_len = calculate_skb_padding(skb);
        trailer_len = padding_len + noise_encrypted_len(0);
@@ -295,7 +300,7 @@ void wg_packet_encrypt_worker(struct work_struct *work)
                skb_list_walk_safe(first, skb, next) {
                        if (likely(encrypt_packet(skb,
                                        PACKET_CB(first)->keypair))) {
-                               wg_reset_packet(skb);
+                               wg_reset_packet(skb, true);
                        } else {
                                state = PACKET_STATE_DEAD;
                                break;
@@ -344,7 +349,6 @@ void wg_packet_purge_staged_packets(struct wg_peer *peer)
 
 void wg_packet_send_staged_packets(struct wg_peer *peer)
 {
-       struct noise_symmetric_key *key;
        struct noise_keypair *keypair;
        struct sk_buff_head packets;
        struct sk_buff *skb;
@@ -364,10 +368,9 @@ void wg_packet_send_staged_packets(struct wg_peer *peer)
        rcu_read_unlock_bh();
        if (unlikely(!keypair))
                goto out_nokey;
-       key = &keypair->sending;
-       if (unlikely(!READ_ONCE(key->is_valid)))
+       if (unlikely(!READ_ONCE(keypair->sending.is_valid)))
                goto out_nokey;
-       if (unlikely(wg_birthdate_has_expired(key->birthdate,
+       if (unlikely(wg_birthdate_has_expired(keypair->sending.birthdate,
                                              REJECT_AFTER_TIME)))
                goto out_invalid;
 
@@ -382,7 +385,7 @@ void wg_packet_send_staged_packets(struct wg_peer *peer)
                 */
                PACKET_CB(skb)->ds = ip_tunnel_ecn_encap(0, ip_hdr(skb), skb);
                PACKET_CB(skb)->nonce =
-                               atomic64_inc_return(&key->counter.counter) - 1;
+                               atomic64_inc_return(&keypair->sending_counter) - 1;
                if (unlikely(PACKET_CB(skb)->nonce >= REJECT_AFTER_MESSAGES))
                        goto out_invalid;
        }
@@ -394,7 +397,7 @@ void wg_packet_send_staged_packets(struct wg_peer *peer)
        return;
 
 out_invalid:
-       WRITE_ONCE(key->is_valid, false);
+       WRITE_ONCE(keypair->sending.is_valid, false);
 out_nokey:
        wg_noise_keypair_put(keypair, false);
 
index 6744c02..29971c2 100644 (file)
@@ -1092,6 +1092,10 @@ static int iwl_pci_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
                        iwl_trans->cfg = &iwl_ax101_cfg_quz_hr;
                else if (iwl_trans->cfg == &iwl_ax201_cfg_qu_hr)
                        iwl_trans->cfg = &iwl_ax201_cfg_quz_hr;
+               else if (iwl_trans->cfg == &killer1650s_2ax_cfg_qu_b0_hr_b0)
+                       iwl_trans->cfg = &iwl_ax1650s_cfg_quz_hr;
+               else if (iwl_trans->cfg == &killer1650i_2ax_cfg_qu_b0_hr_b0)
+                       iwl_trans->cfg = &iwl_ax1650i_cfg_quz_hr;
        }
 
 #endif
index 8155f59..10af330 100644 (file)
@@ -877,6 +877,11 @@ rio_dma_transfer(struct file *filp, u32 transfer_mode,
                                rmcd_error("pinned %ld out of %ld pages",
                                           pinned, nr_pages);
                        ret = -EFAULT;
+                       /*
+                        * Set nr_pages up to mean "how many pages to unpin, in
+                        * the error handler:
+                        */
+                       nr_pages = pinned;
                        goto err_pg;
                }
 
index 953c7fb..79bdc12 100644 (file)
@@ -832,11 +832,11 @@ static const struct rpmh_vreg_init_data pm8150_vreg_data[] = {
        RPMH_VREG("ldo10",  "ldo%s10", &pmic5_pldo,      "vdd-l2-l10"),
        RPMH_VREG("ldo11",  "ldo%s11", &pmic5_nldo,      "vdd-l1-l8-l11"),
        RPMH_VREG("ldo12",  "ldo%s12", &pmic5_pldo_lv,   "vdd-l7-l12-l14-l15"),
-       RPMH_VREG("ldo13",  "ldo%s13", &pmic5_pldo,      "vdd-l13-l6-l17"),
+       RPMH_VREG("ldo13",  "ldo%s13", &pmic5_pldo,      "vdd-l13-l16-l17"),
        RPMH_VREG("ldo14",  "ldo%s14", &pmic5_pldo_lv,   "vdd-l7-l12-l14-l15"),
        RPMH_VREG("ldo15",  "ldo%s15", &pmic5_pldo_lv,   "vdd-l7-l12-l14-l15"),
-       RPMH_VREG("ldo16",  "ldo%s16", &pmic5_pldo,      "vdd-l13-l6-l17"),
-       RPMH_VREG("ldo17",  "ldo%s17", &pmic5_pldo,      "vdd-l13-l6-l17"),
+       RPMH_VREG("ldo16",  "ldo%s16", &pmic5_pldo,      "vdd-l13-l16-l17"),
+       RPMH_VREG("ldo17",  "ldo%s17", &pmic5_pldo,      "vdd-l13-l16-l17"),
        RPMH_VREG("ldo18",  "ldo%s18", &pmic5_nldo,      "vdd-l3-l4-l5-l18"),
        {},
 };
@@ -857,7 +857,7 @@ static const struct rpmh_vreg_init_data pm8150l_vreg_data[] = {
        RPMH_VREG("ldo5",   "ldo%s5",  &pmic5_pldo,      "vdd-l4-l5-l6"),
        RPMH_VREG("ldo6",   "ldo%s6",  &pmic5_pldo,      "vdd-l4-l5-l6"),
        RPMH_VREG("ldo7",   "ldo%s7",  &pmic5_pldo,      "vdd-l7-l11"),
-       RPMH_VREG("ldo8",   "ldo%s8",  &pmic5_pldo_lv,   "vdd-l1-l8-l11"),
+       RPMH_VREG("ldo8",   "ldo%s8",  &pmic5_pldo_lv,   "vdd-l1-l8"),
        RPMH_VREG("ldo9",   "ldo%s9",  &pmic5_pldo,      "vdd-l9-l10"),
        RPMH_VREG("ldo10",  "ldo%s10", &pmic5_pldo,      "vdd-l9-l10"),
        RPMH_VREG("ldo11",  "ldo%s11", &pmic5_pldo,      "vdd-l7-l11"),
index 3325596..2c9e5ac 100644 (file)
@@ -1850,9 +1850,6 @@ qla2x00_port_speed_show(struct device *dev, struct device_attribute *attr,
                return -EINVAL;
        }
 
-       ql_log(ql_log_info, vha, 0x70d6,
-           "port speed:%d\n", ha->link_data_rate);
-
        return scnprintf(buf, PAGE_SIZE, "%s\n", spd[ha->link_data_rate]);
 }
 
index 3717eea..5f0ad8b 100644 (file)
@@ -80,6 +80,10 @@ static int scsi_dev_type_resume(struct device *dev,
        dev_dbg(dev, "scsi resume: %d\n", err);
 
        if (err == 0) {
+               bool was_runtime_suspended;
+
+               was_runtime_suspended = pm_runtime_suspended(dev);
+
                pm_runtime_disable(dev);
                err = pm_runtime_set_active(dev);
                pm_runtime_enable(dev);
@@ -93,8 +97,10 @@ static int scsi_dev_type_resume(struct device *dev,
                 */
                if (!err && scsi_is_sdev_device(dev)) {
                        struct scsi_device *sdev = to_scsi_device(dev);
-
-                       blk_set_runtime_active(sdev->request_queue);
+                       if (was_runtime_suspended)
+                               blk_post_runtime_resume(sdev->request_queue, 0);
+                       else
+                               blk_set_runtime_active(sdev->request_queue);
                }
        }
 
index 55c5114..4ffb334 100644 (file)
@@ -537,9 +537,9 @@ static void gb_tty_set_termios(struct tty_struct *tty,
        }
 
        if (C_CRTSCTS(tty) && C_BAUD(tty) != B0)
-               newline.flow_control |= GB_SERIAL_AUTO_RTSCTS_EN;
+               newline.flow_control = GB_SERIAL_AUTO_RTSCTS_EN;
        else
-               newline.flow_control &= ~GB_SERIAL_AUTO_RTSCTS_EN;
+               newline.flow_control = 0;
 
        if (memcmp(&gb_tty->line_coding, &newline, sizeof(newline))) {
                memcpy(&gb_tty->line_coding, &newline, sizeof(newline));
index 4b25a3a..ed40435 100644 (file)
@@ -130,17 +130,24 @@ static int ad2s1210_config_write(struct ad2s1210_state *st, u8 data)
 static int ad2s1210_config_read(struct ad2s1210_state *st,
                                unsigned char address)
 {
-       struct spi_transfer xfer = {
-               .len = 2,
-               .rx_buf = st->rx,
-               .tx_buf = st->tx,
+       struct spi_transfer xfers[] = {
+               {
+                       .len = 1,
+                       .rx_buf = &st->rx[0],
+                       .tx_buf = &st->tx[0],
+                       .cs_change = 1,
+               }, {
+                       .len = 1,
+                       .rx_buf = &st->rx[1],
+                       .tx_buf = &st->tx[1],
+               },
        };
        int ret = 0;
 
        ad2s1210_set_mode(MOD_CONFIG, st);
        st->tx[0] = address | AD2S1210_MSB_IS_HIGH;
        st->tx[1] = AD2S1210_REG_FAULT;
-       ret = spi_sync_transfer(st->sdev, &xfer, 1);
+       ret = spi_sync_transfer(st->sdev, xfers, 2);
        if (ret < 0)
                return ret;
 
index 7b00d70..358d7b2 100644 (file)
@@ -298,7 +298,6 @@ static int kp2000_pcie_probe(struct pci_dev *pdev,
 {
        int err = 0;
        struct kp2000_device *pcard;
-       int rv;
        unsigned long reg_bar_phys_addr;
        unsigned long reg_bar_phys_len;
        unsigned long dma_bar_phys_addr;
@@ -445,11 +444,11 @@ static int kp2000_pcie_probe(struct pci_dev *pdev,
        if (err < 0)
                goto err_release_dma;
 
-       rv = request_irq(pcard->pdev->irq, kp2000_irq_handler, IRQF_SHARED,
-                        pcard->name, pcard);
-       if (rv) {
+       err = request_irq(pcard->pdev->irq, kp2000_irq_handler, IRQF_SHARED,
+                         pcard->name, pcard);
+       if (err) {
                dev_err(&pcard->pdev->dev,
-                       "%s: failed to request_irq: %d\n", __func__, rv);
+                       "%s: failed to request_irq: %d\n", __func__, err);
                goto err_disable_msi;
        }
 
index 6e1e500..9aa1433 100644 (file)
@@ -57,8 +57,10 @@ static int send_scan_req(struct wfx_vif *wvif,
        wvif->scan_abort = false;
        reinit_completion(&wvif->scan_complete);
        timeout = hif_scan(wvif, req, start_idx, i - start_idx);
-       if (timeout < 0)
+       if (timeout < 0) {
+               wfx_tx_unlock(wvif->wdev);
                return timeout;
+       }
        ret = wait_for_completion_timeout(&wvif->scan_complete, timeout);
        if (req->channels[start_idx]->max_power != wvif->vif->bss_conf.txpower)
                hif_set_output_power(wvif, wvif->vif->bss_conf.txpower);
index 594b724..264a822 100644 (file)
@@ -3350,6 +3350,7 @@ static void target_tmr_work(struct work_struct *work)
 
        cmd->se_tfo->queue_tm_rsp(cmd);
 
+       transport_lun_remove_cmd(cmd);
        transport_cmd_check_stop_to_fabric(cmd);
        return;
 
index 13eadcb..0b5110d 100644 (file)
@@ -883,6 +883,7 @@ console_initcall(sifive_console_init);
 
 static void __ssp_add_console_port(struct sifive_serial_port *ssp)
 {
+       spin_lock_init(&ssp->port.lock);
        sifive_serial_console_ports[ssp->port.line] = ssp;
 }
 
index 7957d2d..01c456f 100644 (file)
@@ -89,15 +89,14 @@ static struct vdpasim *dev_to_sim(struct device *dev)
 static void vdpasim_queue_ready(struct vdpasim *vdpasim, unsigned int idx)
 {
        struct vdpasim_virtqueue *vq = &vdpasim->vqs[idx];
-       int ret;
 
-       ret = vringh_init_iotlb(&vq->vring, vdpasim_features,
-                               VDPASIM_QUEUE_MAX, false,
-                               (struct vring_desc *)(uintptr_t)vq->desc_addr,
-                               (struct vring_avail *)
-                               (uintptr_t)vq->driver_addr,
-                               (struct vring_used *)
-                               (uintptr_t)vq->device_addr);
+       vringh_init_iotlb(&vq->vring, vdpasim_features,
+                         VDPASIM_QUEUE_MAX, false,
+                         (struct vring_desc *)(uintptr_t)vq->desc_addr,
+                         (struct vring_avail *)
+                         (uintptr_t)vq->driver_addr,
+                         (struct vring_used *)
+                         (uintptr_t)vq->device_addr);
 }
 
 static void vdpasim_vq_reset(struct vdpasim_virtqueue *vq)
index d450e16..21a59b5 100644 (file)
@@ -730,7 +730,7 @@ static inline void __user *vhost_vq_meta_fetch(struct vhost_virtqueue *vq,
        if (!map)
                return NULL;
 
-       return (void *)(uintptr_t)(map->addr + addr - map->start);
+       return (void __user *)(uintptr_t)(map->addr + addr - map->start);
 }
 
 /* Can we switch to this memory table? */
@@ -869,7 +869,7 @@ static void __user *__vhost_get_user_slow(struct vhost_virtqueue *vq,
  * not happen in this case.
  */
 static inline void __user *__vhost_get_user(struct vhost_virtqueue *vq,
-                                           void *addr, unsigned int size,
+                                           void __user *addr, unsigned int size,
                                            int type)
 {
        void __user *uaddr = vhost_vq_meta_fetch(vq,
index a587767..37d1bba 100644 (file)
@@ -32,9 +32,8 @@ void afs_fileserver_probe_result(struct afs_call *call)
        struct afs_server *server = call->server;
        unsigned int server_index = call->server_index;
        unsigned int index = call->addr_ix;
-       unsigned int rtt = UINT_MAX;
+       unsigned int rtt_us = 0;
        bool have_result = false;
-       u64 _rtt;
        int ret = call->error;
 
        _enter("%pU,%u", &server->uuid, index);
@@ -93,15 +92,9 @@ responded:
                }
        }
 
-       /* Get the RTT and scale it to fit into a 32-bit value that represents
-        * over a minute of time so that we can access it with one instruction
-        * on a 32-bit system.
-        */
-       _rtt = rxrpc_kernel_get_rtt(call->net->socket, call->rxcall);
-       _rtt /= 64;
-       rtt = (_rtt > UINT_MAX) ? UINT_MAX : _rtt;
-       if (rtt < server->probe.rtt) {
-               server->probe.rtt = rtt;
+       rtt_us = rxrpc_kernel_get_srtt(call->net->socket, call->rxcall);
+       if (rtt_us < server->probe.rtt) {
+               server->probe.rtt = rtt_us;
                alist->preferred = index;
                have_result = true;
        }
@@ -113,8 +106,7 @@ out:
        spin_unlock(&server->probe_lock);
 
        _debug("probe [%u][%u] %pISpc rtt=%u ret=%d",
-              server_index, index, &alist->addrs[index].transport,
-              (unsigned int)rtt, ret);
+              server_index, index, &alist->addrs[index].transport, rtt_us, ret);
 
        have_result |= afs_fs_probe_done(server);
        if (have_result)
index 68fc466..d2b3798 100644 (file)
@@ -385,8 +385,6 @@ static int afs_deliver_fs_fetch_data(struct afs_call *call)
                ASSERTCMP(req->offset, <=, PAGE_SIZE);
                if (req->offset == PAGE_SIZE) {
                        req->offset = 0;
-                       if (req->page_done)
-                               req->page_done(req);
                        req->index++;
                        if (req->remain > 0)
                                goto begin_page;
@@ -440,11 +438,13 @@ static int afs_deliver_fs_fetch_data(struct afs_call *call)
                if (req->offset < PAGE_SIZE)
                        zero_user_segment(req->pages[req->index],
                                          req->offset, PAGE_SIZE);
-               if (req->page_done)
-                       req->page_done(req);
                req->offset = 0;
        }
 
+       if (req->page_done)
+               for (req->index = 0; req->index < req->nr_pages; req->index++)
+                       req->page_done(req);
+
        _leave(" = 0 [done]");
        return 0;
 }
index 858498c..e3aa013 100644 (file)
@@ -31,10 +31,9 @@ void afs_vlserver_probe_result(struct afs_call *call)
        struct afs_addr_list *alist = call->alist;
        struct afs_vlserver *server = call->vlserver;
        unsigned int server_index = call->server_index;
+       unsigned int rtt_us = 0;
        unsigned int index = call->addr_ix;
-       unsigned int rtt = UINT_MAX;
        bool have_result = false;
-       u64 _rtt;
        int ret = call->error;
 
        _enter("%s,%u,%u,%d,%d", server->name, server_index, index, ret, call->abort_code);
@@ -93,15 +92,9 @@ responded:
                }
        }
 
-       /* Get the RTT and scale it to fit into a 32-bit value that represents
-        * over a minute of time so that we can access it with one instruction
-        * on a 32-bit system.
-        */
-       _rtt = rxrpc_kernel_get_rtt(call->net->socket, call->rxcall);
-       _rtt /= 64;
-       rtt = (_rtt > UINT_MAX) ? UINT_MAX : _rtt;
-       if (rtt < server->probe.rtt) {
-               server->probe.rtt = rtt;
+       rtt_us = rxrpc_kernel_get_srtt(call->net->socket, call->rxcall);
+       if (rtt_us < server->probe.rtt) {
+               server->probe.rtt = rtt_us;
                alist->preferred = index;
                have_result = true;
        }
@@ -113,8 +106,7 @@ out:
        spin_unlock(&server->probe_lock);
 
        _debug("probe [%u][%u] %pISpc rtt=%u ret=%d",
-              server_index, index, &alist->addrs[index].transport,
-              (unsigned int)rtt, ret);
+              server_index, index, &alist->addrs[index].transport, rtt_us, ret);
 
        have_result |= afs_vl_probe_done(server);
        if (have_result) {
index b5b45c5..fe413e7 100644 (file)
@@ -497,8 +497,6 @@ static int yfs_deliver_fs_fetch_data64(struct afs_call *call)
                ASSERTCMP(req->offset, <=, PAGE_SIZE);
                if (req->offset == PAGE_SIZE) {
                        req->offset = 0;
-                       if (req->page_done)
-                               req->page_done(req);
                        req->index++;
                        if (req->remain > 0)
                                goto begin_page;
@@ -556,11 +554,13 @@ static int yfs_deliver_fs_fetch_data64(struct afs_call *call)
                if (req->offset < PAGE_SIZE)
                        zero_user_segment(req->pages[req->index],
                                          req->offset, PAGE_SIZE);
-               if (req->page_done)
-                       req->page_done(req);
                req->offset = 0;
        }
 
+       if (req->page_done)
+               for (req->index = 0; req->index < req->nr_pages; req->index++)
+                       req->page_done(req);
+
        _leave(" = 0 [done]");
        return 0;
 }
index 4f76764..c9db8eb 100644 (file)
@@ -348,12 +348,13 @@ out:
 }
 
 const struct file_operations exfat_file_operations = {
-       .llseek      = generic_file_llseek,
-       .read_iter   = generic_file_read_iter,
-       .write_iter  = generic_file_write_iter,
-       .mmap        = generic_file_mmap,
-       .fsync       = generic_file_fsync,
-       .splice_read = generic_file_splice_read,
+       .llseek         = generic_file_llseek,
+       .read_iter      = generic_file_read_iter,
+       .write_iter     = generic_file_write_iter,
+       .mmap           = generic_file_mmap,
+       .fsync          = generic_file_fsync,
+       .splice_read    = generic_file_splice_read,
+       .splice_write   = iter_file_splice_write,
 };
 
 const struct inode_operations exfat_file_inode_operations = {
index b72d782..a2659a8 100644 (file)
@@ -692,6 +692,7 @@ static int exfat_find(struct inode *dir, struct qstr *qname,
                        exfat_fs_error(sb,
                                "non-zero size file starts with zero cluster (size : %llu, p_dir : %u, entry : 0x%08x)",
                                i_size_read(dir), ei->dir.dir, ei->entry);
+                       kfree(es);
                        return -EIO;
                }
 
index 0565d55..a846ff5 100644 (file)
@@ -203,6 +203,12 @@ enum {
        Opt_errors,
        Opt_discard,
        Opt_time_offset,
+
+       /* Deprecated options */
+       Opt_utf8,
+       Opt_debug,
+       Opt_namecase,
+       Opt_codepage,
 };
 
 static const struct constant_table exfat_param_enums[] = {
@@ -223,6 +229,14 @@ static const struct fs_parameter_spec exfat_parameters[] = {
        fsparam_enum("errors",                  Opt_errors, exfat_param_enums),
        fsparam_flag("discard",                 Opt_discard),
        fsparam_s32("time_offset",              Opt_time_offset),
+       __fsparam(NULL, "utf8",                 Opt_utf8, fs_param_deprecated,
+                 NULL),
+       __fsparam(NULL, "debug",                Opt_debug, fs_param_deprecated,
+                 NULL),
+       __fsparam(fs_param_is_u32, "namecase",  Opt_namecase,
+                 fs_param_deprecated, NULL),
+       __fsparam(fs_param_is_u32, "codepage",  Opt_codepage,
+                 fs_param_deprecated, NULL),
        {}
 };
 
@@ -278,6 +292,11 @@ static int exfat_parse_param(struct fs_context *fc, struct fs_parameter *param)
                        return -EINVAL;
                opts->time_offset = result.int_32;
                break;
+       case Opt_utf8:
+       case Opt_debug:
+       case Opt_namecase:
+       case Opt_codepage:
+               break;
        default:
                return -EINVAL;
        }
index 91eb438..ad2dbf6 100644 (file)
@@ -722,7 +722,7 @@ enum {
 #define EXT4_MAX_BLOCK_FILE_PHYS       0xFFFFFFFF
 
 /* Max logical block we can support */
-#define EXT4_MAX_LOGICAL_BLOCK         0xFFFFFFFF
+#define EXT4_MAX_LOGICAL_BLOCK         0xFFFFFFFE
 
 /*
  * Structure of an inode on the disk
index f2b577b..2b4b945 100644 (file)
@@ -4832,6 +4832,28 @@ static const struct iomap_ops ext4_iomap_xattr_ops = {
        .iomap_begin            = ext4_iomap_xattr_begin,
 };
 
+static int ext4_fiemap_check_ranges(struct inode *inode, u64 start, u64 *len)
+{
+       u64 maxbytes;
+
+       if (ext4_test_inode_flag(inode, EXT4_INODE_EXTENTS))
+               maxbytes = inode->i_sb->s_maxbytes;
+       else
+               maxbytes = EXT4_SB(inode->i_sb)->s_bitmap_maxbytes;
+
+       if (*len == 0)
+               return -EINVAL;
+       if (start > maxbytes)
+               return -EFBIG;
+
+       /*
+        * Shrink request scope to what the fs can actually handle.
+        */
+       if (*len > maxbytes || (maxbytes - *len) < start)
+               *len = maxbytes - start;
+       return 0;
+}
+
 static int _ext4_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo,
                        __u64 start, __u64 len, bool from_es_cache)
 {
@@ -4852,6 +4874,15 @@ static int _ext4_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo,
        if (fiemap_check_flags(fieinfo, ext4_fiemap_flags))
                return -EBADR;
 
+       /*
+        * For bitmap files the maximum size limit could be smaller than
+        * s_maxbytes, so check len here manually instead of just relying on the
+        * generic check.
+        */
+       error = ext4_fiemap_check_ranges(inode, start, &len);
+       if (error)
+               return error;
+
        if (fieinfo->fi_flags & FIEMAP_FLAG_XATTR) {
                fieinfo->fi_flags &= ~FIEMAP_FLAG_XATTR;
                error = iomap_fiemap(inode, fieinfo, start, len,
index bfc1281..0746532 100644 (file)
@@ -733,29 +733,6 @@ static void ext4_fill_fsxattr(struct inode *inode, struct fsxattr *fa)
                fa->fsx_projid = from_kprojid(&init_user_ns, ei->i_projid);
 }
 
-/* copied from fs/ioctl.c */
-static int fiemap_check_ranges(struct super_block *sb,
-                              u64 start, u64 len, u64 *new_len)
-{
-       u64 maxbytes = (u64) sb->s_maxbytes;
-
-       *new_len = len;
-
-       if (len == 0)
-               return -EINVAL;
-
-       if (start > maxbytes)
-               return -EFBIG;
-
-       /*
-        * Shrink request scope to what the fs can actually handle.
-        */
-       if (len > maxbytes || (maxbytes - len) < start)
-               *new_len = maxbytes - start;
-
-       return 0;
-}
-
 /* So that the fiemap access checks can't overflow on 32 bit machines. */
 #define FIEMAP_MAX_EXTENTS     (UINT_MAX / sizeof(struct fiemap_extent))
 
@@ -765,8 +742,6 @@ static int ext4_ioctl_get_es_cache(struct file *filp, unsigned long arg)
        struct fiemap __user *ufiemap = (struct fiemap __user *) arg;
        struct fiemap_extent_info fieinfo = { 0, };
        struct inode *inode = file_inode(filp);
-       struct super_block *sb = inode->i_sb;
-       u64 len;
        int error;
 
        if (copy_from_user(&fiemap, ufiemap, sizeof(fiemap)))
@@ -775,11 +750,6 @@ static int ext4_ioctl_get_es_cache(struct file *filp, unsigned long arg)
        if (fiemap.fm_extent_count > FIEMAP_MAX_EXTENTS)
                return -EINVAL;
 
-       error = fiemap_check_ranges(sb, fiemap.fm_start, fiemap.fm_length,
-                                   &len);
-       if (error)
-               return error;
-
        fieinfo.fi_flags = fiemap.fm_flags;
        fieinfo.fi_extents_max = fiemap.fm_extent_count;
        fieinfo.fi_extents_start = ufiemap->fm_extents;
@@ -792,7 +762,8 @@ static int ext4_ioctl_get_es_cache(struct file *filp, unsigned long arg)
        if (fieinfo.fi_flags & FIEMAP_FLAG_SYNC)
                filemap_write_and_wait(inode->i_mapping);
 
-       error = ext4_get_es_cache(inode, &fieinfo, fiemap.fm_start, len);
+       error = ext4_get_es_cache(inode, &fieinfo, fiemap.fm_start,
+                       fiemap.fm_length);
        fiemap.fm_flags = fieinfo.fi_flags;
        fiemap.fm_mapped_extents = fieinfo.fi_extents_mapped;
        if (copy_to_user(ufiemap, &fiemap, sizeof(fiemap)))
index c8a4e4c..abb8b70 100644 (file)
--- a/fs/file.c
+++ b/fs/file.c
@@ -70,7 +70,7 @@ static void copy_fd_bitmaps(struct fdtable *nfdt, struct fdtable *ofdt,
  */
 static void copy_fdtable(struct fdtable *nfdt, struct fdtable *ofdt)
 {
-       unsigned int cpy, set;
+       size_t cpy, set;
 
        BUG_ON(nfdt->max_fds < ofdt->max_fds);
 
index 70ae7e8..bb25e39 100644 (file)
@@ -619,6 +619,8 @@ struct io_kiocb {
        bool                            needs_fixed_file;
        u8                              opcode;
 
+       u16                             buf_index;
+
        struct io_ring_ctx      *ctx;
        struct list_head        list;
        unsigned int            flags;
@@ -924,6 +926,7 @@ static struct io_ring_ctx *io_ring_ctx_alloc(struct io_uring_params *p)
                goto err;
 
        ctx->flags = p->flags;
+       init_waitqueue_head(&ctx->sqo_wait);
        init_waitqueue_head(&ctx->cq_wait);
        INIT_LIST_HEAD(&ctx->cq_overflow_list);
        init_completion(&ctx->completions[0]);
@@ -2100,9 +2103,7 @@ static int io_prep_rw(struct io_kiocb *req, const struct io_uring_sqe *sqe,
 
        req->rw.addr = READ_ONCE(sqe->addr);
        req->rw.len = READ_ONCE(sqe->len);
-       /* we own ->private, reuse it for the buffer index  / buffer ID */
-       req->rw.kiocb.private = (void *) (unsigned long)
-                                       READ_ONCE(sqe->buf_index);
+       req->buf_index = READ_ONCE(sqe->buf_index);
        return 0;
 }
 
@@ -2145,7 +2146,7 @@ static ssize_t io_import_fixed(struct io_kiocb *req, int rw,
        struct io_ring_ctx *ctx = req->ctx;
        size_t len = req->rw.len;
        struct io_mapped_ubuf *imu;
-       unsigned index, buf_index;
+       u16 index, buf_index;
        size_t offset;
        u64 buf_addr;
 
@@ -2153,7 +2154,7 @@ static ssize_t io_import_fixed(struct io_kiocb *req, int rw,
        if (unlikely(!ctx->user_bufs))
                return -EFAULT;
 
-       buf_index = (unsigned long) req->rw.kiocb.private;
+       buf_index = req->buf_index;
        if (unlikely(buf_index >= ctx->nr_user_bufs))
                return -EFAULT;
 
@@ -2269,10 +2270,10 @@ static void __user *io_rw_buffer_select(struct io_kiocb *req, size_t *len,
                                        bool needs_lock)
 {
        struct io_buffer *kbuf;
-       int bgid;
+       u16 bgid;
 
        kbuf = (struct io_buffer *) (unsigned long) req->rw.addr;
-       bgid = (int) (unsigned long) req->rw.kiocb.private;
+       bgid = req->buf_index;
        kbuf = io_buffer_select(req, len, bgid, kbuf, needs_lock);
        if (IS_ERR(kbuf))
                return kbuf;
@@ -2363,7 +2364,7 @@ static ssize_t io_import_iovec(int rw, struct io_kiocb *req,
        }
 
        /* buffer index only valid with fixed read/write, or buffer select  */
-       if (req->rw.kiocb.private && !(req->flags & REQ_F_BUFFER_SELECT))
+       if (req->buf_index && !(req->flags & REQ_F_BUFFER_SELECT))
                return -EINVAL;
 
        if (opcode == IORING_OP_READ || opcode == IORING_OP_WRITE) {
@@ -2771,11 +2772,8 @@ static int io_splice(struct io_kiocb *req, bool force_nonblock)
        poff_in = (sp->off_in == -1) ? NULL : &sp->off_in;
        poff_out = (sp->off_out == -1) ? NULL : &sp->off_out;
 
-       if (sp->len) {
+       if (sp->len)
                ret = do_splice(in, poff_in, out, poff_out, sp->len, flags);
-               if (force_nonblock && ret == -EAGAIN)
-                       return -EAGAIN;
-       }
 
        io_put_file(req, in, (sp->flags & SPLICE_F_FD_IN_FIXED));
        req->flags &= ~REQ_F_NEED_CLEANUP;
@@ -4137,12 +4135,14 @@ static int __io_async_wake(struct io_kiocb *req, struct io_poll_iocb *poll,
        req->result = mask;
        init_task_work(&req->task_work, func);
        /*
-        * If this fails, then the task is exiting. Punt to one of the io-wq
-        * threads to ensure the work gets run, we can't always rely on exit
-        * cancelation taking care of this.
+        * If this fails, then the task is exiting. When a task exits, the
+        * work gets canceled, so just cancel this request as well instead
+        * of executing it. We can't safely execute it anyway, as we may not
+        * have the needed state needed for it anyway.
         */
        ret = task_work_add(tsk, &req->task_work, true);
        if (unlikely(ret)) {
+               WRITE_ONCE(poll->canceled, true);
                tsk = io_wq_get_task(req->ctx->io_wq);
                task_work_add(tsk, &req->task_work, true);
        }
@@ -5013,12 +5013,13 @@ static int io_req_defer(struct io_kiocb *req, const struct io_uring_sqe *sqe)
        if (!req_need_defer(req) && list_empty_careful(&ctx->defer_list))
                return 0;
 
-       if (!req->io && io_alloc_async_ctx(req))
-               return -EAGAIN;
-
-       ret = io_req_defer_prep(req, sqe);
-       if (ret < 0)
-               return ret;
+       if (!req->io) {
+               if (io_alloc_async_ctx(req))
+                       return -EAGAIN;
+               ret = io_req_defer_prep(req, sqe);
+               if (ret < 0)
+                       return ret;
+       }
 
        spin_lock_irq(&ctx->completion_lock);
        if (!req_need_defer(req) && list_empty(&ctx->defer_list)) {
@@ -5305,7 +5306,8 @@ static int io_issue_sqe(struct io_kiocb *req, const struct io_uring_sqe *sqe,
        if (ret)
                return ret;
 
-       if (ctx->flags & IORING_SETUP_IOPOLL) {
+       /* If the op doesn't have a file, we're not polling for it */
+       if ((ctx->flags & IORING_SETUP_IOPOLL) && req->file) {
                const bool in_async = io_wq_current_is_worker();
 
                if (req->result == -EAGAIN)
@@ -5606,9 +5608,15 @@ fail_req:
                        io_double_put_req(req);
                }
        } else if (req->flags & REQ_F_FORCE_ASYNC) {
-               ret = io_req_defer_prep(req, sqe);
-               if (unlikely(ret < 0))
-                       goto fail_req;
+               if (!req->io) {
+                       ret = -EAGAIN;
+                       if (io_alloc_async_ctx(req))
+                               goto fail_req;
+                       ret = io_req_defer_prep(req, sqe);
+                       if (unlikely(ret < 0))
+                               goto fail_req;
+               }
+
                /*
                 * Never try inline submit of IOSQE_ASYNC is set, go straight
                 * to async execution.
@@ -6024,6 +6032,7 @@ static int io_sq_thread(void *data)
                                finish_wait(&ctx->sqo_wait, &wait);
 
                                ctx->rings->sq_flags &= ~IORING_SQ_NEED_WAKEUP;
+                               ret = 0;
                                continue;
                        }
                        finish_wait(&ctx->sqo_wait, &wait);
@@ -6837,7 +6846,6 @@ static int io_sq_offload_start(struct io_ring_ctx *ctx,
 {
        int ret;
 
-       init_waitqueue_head(&ctx->sqo_wait);
        mmgrab(current->mm);
        ctx->sqo_mm = current->mm;
 
index 475c61f..ed5c107 100644 (file)
@@ -783,6 +783,9 @@ static struct ovl_fh *ovl_fid_to_fh(struct fid *fid, int buflen, int fh_type)
        if (fh_type != OVL_FILEID_V0)
                return ERR_PTR(-EINVAL);
 
+       if (buflen <= OVL_FH_WIRE_OFFSET)
+               return ERR_PTR(-EINVAL);
+
        fh = kzalloc(buflen, GFP_KERNEL);
        if (!fh)
                return ERR_PTR(-ENOMEM);
index b0d42ec..981f11e 100644 (file)
@@ -58,6 +58,24 @@ int ovl_setattr(struct dentry *dentry, struct iattr *attr)
                if (attr->ia_valid & (ATTR_KILL_SUID|ATTR_KILL_SGID))
                        attr->ia_valid &= ~ATTR_MODE;
 
+               /*
+                * We might have to translate ovl file into real file object
+                * once use cases emerge.  For now, simply don't let underlying
+                * filesystem rely on attr->ia_file
+                */
+               attr->ia_valid &= ~ATTR_FILE;
+
+               /*
+                * If open(O_TRUNC) is done, VFS calls ->setattr with ATTR_OPEN
+                * set.  Overlayfs does not pass O_TRUNC flag to underlying
+                * filesystem during open -> do not pass ATTR_OPEN.  This
+                * disables optimization in fuse which assumes open(O_TRUNC)
+                * already set file size to 0.  But we never passed O_TRUNC to
+                * fuse.  So by clearing ATTR_OPEN, fuse will be forced to send
+                * setattr request to server.
+                */
+               attr->ia_valid &= ~ATTR_OPEN;
+
                inode_lock(upperdentry->d_inode);
                old_cred = ovl_override_creds(dentry->d_sb);
                err = notify_change(upperdentry, attr, NULL);
index fd0a1e7..4e53efb 100644 (file)
@@ -1494,7 +1494,7 @@ static int opipe_prep(struct pipe_inode_info *pipe, unsigned int flags)
         * Check pipe occupancy without the inode lock first. This function
         * is speculative anyways, so missing one is ok.
         */
-       if (pipe_full(pipe->head, pipe->tail, pipe->max_usage))
+       if (!pipe_full(pipe->head, pipe->tail, pipe->max_usage))
                return 0;
 
        ret = 0;
index 8cdbd53..f985a3f 100644 (file)
@@ -79,13 +79,9 @@ int ubifs_prepare_auth_node(struct ubifs_info *c, void *node,
                             struct shash_desc *inhash)
 {
        struct ubifs_auth_node *auth = node;
-       u8 *hash;
+       u8 hash[UBIFS_HASH_ARR_SZ];
        int err;
 
-       hash = kmalloc(crypto_shash_descsize(c->hash_tfm), GFP_NOFS);
-       if (!hash)
-               return -ENOMEM;
-
        {
                SHASH_DESC_ON_STACK(hash_desc, c->hash_tfm);
 
@@ -94,21 +90,16 @@ int ubifs_prepare_auth_node(struct ubifs_info *c, void *node,
 
                err = crypto_shash_final(hash_desc, hash);
                if (err)
-                       goto out;
+                       return err;
        }
 
        err = ubifs_hash_calc_hmac(c, hash, auth->hmac);
        if (err)
-               goto out;
+               return err;
 
        auth->ch.node_type = UBIFS_AUTH_NODE;
        ubifs_prepare_node(c, auth, ubifs_auth_node_sz(c), 0);
-
-       err = 0;
-out:
-       kfree(hash);
-
-       return err;
+       return 0;
 }
 
 static struct shash_desc *ubifs_get_desc(const struct ubifs_info *c,
index 743928e..49fe062 100644 (file)
@@ -1375,7 +1375,6 @@ int ubifs_update_time(struct inode *inode, struct timespec64 *time,
        struct ubifs_info *c = inode->i_sb->s_fs_info;
        struct ubifs_budget_req req = { .dirtied_ino = 1,
                        .dirtied_ino_d = ALIGN(ui->data_len, 8) };
-       int iflags = I_DIRTY_TIME;
        int err, release;
 
        if (!IS_ENABLED(CONFIG_UBIFS_ATIME_SUPPORT))
@@ -1393,11 +1392,8 @@ int ubifs_update_time(struct inode *inode, struct timespec64 *time,
        if (flags & S_MTIME)
                inode->i_mtime = *time;
 
-       if (!(inode->i_sb->s_flags & SB_LAZYTIME))
-               iflags |= I_DIRTY_SYNC;
-
        release = ui->dirty;
-       __mark_inode_dirty(inode, iflags);
+       __mark_inode_dirty(inode, I_DIRTY_SYNC);
        mutex_unlock(&ui->ui_mutex);
        if (release)
                ubifs_release_budget(c, &req);
index b28ac4d..01fcf79 100644 (file)
@@ -601,18 +601,12 @@ static int authenticate_sleb(struct ubifs_info *c, struct ubifs_scan_leb *sleb,
        struct ubifs_scan_node *snod;
        int n_nodes = 0;
        int err;
-       u8 *hash, *hmac;
+       u8 hash[UBIFS_HASH_ARR_SZ];
+       u8 hmac[UBIFS_HMAC_ARR_SZ];
 
        if (!ubifs_authenticated(c))
                return sleb->nodes_cnt;
 
-       hash = kmalloc(crypto_shash_descsize(c->hash_tfm), GFP_NOFS);
-       hmac = kmalloc(c->hmac_desc_len, GFP_NOFS);
-       if (!hash || !hmac) {
-               err = -ENOMEM;
-               goto out;
-       }
-
        list_for_each_entry(snod, &sleb->nodes, list) {
 
                n_nodes++;
@@ -662,9 +656,6 @@ static int authenticate_sleb(struct ubifs_info *c, struct ubifs_scan_leb *sleb,
                err = 0;
        }
 out:
-       kfree(hash);
-       kfree(hmac);
-
        return err ? err : n_nodes - n_not_auth;
 }
 
index 4f005d9..8537e92 100644 (file)
@@ -521,6 +521,15 @@ struct cper_sec_pcie {
        u8      aer_info[96];
 };
 
+/* Firmware Error Record Reference, UEFI v2.7 sec N.2.10  */
+struct cper_sec_fw_err_rec_ref {
+       u8 record_type;
+       u8 revision;
+       u8 reserved[6];
+       u64 record_identifier;
+       guid_t record_identifier_guid;
+};
+
 /* Reset to default packing */
 #pragma pack()
 
index 251f1f7..9430d01 100644 (file)
@@ -1245,4 +1245,6 @@ struct linux_efi_memreserve {
 
 void __init efi_arch_mem_reserve(phys_addr_t addr, u64 size);
 
+char *efi_systab_show_arch(char *str);
+
 #endif /* _LINUX_EFI_H */
index c5a9773..98ef73b 100644 (file)
@@ -29,7 +29,7 @@ struct i2c_mux_core {
 
        int num_adapters;
        int max_adapters;
-       struct i2c_adapter *adapter[0];
+       struct i2c_adapter *adapter[];
 };
 
 struct i2c_mux_core *i2c_mux_alloc(struct i2c_adapter *parent,
index 45d36ba..49d2905 100644 (file)
@@ -2,7 +2,7 @@
 /*
  * i2c.h - definitions for the Linux i2c bus interface
  * Copyright (C) 1995-2000 Simon G. Vogl
- * Copyright (C) 2013-2019 Wolfram Sang <wsa@the-dreams.de>
+ * Copyright (C) 2013-2019 Wolfram Sang <wsa@kernel.org>
  *
  * With some changes from Kyösti Mälkki <kmalkki@cc.hut.fi> and
  * Frodo Looijaard <frodol@dds.nl>
index 6f8f79e..8397b65 100644 (file)
@@ -213,6 +213,12 @@ enum mlx5_port_status {
        MLX5_PORT_DOWN      = 2,
 };
 
+enum mlx5_cmdif_state {
+       MLX5_CMDIF_STATE_UNINITIALIZED,
+       MLX5_CMDIF_STATE_UP,
+       MLX5_CMDIF_STATE_DOWN,
+};
+
 struct mlx5_cmd_first {
        __be32          data[4];
 };
@@ -258,6 +264,7 @@ struct mlx5_cmd_stats {
 struct mlx5_cmd {
        struct mlx5_nb    nb;
 
+       enum mlx5_cmdif_state   state;
        void           *cmd_alloc_buf;
        dma_addr_t      alloc_dma;
        int             alloc_size;
@@ -284,6 +291,7 @@ struct mlx5_cmd {
        struct semaphore sem;
        struct semaphore pages_sem;
        int     mode;
+       u16     allowed_opcode;
        struct mlx5_cmd_work_ent *ent_arr[MLX5_MAX_COMMANDS];
        struct dma_pool *pool;
        struct mlx5_cmd_debug dbg;
@@ -743,6 +751,7 @@ struct mlx5_cmd_work_ent {
        struct delayed_work     cb_timeout_work;
        void                   *context;
        int                     idx;
+       struct completion       handling;
        struct completion       done;
        struct mlx5_cmd        *cmd;
        struct work_struct      work;
@@ -874,10 +883,17 @@ mlx5_frag_buf_get_idx_last_contig_stride(struct mlx5_frag_buf_ctrl *fbc, u32 ix)
        return min_t(u32, last_frag_stride_idx - fbc->strides_offset, fbc->sz_m1);
 }
 
+enum {
+       CMD_ALLOWED_OPCODE_ALL,
+};
+
 int mlx5_cmd_init(struct mlx5_core_dev *dev);
 void mlx5_cmd_cleanup(struct mlx5_core_dev *dev);
+void mlx5_cmd_set_state(struct mlx5_core_dev *dev,
+                       enum mlx5_cmdif_state cmdif_state);
 void mlx5_cmd_use_events(struct mlx5_core_dev *dev);
 void mlx5_cmd_use_polling(struct mlx5_core_dev *dev);
+void mlx5_cmd_allowed_opcode(struct mlx5_core_dev *dev, u16 opcode);
 
 struct mlx5_async_ctx {
        struct mlx5_core_dev *dev;
index c24d764..124bd13 100644 (file)
@@ -75,7 +75,8 @@ static inline void tcf_tm_dump(struct tcf_t *dtm, const struct tcf_t *stm)
 {
        dtm->install = jiffies_to_clock_t(jiffies - stm->install);
        dtm->lastuse = jiffies_to_clock_t(jiffies - stm->lastuse);
-       dtm->firstuse = jiffies_to_clock_t(jiffies - stm->firstuse);
+       dtm->firstuse = stm->firstuse ?
+               jiffies_to_clock_t(jiffies - stm->firstuse) : 0;
        dtm->expires = jiffies_to_clock_t(stm->expires);
 }
 
index 04e97ba..ab98894 100644 (file)
@@ -59,7 +59,7 @@ bool rxrpc_kernel_abort_call(struct socket *, struct rxrpc_call *,
 void rxrpc_kernel_end_call(struct socket *, struct rxrpc_call *);
 void rxrpc_kernel_get_peer(struct socket *, struct rxrpc_call *,
                           struct sockaddr_rxrpc *);
-u64 rxrpc_kernel_get_rtt(struct socket *, struct rxrpc_call *);
+u32 rxrpc_kernel_get_srtt(struct socket *, struct rxrpc_call *);
 int rxrpc_kernel_charge_accept(struct socket *, rxrpc_notify_rx_t,
                               rxrpc_user_attach_call_t, unsigned long, gfp_t,
                               unsigned int);
index 59e0d4e..b219a8f 100644 (file)
@@ -257,7 +257,6 @@ struct fib_dump_filter {
        u32                     table_id;
        /* filter_set is an optimization that an entry is set */
        bool                    filter_set;
-       bool                    dump_all_families;
        bool                    dump_routes;
        bool                    dump_exceptions;
        unsigned char           protocol;
index 191fe44..ba9efdc 100644 (file)
@@ -1112,18 +1112,17 @@ TRACE_EVENT(rxrpc_rtt_tx,
 TRACE_EVENT(rxrpc_rtt_rx,
            TP_PROTO(struct rxrpc_call *call, enum rxrpc_rtt_rx_trace why,
                     rxrpc_serial_t send_serial, rxrpc_serial_t resp_serial,
-                    s64 rtt, u8 nr, s64 avg),
+                    u32 rtt, u32 rto),
 
-           TP_ARGS(call, why, send_serial, resp_serial, rtt, nr, avg),
+           TP_ARGS(call, why, send_serial, resp_serial, rtt, rto),
 
            TP_STRUCT__entry(
                    __field(unsigned int,               call            )
                    __field(enum rxrpc_rtt_rx_trace,    why             )
-                   __field(u8,                         nr              )
                    __field(rxrpc_serial_t,             send_serial     )
                    __field(rxrpc_serial_t,             resp_serial     )
-                   __field(s64,                        rtt             )
-                   __field(u64,                        avg             )
+                   __field(u32,                        rtt             )
+                   __field(u32,                        rto             )
                             ),
 
            TP_fast_assign(
@@ -1132,18 +1131,16 @@ TRACE_EVENT(rxrpc_rtt_rx,
                    __entry->send_serial = send_serial;
                    __entry->resp_serial = resp_serial;
                    __entry->rtt = rtt;
-                   __entry->nr = nr;
-                   __entry->avg = avg;
+                   __entry->rto = rto;
                           ),
 
-           TP_printk("c=%08x %s sr=%08x rr=%08x rtt=%lld nr=%u avg=%lld",
+           TP_printk("c=%08x %s sr=%08x rr=%08x rtt=%u rto=%u",
                      __entry->call,
                      __print_symbolic(__entry->why, rxrpc_rtt_rx_traces),
                      __entry->send_serial,
                      __entry->resp_serial,
                      __entry->rtt,
-                     __entry->nr,
-                     __entry->avg)
+                     __entry->rto)
            );
 
 TRACE_EVENT(rxrpc_timer,
@@ -1544,6 +1541,41 @@ TRACE_EVENT(rxrpc_notify_socket,
                      __entry->serial)
            );
 
+TRACE_EVENT(rxrpc_rx_discard_ack,
+           TP_PROTO(unsigned int debug_id, rxrpc_serial_t serial,
+                    rxrpc_seq_t first_soft_ack, rxrpc_seq_t call_ackr_first,
+                    rxrpc_seq_t prev_pkt, rxrpc_seq_t call_ackr_prev),
+
+           TP_ARGS(debug_id, serial, first_soft_ack, call_ackr_first,
+                   prev_pkt, call_ackr_prev),
+
+           TP_STRUCT__entry(
+                   __field(unsigned int,       debug_id        )
+                   __field(rxrpc_serial_t,     serial          )
+                   __field(rxrpc_seq_t,        first_soft_ack)
+                   __field(rxrpc_seq_t,        call_ackr_first)
+                   __field(rxrpc_seq_t,        prev_pkt)
+                   __field(rxrpc_seq_t,        call_ackr_prev)
+                            ),
+
+           TP_fast_assign(
+                   __entry->debug_id           = debug_id;
+                   __entry->serial             = serial;
+                   __entry->first_soft_ack     = first_soft_ack;
+                   __entry->call_ackr_first    = call_ackr_first;
+                   __entry->prev_pkt           = prev_pkt;
+                   __entry->call_ackr_prev     = call_ackr_prev;
+                          ),
+
+           TP_printk("c=%08x r=%08x %08x<%08x %08x<%08x",
+                     __entry->debug_id,
+                     __entry->serial,
+                     __entry->first_soft_ack,
+                     __entry->call_ackr_first,
+                     __entry->prev_pkt,
+                     __entry->call_ackr_prev)
+           );
+
 #endif /* _TRACE_RXRPC_H */
 
 /* This part must be outside protection */
index 2843bbb..4e6dee1 100644 (file)
@@ -623,9 +623,20 @@ static int bpf_map_mmap(struct file *filp, struct vm_area_struct *vma)
 
        mutex_lock(&map->freeze_mutex);
 
-       if ((vma->vm_flags & VM_WRITE) && map->frozen) {
-               err = -EPERM;
-               goto out;
+       if (vma->vm_flags & VM_WRITE) {
+               if (map->frozen) {
+                       err = -EPERM;
+                       goto out;
+               }
+               /* map is meant to be read-only, so do not allow mapping as
+                * writable, because it's possible to leak a writable page
+                * reference and allows user-space to still modify it after
+                * freezing, while verifier will assume contents do not change
+                */
+               if (map->map_flags & BPF_F_RDONLY_PROG) {
+                       err = -EACCES;
+                       goto out;
+               }
        }
 
        /* set default open/close callbacks */
index a562df5..239970b 100644 (file)
@@ -948,8 +948,8 @@ void proc_sched_show_task(struct task_struct *p, struct pid_namespace *ns,
        P(se.avg.util_est.enqueued);
 #endif
 #ifdef CONFIG_UCLAMP_TASK
-       __PS("uclamp.min", p->uclamp[UCLAMP_MIN].value);
-       __PS("uclamp.max", p->uclamp[UCLAMP_MAX].value);
+       __PS("uclamp.min", p->uclamp_req[UCLAMP_MIN].value);
+       __PS("uclamp.max", p->uclamp_req[UCLAMP_MAX].value);
        __PS("effective uclamp.min", uclamp_eff_value(p, UCLAMP_MIN));
        __PS("effective uclamp.max", uclamp_eff_value(p, UCLAMP_MAX));
 #endif
index 02f323b..538ba5d 100644 (file)
@@ -4774,7 +4774,6 @@ void unthrottle_cfs_rq(struct cfs_rq *cfs_rq)
        struct rq *rq = rq_of(cfs_rq);
        struct cfs_bandwidth *cfs_b = tg_cfs_bandwidth(cfs_rq->tg);
        struct sched_entity *se;
-       int enqueue = 1;
        long task_delta, idle_task_delta;
 
        se = cfs_rq->tg->se[cpu_of(rq)];
@@ -4798,26 +4797,44 @@ void unthrottle_cfs_rq(struct cfs_rq *cfs_rq)
        idle_task_delta = cfs_rq->idle_h_nr_running;
        for_each_sched_entity(se) {
                if (se->on_rq)
-                       enqueue = 0;
+                       break;
+               cfs_rq = cfs_rq_of(se);
+               enqueue_entity(cfs_rq, se, ENQUEUE_WAKEUP);
+
+               cfs_rq->h_nr_running += task_delta;
+               cfs_rq->idle_h_nr_running += idle_task_delta;
+
+               /* end evaluation on encountering a throttled cfs_rq */
+               if (cfs_rq_throttled(cfs_rq))
+                       goto unthrottle_throttle;
+       }
 
+       for_each_sched_entity(se) {
                cfs_rq = cfs_rq_of(se);
-               if (enqueue) {
-                       enqueue_entity(cfs_rq, se, ENQUEUE_WAKEUP);
-               } else {
-                       update_load_avg(cfs_rq, se, 0);
-                       se_update_runnable(se);
-               }
+
+               update_load_avg(cfs_rq, se, UPDATE_TG);
+               se_update_runnable(se);
 
                cfs_rq->h_nr_running += task_delta;
                cfs_rq->idle_h_nr_running += idle_task_delta;
 
+
+               /* end evaluation on encountering a throttled cfs_rq */
                if (cfs_rq_throttled(cfs_rq))
-                       break;
+                       goto unthrottle_throttle;
+
+               /*
+                * One parent has been throttled and cfs_rq removed from the
+                * list. Add it back to not break the leaf list.
+                */
+               if (throttled_hierarchy(cfs_rq))
+                       list_add_leaf_cfs_rq(cfs_rq);
        }
 
-       if (!se)
-               add_nr_running(rq, task_delta);
+       /* At this point se is NULL and we are at root level*/
+       add_nr_running(rq, task_delta);
 
+unthrottle_throttle:
        /*
         * The cfs_rq_throttled() breaks in the above iteration can result in
         * incomplete leaf list maintenance, resulting in triggering the
@@ -4826,7 +4843,8 @@ void unthrottle_cfs_rq(struct cfs_rq *cfs_rq)
        for_each_sched_entity(se) {
                cfs_rq = cfs_rq_of(se);
 
-               list_add_leaf_cfs_rq(cfs_rq);
+               if (list_add_leaf_cfs_rq(cfs_rq))
+                       break;
        }
 
        assert_list_leaf_cfs_rq(rq);
@@ -5479,6 +5497,13 @@ enqueue_task_fair(struct rq *rq, struct task_struct *p, int flags)
                /* end evaluation on encountering a throttled cfs_rq */
                if (cfs_rq_throttled(cfs_rq))
                        goto enqueue_throttle;
+
+               /*
+                * One parent has been throttled and cfs_rq removed from the
+                * list. Add it back to not break the leaf list.
+                */
+               if (throttled_hierarchy(cfs_rq))
+                       list_add_leaf_cfs_rq(cfs_rq);
        }
 
 enqueue_throttle:
index 2d9f520..6b1622f 100644 (file)
@@ -214,6 +214,7 @@ test_string(void)
 #define PTR_STR "ffff0123456789ab"
 #define PTR_VAL_NO_CRNG "(____ptrval____)"
 #define ZEROS "00000000"       /* hex 32 zero bits */
+#define ONES "ffffffff"                /* hex 32 one bits */
 
 static int __init
 plain_format(void)
@@ -245,6 +246,7 @@ plain_format(void)
 #define PTR_STR "456789ab"
 #define PTR_VAL_NO_CRNG "(ptrval)"
 #define ZEROS ""
+#define ONES ""
 
 static int __init
 plain_format(void)
@@ -330,14 +332,28 @@ test_hashed(const char *fmt, const void *p)
        test(buf, fmt, p);
 }
 
+/*
+ * NULL pointers aren't hashed.
+ */
 static void __init
 null_pointer(void)
 {
-       test_hashed("%p", NULL);
+       test(ZEROS "00000000", "%p", NULL);
        test(ZEROS "00000000", "%px", NULL);
        test("(null)", "%pE", NULL);
 }
 
+/*
+ * Error pointers aren't hashed.
+ */
+static void __init
+error_pointer(void)
+{
+       test(ONES "fffffff5", "%p", ERR_PTR(-11));
+       test(ONES "fffffff5", "%px", ERR_PTR(-11));
+       test("(efault)", "%pE", ERR_PTR(-11));
+}
+
 #define PTR_INVALID ((void *)0x000000ab)
 
 static void __init
@@ -649,6 +665,7 @@ test_pointer(void)
 {
        plain();
        null_pointer();
+       error_pointer();
        invalid_pointer();
        symbol_ptr();
        kernel_ptr();
index 532b660..7c47ad5 100644 (file)
@@ -794,6 +794,13 @@ static char *ptr_to_id(char *buf, char *end, const void *ptr,
        unsigned long hashval;
        int ret;
 
+       /*
+        * Print the real pointer value for NULL and error pointers,
+        * as they are not actual addresses.
+        */
+       if (IS_ERR_OR_NULL(ptr))
+               return pointer_string(buf, end, ptr, spec);
+
        /* When debugging early boot use non-cryptographically secure hash. */
        if (unlikely(debug_boot_weak_hash)) {
                hashval = hash_long((unsigned long)ptr, 32);
index 434d503..de31218 100644 (file)
@@ -15,14 +15,14 @@ CFLAGS_REMOVE_tags_report.o = $(CC_FLAGS_FTRACE)
 
 # Function splitter causes unnecessary splits in __asan_load1/__asan_store1
 # see: https://gcc.gnu.org/bugzilla/show_bug.cgi?id=63533
-CFLAGS_common.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector)
-CFLAGS_generic.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector)
-CFLAGS_generic_report.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector)
-CFLAGS_init.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector)
-CFLAGS_quarantine.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector)
-CFLAGS_report.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector)
-CFLAGS_tags.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector)
-CFLAGS_tags_report.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector)
+CFLAGS_common.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector) -DDISABLE_BRANCH_PROFILING
+CFLAGS_generic.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector) -DDISABLE_BRANCH_PROFILING
+CFLAGS_generic_report.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector) -DDISABLE_BRANCH_PROFILING
+CFLAGS_init.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector) -DDISABLE_BRANCH_PROFILING
+CFLAGS_quarantine.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector) -DDISABLE_BRANCH_PROFILING
+CFLAGS_report.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector) -DDISABLE_BRANCH_PROFILING
+CFLAGS_tags.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector) -DDISABLE_BRANCH_PROFILING
+CFLAGS_tags_report.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector) -DDISABLE_BRANCH_PROFILING
 
 obj-$(CONFIG_KASAN) := common.o init.o report.o
 obj-$(CONFIG_KASAN_GENERIC) += generic.o generic_report.o quarantine.o
index 56ff888..098a7db 100644 (file)
@@ -15,7 +15,6 @@
  */
 
 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
-#define DISABLE_BRANCH_PROFILING
 
 #include <linux/export.h>
 #include <linux/interrupt.h>
index 25b7734..8a959fd 100644 (file)
@@ -12,7 +12,6 @@
  */
 
 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
-#define DISABLE_BRANCH_PROFILING
 
 #include <linux/export.h>
 #include <linux/interrupt.h>
index 42f31c4..8c3bb5e 100644 (file)
@@ -318,16 +318,16 @@ static inline void free_handle(unsigned long handle)
        slots = handle_to_slots(handle);
        write_lock(&slots->lock);
        *(unsigned long *)handle = 0;
-       write_unlock(&slots->lock);
-       if (zhdr->slots == slots)
+       if (zhdr->slots == slots) {
+               write_unlock(&slots->lock);
                return; /* simple case, nothing else to do */
+       }
 
        /* we are freeing a foreign handle if we are here */
        zhdr->foreign_handles--;
        is_free = true;
-       read_lock(&slots->lock);
        if (!test_bit(HANDLES_ORPHANED, &slots->pool)) {
-               read_unlock(&slots->lock);
+               write_unlock(&slots->lock);
                return;
        }
        for (i = 0; i <= BUDDY_MASK; i++) {
@@ -336,7 +336,7 @@ static inline void free_handle(unsigned long handle)
                        break;
                }
        }
-       read_unlock(&slots->lock);
+       write_unlock(&slots->lock);
 
        if (is_free) {
                struct z3fold_pool *pool = slots_to_pool(slots);
@@ -422,6 +422,7 @@ static struct z3fold_header *init_z3fold_page(struct page *page, bool headless,
        zhdr->start_middle = 0;
        zhdr->cpu = -1;
        zhdr->foreign_handles = 0;
+       zhdr->mapped_count = 0;
        zhdr->slots = slots;
        zhdr->pool = pool;
        INIT_LIST_HEAD(&zhdr->buddy);
index ff57ea8..fd91cd3 100644 (file)
@@ -635,8 +635,10 @@ static int ax25_setsockopt(struct socket *sock, int level, int optname,
                break;
 
        case SO_BINDTODEVICE:
-               if (optlen > IFNAMSIZ)
-                       optlen = IFNAMSIZ;
+               if (optlen > IFNAMSIZ - 1)
+                       optlen = IFNAMSIZ - 1;
+
+               memset(devname, 0, sizeof(devname));
 
                if (copy_from_user(devname, optval, optlen)) {
                        res = -EFAULT;
index 6d327b7..2d8acee 100644 (file)
@@ -4988,11 +4988,12 @@ static inline int nf_ingress(struct sk_buff *skb, struct packet_type **pt_prev,
        return 0;
 }
 
-static int __netif_receive_skb_core(struct sk_buff *skb, bool pfmemalloc,
+static int __netif_receive_skb_core(struct sk_buff **pskb, bool pfmemalloc,
                                    struct packet_type **ppt_prev)
 {
        struct packet_type *ptype, *pt_prev;
        rx_handler_func_t *rx_handler;
+       struct sk_buff *skb = *pskb;
        struct net_device *orig_dev;
        bool deliver_exact = false;
        int ret = NET_RX_DROP;
@@ -5023,8 +5024,10 @@ another_round:
                ret2 = do_xdp_generic(rcu_dereference(skb->dev->xdp_prog), skb);
                preempt_enable();
 
-               if (ret2 != XDP_PASS)
-                       return NET_RX_DROP;
+               if (ret2 != XDP_PASS) {
+                       ret = NET_RX_DROP;
+                       goto out;
+               }
                skb_reset_mac_len(skb);
        }
 
@@ -5174,6 +5177,13 @@ drop:
        }
 
 out:
+       /* The invariant here is that if *ppt_prev is not NULL
+        * then skb should also be non-NULL.
+        *
+        * Apparently *ppt_prev assignment above holds this invariant due to
+        * skb dereferencing near it.
+        */
+       *pskb = skb;
        return ret;
 }
 
@@ -5183,7 +5193,7 @@ static int __netif_receive_skb_one_core(struct sk_buff *skb, bool pfmemalloc)
        struct packet_type *pt_prev = NULL;
        int ret;
 
-       ret = __netif_receive_skb_core(skb, pfmemalloc, &pt_prev);
+       ret = __netif_receive_skb_core(&skb, pfmemalloc, &pt_prev);
        if (pt_prev)
                ret = INDIRECT_CALL_INET(pt_prev->func, ipv6_rcv, ip_rcv, skb,
                                         skb->dev, pt_prev, orig_dev);
@@ -5261,7 +5271,7 @@ static void __netif_receive_skb_list_core(struct list_head *head, bool pfmemallo
                struct packet_type *pt_prev = NULL;
 
                skb_list_del_init(skb);
-               __netif_receive_skb_core(skb, pfmemalloc, &pt_prev);
+               __netif_receive_skb_core(&skb, pfmemalloc, &pt_prev);
                if (!pt_prev)
                        continue;
                if (pt_curr != pt_prev || od_curr != orig_dev) {
index 3eff848..5dceed4 100644 (file)
@@ -160,12 +160,10 @@ out:
        return ret;
 }
 
-int skb_flow_dissector_bpf_prog_detach(const union bpf_attr *attr)
+static int flow_dissector_bpf_prog_detach(struct net *net)
 {
        struct bpf_prog *attached;
-       struct net *net;
 
-       net = current->nsproxy->net_ns;
        mutex_lock(&flow_dissector_mutex);
        attached = rcu_dereference_protected(net->flow_dissector_prog,
                                             lockdep_is_held(&flow_dissector_mutex));
@@ -179,6 +177,24 @@ int skb_flow_dissector_bpf_prog_detach(const union bpf_attr *attr)
        return 0;
 }
 
+int skb_flow_dissector_bpf_prog_detach(const union bpf_attr *attr)
+{
+       return flow_dissector_bpf_prog_detach(current->nsproxy->net_ns);
+}
+
+static void __net_exit flow_dissector_pernet_pre_exit(struct net *net)
+{
+       /* We're not racing with attach/detach because there are no
+        * references to netns left when pre_exit gets called.
+        */
+       if (rcu_access_pointer(net->flow_dissector_prog))
+               flow_dissector_bpf_prog_detach(net);
+}
+
+static struct pernet_operations flow_dissector_pernet_ops __net_initdata = {
+       .pre_exit = flow_dissector_pernet_pre_exit,
+};
+
 /**
  * __skb_flow_get_ports - extract the upper layer ports and return them
  * @skb: sk_buff to extract the ports from
@@ -1836,7 +1852,7 @@ static int __init init_default_flow_dissectors(void)
        skb_flow_dissector_init(&flow_keys_basic_dissector,
                                flow_keys_basic_dissector_keys,
                                ARRAY_SIZE(flow_keys_basic_dissector_keys));
-       return 0;
-}
 
+       return register_pernet_subsys(&flow_dissector_pernet_ops);
+}
 core_initcall(init_default_flow_dissectors);
index b5705cb..d6619ed 100644 (file)
@@ -15,6 +15,7 @@
 #define MTK_HDR_XMIT_TAGGED_TPID_8100  1
 #define MTK_HDR_RECV_SOURCE_PORT_MASK  GENMASK(2, 0)
 #define MTK_HDR_XMIT_DP_BIT_MASK       GENMASK(5, 0)
+#define MTK_HDR_XMIT_SA_DIS            BIT(6)
 
 static struct sk_buff *mtk_tag_xmit(struct sk_buff *skb,
                                    struct net_device *dev)
@@ -22,6 +23,9 @@ static struct sk_buff *mtk_tag_xmit(struct sk_buff *skb,
        struct dsa_port *dp = dsa_slave_to_port(dev);
        u8 *mtk_tag;
        bool is_vlan_skb = true;
+       unsigned char *dest = eth_hdr(skb)->h_dest;
+       bool is_multicast_skb = is_multicast_ether_addr(dest) &&
+                               !is_broadcast_ether_addr(dest);
 
        /* Build the special tag after the MAC Source Address. If VLAN header
         * is present, it's required that VLAN header and special tag is
@@ -47,6 +51,10 @@ static struct sk_buff *mtk_tag_xmit(struct sk_buff *skb,
                     MTK_HDR_XMIT_UNTAGGED;
        mtk_tag[1] = (1 << dp->index) & MTK_HDR_XMIT_DP_BIT_MASK;
 
+       /* Disable SA learning for multicast frames */
+       if (unlikely(is_multicast_skb))
+               mtk_tag[1] |= MTK_HDR_XMIT_SA_DIS;
+
        /* Tag control information is kept for 802.1Q */
        if (!is_vlan_skb) {
                mtk_tag[2] = 0;
@@ -61,6 +69,9 @@ static struct sk_buff *mtk_tag_rcv(struct sk_buff *skb, struct net_device *dev,
 {
        int port;
        __be16 *phdr, hdr;
+       unsigned char *dest = eth_hdr(skb)->h_dest;
+       bool is_multicast_skb = is_multicast_ether_addr(dest) &&
+                               !is_broadcast_ether_addr(dest);
 
        if (unlikely(!pskb_may_pull(skb, MTK_HDR_LEN)))
                return NULL;
@@ -86,6 +97,10 @@ static struct sk_buff *mtk_tag_rcv(struct sk_buff *skb, struct net_device *dev,
        if (!skb->dev)
                return NULL;
 
+       /* Only unicast or broadcast frames are offloaded */
+       if (likely(!is_multicast_skb))
+               skb->offload_fwd_mark = 1;
+
        return skb;
 }
 
index 0c77231..ed53572 100644 (file)
@@ -342,7 +342,7 @@ static int ethnl_default_doit(struct sk_buff *skb, struct genl_info *info)
        ret = ops->reply_size(req_info, reply_data);
        if (ret < 0)
                goto err_cleanup;
-       reply_len = ret;
+       reply_len = ret + ethnl_reply_header_size();
        ret = -ENOMEM;
        rskb = ethnl_reply_init(reply_len, req_info->dev, ops->reply_cmd,
                                ops->hdr_attr, info, &reply_payload);
@@ -588,7 +588,7 @@ static void ethnl_default_notify(struct net_device *dev, unsigned int cmd,
        ret = ops->reply_size(req_info, reply_data);
        if (ret < 0)
                goto err_cleanup;
-       reply_len = ret;
+       reply_len = ret + ethnl_reply_header_size();
        ret = -ENOMEM;
        skb = genlmsg_new(reply_len, GFP_KERNEL);
        if (!skb)
index 95eae5c..0eed4e4 100644 (file)
@@ -324,7 +324,6 @@ static int strset_reply_size(const struct ethnl_req_info *req_base,
        int len = 0;
        int ret;
 
-       len += ethnl_reply_header_size();
        for (i = 0; i < ETH_SS_COUNT; i++) {
                const struct strset_info *set_info = &data->sets[i];
 
index 213be9c..1bf9da3 100644 (file)
@@ -918,7 +918,6 @@ int ip_valid_fib_dump_req(struct net *net, const struct nlmsghdr *nlh,
        else
                filter->dump_exceptions = false;
 
-       filter->dump_all_families = (rtm->rtm_family == AF_UNSPEC);
        filter->flags    = rtm->rtm_flags;
        filter->protocol = rtm->rtm_protocol;
        filter->rt_type  = rtm->rtm_type;
@@ -990,7 +989,7 @@ static int inet_dump_fib(struct sk_buff *skb, struct netlink_callback *cb)
        if (filter.table_id) {
                tb = fib_get_table(net, filter.table_id);
                if (!tb) {
-                       if (filter.dump_all_families)
+                       if (rtnl_msg_family(cb->nlh) != PF_INET)
                                return skb->len;
 
                        NL_SET_ERR_MSG(cb->extack, "ipv4: FIB table does not exist");
index 5f34eb9..65c29f2 100644 (file)
 #include <net/addrconf.h>
 
 #if IS_ENABLED(CONFIG_IPV6)
-/* match_wildcard == true:  IPV6_ADDR_ANY equals to any IPv6 addresses if IPv6
- *                          only, and any IPv4 addresses if not IPv6 only
- * match_wildcard == false: addresses must be exactly the same, i.e.
- *                          IPV6_ADDR_ANY only equals to IPV6_ADDR_ANY,
- *                          and 0.0.0.0 equals to 0.0.0.0 only
+/* match_sk*_wildcard == true:  IPV6_ADDR_ANY equals to any IPv6 addresses
+ *                             if IPv6 only, and any IPv4 addresses
+ *                             if not IPv6 only
+ * match_sk*_wildcard == false: addresses must be exactly the same, i.e.
+ *                             IPV6_ADDR_ANY only equals to IPV6_ADDR_ANY,
+ *                             and 0.0.0.0 equals to 0.0.0.0 only
  */
 static bool ipv6_rcv_saddr_equal(const struct in6_addr *sk1_rcv_saddr6,
                                 const struct in6_addr *sk2_rcv_saddr6,
                                 __be32 sk1_rcv_saddr, __be32 sk2_rcv_saddr,
                                 bool sk1_ipv6only, bool sk2_ipv6only,
-                                bool match_wildcard)
+                                bool match_sk1_wildcard,
+                                bool match_sk2_wildcard)
 {
        int addr_type = ipv6_addr_type(sk1_rcv_saddr6);
        int addr_type2 = sk2_rcv_saddr6 ? ipv6_addr_type(sk2_rcv_saddr6) : IPV6_ADDR_MAPPED;
@@ -44,8 +46,8 @@ static bool ipv6_rcv_saddr_equal(const struct in6_addr *sk1_rcv_saddr6,
                if (!sk2_ipv6only) {
                        if (sk1_rcv_saddr == sk2_rcv_saddr)
                                return true;
-                       if (!sk1_rcv_saddr || !sk2_rcv_saddr)
-                               return match_wildcard;
+                       return (match_sk1_wildcard && !sk1_rcv_saddr) ||
+                               (match_sk2_wildcard && !sk2_rcv_saddr);
                }
                return false;
        }
@@ -53,11 +55,11 @@ static bool ipv6_rcv_saddr_equal(const struct in6_addr *sk1_rcv_saddr6,
        if (addr_type == IPV6_ADDR_ANY && addr_type2 == IPV6_ADDR_ANY)
                return true;
 
-       if (addr_type2 == IPV6_ADDR_ANY && match_wildcard &&
+       if (addr_type2 == IPV6_ADDR_ANY && match_sk2_wildcard &&
            !(sk2_ipv6only && addr_type == IPV6_ADDR_MAPPED))
                return true;
 
-       if (addr_type == IPV6_ADDR_ANY && match_wildcard &&
+       if (addr_type == IPV6_ADDR_ANY && match_sk1_wildcard &&
            !(sk1_ipv6only && addr_type2 == IPV6_ADDR_MAPPED))
                return true;
 
@@ -69,18 +71,19 @@ static bool ipv6_rcv_saddr_equal(const struct in6_addr *sk1_rcv_saddr6,
 }
 #endif
 
-/* match_wildcard == true:  0.0.0.0 equals to any IPv4 addresses
- * match_wildcard == false: addresses must be exactly the same, i.e.
- *                          0.0.0.0 only equals to 0.0.0.0
+/* match_sk*_wildcard == true:  0.0.0.0 equals to any IPv4 addresses
+ * match_sk*_wildcard == false: addresses must be exactly the same, i.e.
+ *                             0.0.0.0 only equals to 0.0.0.0
  */
 static bool ipv4_rcv_saddr_equal(__be32 sk1_rcv_saddr, __be32 sk2_rcv_saddr,
-                                bool sk2_ipv6only, bool match_wildcard)
+                                bool sk2_ipv6only, bool match_sk1_wildcard,
+                                bool match_sk2_wildcard)
 {
        if (!sk2_ipv6only) {
                if (sk1_rcv_saddr == sk2_rcv_saddr)
                        return true;
-               if (!sk1_rcv_saddr || !sk2_rcv_saddr)
-                       return match_wildcard;
+               return (match_sk1_wildcard && !sk1_rcv_saddr) ||
+                       (match_sk2_wildcard && !sk2_rcv_saddr);
        }
        return false;
 }
@@ -96,10 +99,12 @@ bool inet_rcv_saddr_equal(const struct sock *sk, const struct sock *sk2,
                                            sk2->sk_rcv_saddr,
                                            ipv6_only_sock(sk),
                                            ipv6_only_sock(sk2),
+                                           match_wildcard,
                                            match_wildcard);
 #endif
        return ipv4_rcv_saddr_equal(sk->sk_rcv_saddr, sk2->sk_rcv_saddr,
-                                   ipv6_only_sock(sk2), match_wildcard);
+                                   ipv6_only_sock(sk2), match_wildcard,
+                                   match_wildcard);
 }
 EXPORT_SYMBOL(inet_rcv_saddr_equal);
 
@@ -285,10 +290,10 @@ static inline int sk_reuseport_match(struct inet_bind_bucket *tb,
                                            tb->fast_rcv_saddr,
                                            sk->sk_rcv_saddr,
                                            tb->fast_ipv6_only,
-                                           ipv6_only_sock(sk), true);
+                                           ipv6_only_sock(sk), true, false);
 #endif
        return ipv4_rcv_saddr_equal(tb->fast_rcv_saddr, sk->sk_rcv_saddr,
-                                   ipv6_only_sock(sk), true);
+                                   ipv6_only_sock(sk), true, false);
 }
 
 /* Obtain a reference to a local port for the given sock,
index 2f01cf6..678575a 100644 (file)
@@ -698,7 +698,7 @@ out:
 
 rtnl_link_failed:
 #if IS_ENABLED(CONFIG_MPLS)
-       xfrm4_tunnel_deregister(&mplsip_handler, AF_INET);
+       xfrm4_tunnel_deregister(&mplsip_handler, AF_MPLS);
 xfrm_tunnel_mplsip_failed:
 
 #endif
index 5c218db..b2363b8 100644 (file)
@@ -2613,7 +2613,7 @@ static int ipmr_rtm_dumproute(struct sk_buff *skb, struct netlink_callback *cb)
 
                mrt = ipmr_get_table(sock_net(skb->sk), filter.table_id);
                if (!mrt) {
-                       if (filter.dump_all_families)
+                       if (rtnl_msg_family(cb->nlh) != RTNL_FAMILY_IPMR)
                                return skb->len;
 
                        NL_SET_ERR_MSG(cb->extack, "ipv4: MR table does not exist");
index fdfca53..715e144 100644 (file)
@@ -276,6 +276,7 @@ out:
        return 0;
 
 nla_put_failure:
+       nlmsg_cancel(skb, nlh);
        return -EMSGSIZE;
 }
 
@@ -433,7 +434,7 @@ static int nh_check_attr_group(struct net *net, struct nlattr *tb[],
                if (!valid_group_nh(nh, len, extack))
                        return -EINVAL;
        }
-       for (i = NHA_GROUP + 1; i < __NHA_MAX; ++i) {
+       for (i = NHA_GROUP_TYPE + 1; i < __NHA_MAX; ++i) {
                if (!tb[i])
                        continue;
 
index fa829f3..b73f540 100644 (file)
@@ -491,18 +491,16 @@ u32 ip_idents_reserve(u32 hash, int segs)
        atomic_t *p_id = ip_idents + hash % IP_IDENTS_SZ;
        u32 old = READ_ONCE(*p_tstamp);
        u32 now = (u32)jiffies;
-       u32 new, delta = 0;
+       u32 delta = 0;
 
        if (old != now && cmpxchg(p_tstamp, old, now) == old)
                delta = prandom_u32_max(now - old);
 
-       /* Do not use atomic_add_return() as it makes UBSAN unhappy */
-       do {
-               old = (u32)atomic_read(p_id);
-               new = old + delta + segs;
-       } while (atomic_cmpxchg(p_id, old, new) != old);
-
-       return new - segs;
+       /* If UBSAN reports an error there, please make sure your compiler
+        * supports -fno-strict-overflow before reporting it that was a bug
+        * in UBSAN, and it has been fixed in GCC-8.
+        */
+       return atomic_add_return(segs + delta, p_id) - segs;
 }
 EXPORT_SYMBOL(ip_idents_reserve);
 
index 46ed567..2031489 100644 (file)
@@ -664,7 +664,7 @@ static int inet6_dump_fib(struct sk_buff *skb, struct netlink_callback *cb)
        if (arg.filter.table_id) {
                tb = fib6_get_table(net, arg.filter.table_id);
                if (!tb) {
-                       if (arg.filter.dump_all_families)
+                       if (rtnl_msg_family(cb->nlh) != PF_INET6)
                                goto out;
 
                        NL_SET_ERR_MSG_MOD(cb->extack, "FIB table does not exist");
index 65a54d7..1f4d20e 100644 (file)
@@ -98,7 +98,8 @@ static void ipmr_expire_process(struct timer_list *t);
 #ifdef CONFIG_IPV6_MROUTE_MULTIPLE_TABLES
 #define ip6mr_for_each_table(mrt, net) \
        list_for_each_entry_rcu(mrt, &net->ipv6.mr6_tables, list, \
-                               lockdep_rtnl_is_held())
+                               lockdep_rtnl_is_held() || \
+                               list_empty(&net->ipv6.mr6_tables))
 
 static struct mr_table *ip6mr_mr_table_iter(struct net *net,
                                            struct mr_table *mrt)
@@ -2502,7 +2503,7 @@ static int ip6mr_rtm_dumproute(struct sk_buff *skb, struct netlink_callback *cb)
 
                mrt = ip6mr_get_table(sock_net(skb->sk), filter.table_id);
                if (!mrt) {
-                       if (filter.dump_all_families)
+                       if (rtnl_msg_family(cb->nlh) != RTNL_FAMILY_IP6MR)
                                return skb->len;
 
                        NL_SET_ERR_MSG_MOD(cb->extack, "MR table does not exist");
index c151628..0f5a414 100644 (file)
@@ -47,8 +47,6 @@ void mptcp_crypto_key_sha(u64 key, u32 *token, u64 *idsn)
 void mptcp_crypto_hmac_sha(u64 key1, u64 key2, u8 *msg, int len, void *hmac)
 {
        u8 input[SHA256_BLOCK_SIZE + SHA256_DIGEST_SIZE];
-       __be32 mptcp_hashed_key[SHA256_DIGEST_WORDS];
-       __be32 *hash_out = (__force __be32 *)hmac;
        struct sha256_state state;
        u8 key1be[8];
        u8 key2be[8];
@@ -86,11 +84,7 @@ void mptcp_crypto_hmac_sha(u64 key1, u64 key2, u8 *msg, int len, void *hmac)
 
        sha256_init(&state);
        sha256_update(&state, input, SHA256_BLOCK_SIZE + SHA256_DIGEST_SIZE);
-       sha256_final(&state, (u8 *)mptcp_hashed_key);
-
-       /* takes only first 160 bits */
-       for (i = 0; i < 5; i++)
-               hash_out[i] = mptcp_hashed_key[i];
+       sha256_final(&state, (u8 *)hmac);
 }
 
 #ifdef CONFIG_MPTCP_HMAC_TEST
@@ -101,29 +95,29 @@ struct test_cast {
 };
 
 /* we can't reuse RFC 4231 test vectors, as we have constraint on the
- * input and key size, and we truncate the output.
+ * input and key size.
  */
 static struct test_cast tests[] = {
        {
                .key = "0b0b0b0b0b0b0b0b",
                .msg = "48692054",
-               .result = "8385e24fb4235ac37556b6b886db106284a1da67",
+               .result = "8385e24fb4235ac37556b6b886db106284a1da671699f46db1f235ec622dcafa",
        },
        {
                .key = "aaaaaaaaaaaaaaaa",
                .msg = "dddddddd",
-               .result = "2c5e219164ff1dca1c4a92318d847bb6b9d44492",
+               .result = "2c5e219164ff1dca1c4a92318d847bb6b9d44492984e1eb71aff9022f71046e9",
        },
        {
                .key = "0102030405060708",
                .msg = "cdcdcdcd",
-               .result = "e73b9ba9969969cefb04aa0d6df18ec2fcc075b6",
+               .result = "e73b9ba9969969cefb04aa0d6df18ec2fcc075b6f23b4d8c4da736a5dbbc6e7d",
        },
 };
 
 static int __init test_mptcp_crypto(void)
 {
-       char hmac[20], hmac_hex[41];
+       char hmac[32], hmac_hex[65];
        u32 nonce1, nonce2;
        u64 key1, key2;
        u8 msg[8];
@@ -140,11 +134,11 @@ static int __init test_mptcp_crypto(void)
                put_unaligned_be32(nonce2, &msg[4]);
 
                mptcp_crypto_hmac_sha(key1, key2, msg, 8, hmac);
-               for (j = 0; j < 20; ++j)
+               for (j = 0; j < 32; ++j)
                        sprintf(&hmac_hex[j << 1], "%02x", hmac[j] & 0xff);
-               hmac_hex[40] = 0;
+               hmac_hex[64] = 0;
 
-               if (memcmp(hmac_hex, tests[i].result, 40))
+               if (memcmp(hmac_hex, tests[i].result, 64))
                        pr_err("test %d failed, got %s expected %s", i,
                               hmac_hex, tests[i].result);
                else
index 45497af..7793b60 100644 (file)
@@ -7,6 +7,7 @@
 #define pr_fmt(fmt) "MPTCP: " fmt
 
 #include <linux/kernel.h>
+#include <crypto/sha.h>
 #include <net/tcp.h>
 #include <net/mptcp.h>
 #include "protocol.h"
@@ -535,7 +536,7 @@ static bool mptcp_established_options_dss(struct sock *sk, struct sk_buff *skb,
 static u64 add_addr_generate_hmac(u64 key1, u64 key2, u8 addr_id,
                                  struct in_addr *addr)
 {
-       u8 hmac[MPTCP_ADDR_HMAC_LEN];
+       u8 hmac[SHA256_DIGEST_SIZE];
        u8 msg[7];
 
        msg[0] = addr_id;
@@ -545,14 +546,14 @@ static u64 add_addr_generate_hmac(u64 key1, u64 key2, u8 addr_id,
 
        mptcp_crypto_hmac_sha(key1, key2, msg, 7, hmac);
 
-       return get_unaligned_be64(hmac);
+       return get_unaligned_be64(&hmac[SHA256_DIGEST_SIZE - sizeof(u64)]);
 }
 
 #if IS_ENABLED(CONFIG_MPTCP_IPV6)
 static u64 add_addr6_generate_hmac(u64 key1, u64 key2, u8 addr_id,
                                   struct in6_addr *addr)
 {
-       u8 hmac[MPTCP_ADDR_HMAC_LEN];
+       u8 hmac[SHA256_DIGEST_SIZE];
        u8 msg[19];
 
        msg[0] = addr_id;
@@ -562,7 +563,7 @@ static u64 add_addr6_generate_hmac(u64 key1, u64 key2, u8 addr_id,
 
        mptcp_crypto_hmac_sha(key1, key2, msg, 19, hmac);
 
-       return get_unaligned_be64(hmac);
+       return get_unaligned_be64(&hmac[SHA256_DIGEST_SIZE - sizeof(u64)]);
 }
 #endif
 
index e4ca632..d0803df 100644 (file)
@@ -81,7 +81,6 @@
 
 /* MPTCP ADD_ADDR flags */
 #define MPTCP_ADDR_ECHO                BIT(0)
-#define MPTCP_ADDR_HMAC_LEN    20
 #define MPTCP_ADDR_IPVERSION_4 4
 #define MPTCP_ADDR_IPVERSION_6 6
 
index 4931a29..8968b2c 100644 (file)
@@ -10,6 +10,7 @@
 #include <linux/module.h>
 #include <linux/netdevice.h>
 #include <crypto/algapi.h>
+#include <crypto/sha.h>
 #include <net/sock.h>
 #include <net/inet_common.h>
 #include <net/inet_hashtables.h>
@@ -89,7 +90,7 @@ static bool subflow_token_join_request(struct request_sock *req,
                                       const struct sk_buff *skb)
 {
        struct mptcp_subflow_request_sock *subflow_req = mptcp_subflow_rsk(req);
-       u8 hmac[MPTCPOPT_HMAC_LEN];
+       u8 hmac[SHA256_DIGEST_SIZE];
        struct mptcp_sock *msk;
        int local_id;
 
@@ -201,7 +202,7 @@ static void subflow_v6_init_req(struct request_sock *req,
 /* validate received truncated hmac and create hmac for third ACK */
 static bool subflow_thmac_valid(struct mptcp_subflow_context *subflow)
 {
-       u8 hmac[MPTCPOPT_HMAC_LEN];
+       u8 hmac[SHA256_DIGEST_SIZE];
        u64 thmac;
 
        subflow_generate_hmac(subflow->remote_key, subflow->local_key,
@@ -267,6 +268,8 @@ static void subflow_finish_connect(struct sock *sk, const struct sk_buff *skb)
                        subflow->ssn_offset = TCP_SKB_CB(skb)->seq;
                }
        } else if (subflow->mp_join) {
+               u8 hmac[SHA256_DIGEST_SIZE];
+
                pr_debug("subflow=%p, thmac=%llu, remote_nonce=%u",
                         subflow, subflow->thmac,
                         subflow->remote_nonce);
@@ -279,7 +282,9 @@ static void subflow_finish_connect(struct sock *sk, const struct sk_buff *skb)
                subflow_generate_hmac(subflow->local_key, subflow->remote_key,
                                      subflow->local_nonce,
                                      subflow->remote_nonce,
-                                     subflow->hmac);
+                                     hmac);
+
+               memcpy(subflow->hmac, hmac, MPTCPOPT_HMAC_LEN);
 
                if (skb)
                        subflow->ssn_offset = TCP_SKB_CB(skb)->seq;
@@ -347,7 +352,7 @@ static bool subflow_hmac_valid(const struct request_sock *req,
                               const struct mptcp_options_received *mp_opt)
 {
        const struct mptcp_subflow_request_sock *subflow_req;
-       u8 hmac[MPTCPOPT_HMAC_LEN];
+       u8 hmac[SHA256_DIGEST_SIZE];
        struct mptcp_sock *msk;
        bool ret;
 
@@ -361,7 +366,7 @@ static bool subflow_hmac_valid(const struct request_sock *req,
                              subflow_req->local_nonce, hmac);
 
        ret = true;
-       if (crypto_memneq(hmac, mp_opt->hmac, sizeof(hmac)))
+       if (crypto_memneq(hmac, mp_opt->hmac, MPTCPOPT_HMAC_LEN))
                ret = false;
 
        sock_put((struct sock *)msk);
index 7ed31b5..2d8d613 100644 (file)
@@ -854,7 +854,7 @@ static int qrtr_bcast_enqueue(struct qrtr_node *node, struct sk_buff *skb,
        }
        mutex_unlock(&qrtr_node_lock);
 
-       qrtr_local_enqueue(node, skb, type, from, to);
+       qrtr_local_enqueue(NULL, skb, type, from, to);
 
        return 0;
 }
index 6ffb7e9..ddd0f95 100644 (file)
@@ -25,6 +25,7 @@ rxrpc-y := \
        peer_event.o \
        peer_object.o \
        recvmsg.o \
+       rtt.o \
        security.o \
        sendmsg.o \
        skbuff.o \
index 3eb1ab4..9fe264b 100644 (file)
@@ -7,6 +7,7 @@
 
 #include <linux/atomic.h>
 #include <linux/seqlock.h>
+#include <linux/win_minmax.h>
 #include <net/net_namespace.h>
 #include <net/netns/generic.h>
 #include <net/sock.h>
@@ -311,11 +312,14 @@ struct rxrpc_peer {
 #define RXRPC_RTT_CACHE_SIZE 32
        spinlock_t              rtt_input_lock; /* RTT lock for input routine */
        ktime_t                 rtt_last_req;   /* Time of last RTT request */
-       u64                     rtt;            /* Current RTT estimate (in nS) */
-       u64                     rtt_sum;        /* Sum of cache contents */
-       u64                     rtt_cache[RXRPC_RTT_CACHE_SIZE]; /* Determined RTT cache */
-       u8                      rtt_cursor;     /* next entry at which to insert */
-       u8                      rtt_usage;      /* amount of cache actually used */
+       unsigned int            rtt_count;      /* Number of samples we've got */
+
+       u32                     srtt_us;        /* smoothed round trip time << 3 in usecs */
+       u32                     mdev_us;        /* medium deviation                     */
+       u32                     mdev_max_us;    /* maximal mdev for the last rtt period */
+       u32                     rttvar_us;      /* smoothed mdev_max                    */
+       u32                     rto_j;          /* Retransmission timeout in jiffies */
+       u8                      backoff;        /* Backoff timeout */
 
        u8                      cong_cwnd;      /* Congestion window size */
 };
@@ -1041,7 +1045,6 @@ extern unsigned long rxrpc_idle_ack_delay;
 extern unsigned int rxrpc_rx_window_size;
 extern unsigned int rxrpc_rx_mtu;
 extern unsigned int rxrpc_rx_jumbo_max;
-extern unsigned long rxrpc_resend_timeout;
 
 extern const s8 rxrpc_ack_priority[];
 
@@ -1069,8 +1072,6 @@ void rxrpc_send_keepalive(struct rxrpc_peer *);
  * peer_event.c
  */
 void rxrpc_error_report(struct sock *);
-void rxrpc_peer_add_rtt(struct rxrpc_call *, enum rxrpc_rtt_rx_trace,
-                       rxrpc_serial_t, rxrpc_serial_t, ktime_t, ktime_t);
 void rxrpc_peer_keepalive_worker(struct work_struct *);
 
 /*
@@ -1102,6 +1103,14 @@ extern const struct seq_operations rxrpc_peer_seq_ops;
 void rxrpc_notify_socket(struct rxrpc_call *);
 int rxrpc_recvmsg(struct socket *, struct msghdr *, size_t, int);
 
+/*
+ * rtt.c
+ */
+void rxrpc_peer_add_rtt(struct rxrpc_call *, enum rxrpc_rtt_rx_trace,
+                       rxrpc_serial_t, rxrpc_serial_t, ktime_t, ktime_t);
+unsigned long rxrpc_get_rto_backoff(struct rxrpc_peer *, bool);
+void rxrpc_peer_init_rtt(struct rxrpc_peer *);
+
 /*
  * rxkad.c
  */
index 70e44ab..b7611cc 100644 (file)
@@ -248,7 +248,7 @@ static void rxrpc_send_ping(struct rxrpc_call *call, struct sk_buff *skb)
        struct rxrpc_skb_priv *sp = rxrpc_skb(skb);
        ktime_t now = skb->tstamp;
 
-       if (call->peer->rtt_usage < 3 ||
+       if (call->peer->rtt_count < 3 ||
            ktime_before(ktime_add_ms(call->peer->rtt_last_req, 1000), now))
                rxrpc_propose_ACK(call, RXRPC_ACK_PING, sp->hdr.serial,
                                  true, true,
index cedbbb3..2a65ac4 100644 (file)
@@ -111,8 +111,8 @@ static void __rxrpc_propose_ACK(struct rxrpc_call *call, u8 ack_reason,
        } else {
                unsigned long now = jiffies, ack_at;
 
-               if (call->peer->rtt_usage > 0)
-                       ack_at = nsecs_to_jiffies(call->peer->rtt);
+               if (call->peer->srtt_us != 0)
+                       ack_at = usecs_to_jiffies(call->peer->srtt_us >> 3);
                else
                        ack_at = expiry;
 
@@ -157,24 +157,18 @@ static void rxrpc_congestion_timeout(struct rxrpc_call *call)
 static void rxrpc_resend(struct rxrpc_call *call, unsigned long now_j)
 {
        struct sk_buff *skb;
-       unsigned long resend_at;
+       unsigned long resend_at, rto_j;
        rxrpc_seq_t cursor, seq, top;
-       ktime_t now, max_age, oldest, ack_ts, timeout, min_timeo;
+       ktime_t now, max_age, oldest, ack_ts;
        int ix;
        u8 annotation, anno_type, retrans = 0, unacked = 0;
 
        _enter("{%d,%d}", call->tx_hard_ack, call->tx_top);
 
-       if (call->peer->rtt_usage > 1)
-               timeout = ns_to_ktime(call->peer->rtt * 3 / 2);
-       else
-               timeout = ms_to_ktime(rxrpc_resend_timeout);
-       min_timeo = ns_to_ktime((1000000000 / HZ) * 4);
-       if (ktime_before(timeout, min_timeo))
-               timeout = min_timeo;
+       rto_j = call->peer->rto_j;
 
        now = ktime_get_real();
-       max_age = ktime_sub(now, timeout);
+       max_age = ktime_sub(now, jiffies_to_usecs(rto_j));
 
        spin_lock_bh(&call->lock);
 
@@ -219,7 +213,7 @@ static void rxrpc_resend(struct rxrpc_call *call, unsigned long now_j)
        }
 
        resend_at = nsecs_to_jiffies(ktime_to_ns(ktime_sub(now, oldest)));
-       resend_at += jiffies + rxrpc_resend_timeout;
+       resend_at += jiffies + rto_j;
        WRITE_ONCE(call->resend_at, resend_at);
 
        if (unacked)
@@ -234,7 +228,7 @@ static void rxrpc_resend(struct rxrpc_call *call, unsigned long now_j)
                                        rxrpc_timer_set_for_resend);
                spin_unlock_bh(&call->lock);
                ack_ts = ktime_sub(now, call->acks_latest_ts);
-               if (ktime_to_ns(ack_ts) < call->peer->rtt)
+               if (ktime_to_us(ack_ts) < (call->peer->srtt_us >> 3))
                        goto out;
                rxrpc_propose_ACK(call, RXRPC_ACK_PING, 0, true, false,
                                  rxrpc_propose_ack_ping_for_lost_ack);
index 69e09d6..3be4177 100644 (file)
@@ -91,11 +91,11 @@ static void rxrpc_congestion_management(struct rxrpc_call *call,
                /* We analyse the number of packets that get ACK'd per RTT
                 * period and increase the window if we managed to fill it.
                 */
-               if (call->peer->rtt_usage == 0)
+               if (call->peer->rtt_count == 0)
                        goto out;
                if (ktime_before(skb->tstamp,
-                                ktime_add_ns(call->cong_tstamp,
-                                             call->peer->rtt)))
+                                ktime_add_us(call->cong_tstamp,
+                                             call->peer->srtt_us >> 3)))
                        goto out_no_clear_ca;
                change = rxrpc_cong_rtt_window_end;
                call->cong_tstamp = skb->tstamp;
@@ -802,6 +802,30 @@ static void rxrpc_input_soft_acks(struct rxrpc_call *call, u8 *acks,
        }
 }
 
+/*
+ * Return true if the ACK is valid - ie. it doesn't appear to have regressed
+ * with respect to the ack state conveyed by preceding ACKs.
+ */
+static bool rxrpc_is_ack_valid(struct rxrpc_call *call,
+                              rxrpc_seq_t first_pkt, rxrpc_seq_t prev_pkt)
+{
+       rxrpc_seq_t base = READ_ONCE(call->ackr_first_seq);
+
+       if (after(first_pkt, base))
+               return true; /* The window advanced */
+
+       if (before(first_pkt, base))
+               return false; /* firstPacket regressed */
+
+       if (after_eq(prev_pkt, call->ackr_prev_seq))
+               return true; /* previousPacket hasn't regressed. */
+
+       /* Some rx implementations put a serial number in previousPacket. */
+       if (after_eq(prev_pkt, base + call->tx_winsize))
+               return false;
+       return true;
+}
+
 /*
  * Process an ACK packet.
  *
@@ -865,9 +889,12 @@ static void rxrpc_input_ack(struct rxrpc_call *call, struct sk_buff *skb)
        }
 
        /* Discard any out-of-order or duplicate ACKs (outside lock). */
-       if (before(first_soft_ack, call->ackr_first_seq) ||
-           before(prev_pkt, call->ackr_prev_seq))
+       if (!rxrpc_is_ack_valid(call, first_soft_ack, prev_pkt)) {
+               trace_rxrpc_rx_discard_ack(call->debug_id, sp->hdr.serial,
+                                          first_soft_ack, call->ackr_first_seq,
+                                          prev_pkt, call->ackr_prev_seq);
                return;
+       }
 
        buf.info.rxMTU = 0;
        ioffset = offset + nr_acks + 3;
@@ -878,9 +905,12 @@ static void rxrpc_input_ack(struct rxrpc_call *call, struct sk_buff *skb)
        spin_lock(&call->input_lock);
 
        /* Discard any out-of-order or duplicate ACKs (inside lock). */
-       if (before(first_soft_ack, call->ackr_first_seq) ||
-           before(prev_pkt, call->ackr_prev_seq))
+       if (!rxrpc_is_ack_valid(call, first_soft_ack, prev_pkt)) {
+               trace_rxrpc_rx_discard_ack(call->debug_id, sp->hdr.serial,
+                                          first_soft_ack, call->ackr_first_seq,
+                                          prev_pkt, call->ackr_prev_seq);
                goto out;
+       }
        call->acks_latest_ts = skb->tstamp;
 
        call->ackr_first_seq = first_soft_ack;
index 214405f..d4144fd 100644 (file)
@@ -63,11 +63,6 @@ unsigned int rxrpc_rx_mtu = 5692;
  */
 unsigned int rxrpc_rx_jumbo_max = 4;
 
-/*
- * Time till packet resend (in milliseconds).
- */
-unsigned long rxrpc_resend_timeout = 4 * HZ;
-
 const s8 rxrpc_ack_priority[] = {
        [0]                             = 0,
        [RXRPC_ACK_DELAY]               = 1,
index 90e263c..f8b632a 100644 (file)
@@ -369,7 +369,7 @@ int rxrpc_send_data_packet(struct rxrpc_call *call, struct sk_buff *skb,
            (test_and_clear_bit(RXRPC_CALL_EV_ACK_LOST, &call->events) ||
             retrans ||
             call->cong_mode == RXRPC_CALL_SLOW_START ||
-            (call->peer->rtt_usage < 3 && sp->hdr.seq & 1) ||
+            (call->peer->rtt_count < 3 && sp->hdr.seq & 1) ||
             ktime_before(ktime_add_ms(call->peer->rtt_last_req, 1000),
                          ktime_get_real())))
                whdr.flags |= RXRPC_REQUEST_ACK;
@@ -423,13 +423,10 @@ done:
                if (whdr.flags & RXRPC_REQUEST_ACK) {
                        call->peer->rtt_last_req = skb->tstamp;
                        trace_rxrpc_rtt_tx(call, rxrpc_rtt_tx_data, serial);
-                       if (call->peer->rtt_usage > 1) {
+                       if (call->peer->rtt_count > 1) {
                                unsigned long nowj = jiffies, ack_lost_at;
 
-                               ack_lost_at = nsecs_to_jiffies(2 * call->peer->rtt);
-                               if (ack_lost_at < 1)
-                                       ack_lost_at = 1;
-
+                               ack_lost_at = rxrpc_get_rto_backoff(call->peer, retrans);
                                ack_lost_at += nowj;
                                WRITE_ONCE(call->ack_lost_at, ack_lost_at);
                                rxrpc_reduce_call_timer(call, ack_lost_at, nowj,
index 923b263..b1449d9 100644 (file)
@@ -295,52 +295,6 @@ static void rxrpc_distribute_error(struct rxrpc_peer *peer, int error,
        }
 }
 
-/*
- * Add RTT information to cache.  This is called in softirq mode and has
- * exclusive access to the peer RTT data.
- */
-void rxrpc_peer_add_rtt(struct rxrpc_call *call, enum rxrpc_rtt_rx_trace why,
-                       rxrpc_serial_t send_serial, rxrpc_serial_t resp_serial,
-                       ktime_t send_time, ktime_t resp_time)
-{
-       struct rxrpc_peer *peer = call->peer;
-       s64 rtt;
-       u64 sum = peer->rtt_sum, avg;
-       u8 cursor = peer->rtt_cursor, usage = peer->rtt_usage;
-
-       rtt = ktime_to_ns(ktime_sub(resp_time, send_time));
-       if (rtt < 0)
-               return;
-
-       spin_lock(&peer->rtt_input_lock);
-
-       /* Replace the oldest datum in the RTT buffer */
-       sum -= peer->rtt_cache[cursor];
-       sum += rtt;
-       peer->rtt_cache[cursor] = rtt;
-       peer->rtt_cursor = (cursor + 1) & (RXRPC_RTT_CACHE_SIZE - 1);
-       peer->rtt_sum = sum;
-       if (usage < RXRPC_RTT_CACHE_SIZE) {
-               usage++;
-               peer->rtt_usage = usage;
-       }
-
-       spin_unlock(&peer->rtt_input_lock);
-
-       /* Now recalculate the average */
-       if (usage == RXRPC_RTT_CACHE_SIZE) {
-               avg = sum / RXRPC_RTT_CACHE_SIZE;
-       } else {
-               avg = sum;
-               do_div(avg, usage);
-       }
-
-       /* Don't need to update this under lock */
-       peer->rtt = avg;
-       trace_rxrpc_rtt_rx(call, why, send_serial, resp_serial, rtt,
-                          usage, avg);
-}
-
 /*
  * Perform keep-alive pings.
  */
index 452163e..ca29976 100644 (file)
@@ -225,6 +225,8 @@ struct rxrpc_peer *rxrpc_alloc_peer(struct rxrpc_local *local, gfp_t gfp)
                spin_lock_init(&peer->rtt_input_lock);
                peer->debug_id = atomic_inc_return(&rxrpc_debug_id);
 
+               rxrpc_peer_init_rtt(peer);
+
                if (RXRPC_TX_SMSS > 2190)
                        peer->cong_cwnd = 2;
                else if (RXRPC_TX_SMSS > 1095)
@@ -497,14 +499,14 @@ void rxrpc_kernel_get_peer(struct socket *sock, struct rxrpc_call *call,
 EXPORT_SYMBOL(rxrpc_kernel_get_peer);
 
 /**
- * rxrpc_kernel_get_rtt - Get a call's peer RTT
+ * rxrpc_kernel_get_srtt - Get a call's peer smoothed RTT
  * @sock: The socket on which the call is in progress.
  * @call: The call to query
  *
- * Get the call's peer RTT.
+ * Get the call's peer smoothed RTT.
  */
-u64 rxrpc_kernel_get_rtt(struct socket *sock, struct rxrpc_call *call)
+u32 rxrpc_kernel_get_srtt(struct socket *sock, struct rxrpc_call *call)
 {
-       return call->peer->rtt;
+       return call->peer->srtt_us >> 3;
 }
-EXPORT_SYMBOL(rxrpc_kernel_get_rtt);
+EXPORT_SYMBOL(rxrpc_kernel_get_srtt);
index b9d053e..8b179e3 100644 (file)
@@ -222,7 +222,7 @@ static int rxrpc_peer_seq_show(struct seq_file *seq, void *v)
                seq_puts(seq,
                         "Proto Local                                          "
                         " Remote                                         "
-                        " Use CW  MTU   LastUse          RTT Rc\n"
+                        " Use  CW   MTU LastUse      RTT      RTO\n"
                         );
                return 0;
        }
@@ -236,15 +236,15 @@ static int rxrpc_peer_seq_show(struct seq_file *seq, void *v)
        now = ktime_get_seconds();
        seq_printf(seq,
                   "UDP   %-47.47s %-47.47s %3u"
-                  " %3u %5u %6llus %12llu %2u\n",
+                  " %3u %5u %6llus %8u %8u\n",
                   lbuff,
                   rbuff,
                   atomic_read(&peer->usage),
                   peer->cong_cwnd,
                   peer->mtu,
                   now - peer->last_tx_at,
-                  peer->rtt,
-                  peer->rtt_cursor);
+                  peer->srtt_us >> 3,
+                  jiffies_to_usecs(peer->rto_j));
 
        return 0;
 }
diff --git a/net/rxrpc/rtt.c b/net/rxrpc/rtt.c
new file mode 100644 (file)
index 0000000..928d8b3
--- /dev/null
@@ -0,0 +1,195 @@
+// SPDX-License-Identifier: GPL-2.0
+/* RTT/RTO calculation.
+ *
+ * Adapted from TCP for AF_RXRPC by David Howells (dhowells@redhat.com)
+ *
+ * https://tools.ietf.org/html/rfc6298
+ * https://tools.ietf.org/html/rfc1122#section-4.2.3.1
+ * http://ccr.sigcomm.org/archive/1995/jan95/ccr-9501-partridge87.pdf
+ */
+
+#include <linux/net.h>
+#include "ar-internal.h"
+
+#define RXRPC_RTO_MAX  ((unsigned)(120 * HZ))
+#define RXRPC_TIMEOUT_INIT ((unsigned)(1*HZ))  /* RFC6298 2.1 initial RTO value        */
+#define rxrpc_jiffies32 ((u32)jiffies)         /* As rxrpc_jiffies32 */
+#define rxrpc_min_rtt_wlen 300                 /* As sysctl_tcp_min_rtt_wlen */
+
+static u32 rxrpc_rto_min_us(struct rxrpc_peer *peer)
+{
+       return 200;
+}
+
+static u32 __rxrpc_set_rto(const struct rxrpc_peer *peer)
+{
+       return _usecs_to_jiffies((peer->srtt_us >> 3) + peer->rttvar_us);
+}
+
+static u32 rxrpc_bound_rto(u32 rto)
+{
+       return min(rto, RXRPC_RTO_MAX);
+}
+
+/*
+ * Called to compute a smoothed rtt estimate. The data fed to this
+ * routine either comes from timestamps, or from segments that were
+ * known _not_ to have been retransmitted [see Karn/Partridge
+ * Proceedings SIGCOMM 87]. The algorithm is from the SIGCOMM 88
+ * piece by Van Jacobson.
+ * NOTE: the next three routines used to be one big routine.
+ * To save cycles in the RFC 1323 implementation it was better to break
+ * it up into three procedures. -- erics
+ */
+static void rxrpc_rtt_estimator(struct rxrpc_peer *peer, long sample_rtt_us)
+{
+       long m = sample_rtt_us; /* RTT */
+       u32 srtt = peer->srtt_us;
+
+       /*      The following amusing code comes from Jacobson's
+        *      article in SIGCOMM '88.  Note that rtt and mdev
+        *      are scaled versions of rtt and mean deviation.
+        *      This is designed to be as fast as possible
+        *      m stands for "measurement".
+        *
+        *      On a 1990 paper the rto value is changed to:
+        *      RTO = rtt + 4 * mdev
+        *
+        * Funny. This algorithm seems to be very broken.
+        * These formulae increase RTO, when it should be decreased, increase
+        * too slowly, when it should be increased quickly, decrease too quickly
+        * etc. I guess in BSD RTO takes ONE value, so that it is absolutely
+        * does not matter how to _calculate_ it. Seems, it was trap
+        * that VJ failed to avoid. 8)
+        */
+       if (srtt != 0) {
+               m -= (srtt >> 3);       /* m is now error in rtt est */
+               srtt += m;              /* rtt = 7/8 rtt + 1/8 new */
+               if (m < 0) {
+                       m = -m;         /* m is now abs(error) */
+                       m -= (peer->mdev_us >> 2);   /* similar update on mdev */
+                       /* This is similar to one of Eifel findings.
+                        * Eifel blocks mdev updates when rtt decreases.
+                        * This solution is a bit different: we use finer gain
+                        * for mdev in this case (alpha*beta).
+                        * Like Eifel it also prevents growth of rto,
+                        * but also it limits too fast rto decreases,
+                        * happening in pure Eifel.
+                        */
+                       if (m > 0)
+                               m >>= 3;
+               } else {
+                       m -= (peer->mdev_us >> 2);   /* similar update on mdev */
+               }
+
+               peer->mdev_us += m;             /* mdev = 3/4 mdev + 1/4 new */
+               if (peer->mdev_us > peer->mdev_max_us) {
+                       peer->mdev_max_us = peer->mdev_us;
+                       if (peer->mdev_max_us > peer->rttvar_us)
+                               peer->rttvar_us = peer->mdev_max_us;
+               }
+       } else {
+               /* no previous measure. */
+               srtt = m << 3;          /* take the measured time to be rtt */
+               peer->mdev_us = m << 1; /* make sure rto = 3*rtt */
+               peer->rttvar_us = max(peer->mdev_us, rxrpc_rto_min_us(peer));
+               peer->mdev_max_us = peer->rttvar_us;
+       }
+
+       peer->srtt_us = max(1U, srtt);
+}
+
+/*
+ * Calculate rto without backoff.  This is the second half of Van Jacobson's
+ * routine referred to above.
+ */
+static void rxrpc_set_rto(struct rxrpc_peer *peer)
+{
+       u32 rto;
+
+       /* 1. If rtt variance happened to be less 50msec, it is hallucination.
+        *    It cannot be less due to utterly erratic ACK generation made
+        *    at least by solaris and freebsd. "Erratic ACKs" has _nothing_
+        *    to do with delayed acks, because at cwnd>2 true delack timeout
+        *    is invisible. Actually, Linux-2.4 also generates erratic
+        *    ACKs in some circumstances.
+        */
+       rto = __rxrpc_set_rto(peer);
+
+       /* 2. Fixups made earlier cannot be right.
+        *    If we do not estimate RTO correctly without them,
+        *    all the algo is pure shit and should be replaced
+        *    with correct one. It is exactly, which we pretend to do.
+        */
+
+       /* NOTE: clamping at RXRPC_RTO_MIN is not required, current algo
+        * guarantees that rto is higher.
+        */
+       peer->rto_j = rxrpc_bound_rto(rto);
+}
+
+static void rxrpc_ack_update_rtt(struct rxrpc_peer *peer, long rtt_us)
+{
+       if (rtt_us < 0)
+               return;
+
+       //rxrpc_update_rtt_min(peer, rtt_us);
+       rxrpc_rtt_estimator(peer, rtt_us);
+       rxrpc_set_rto(peer);
+
+       /* RFC6298: only reset backoff on valid RTT measurement. */
+       peer->backoff = 0;
+}
+
+/*
+ * Add RTT information to cache.  This is called in softirq mode and has
+ * exclusive access to the peer RTT data.
+ */
+void rxrpc_peer_add_rtt(struct rxrpc_call *call, enum rxrpc_rtt_rx_trace why,
+                       rxrpc_serial_t send_serial, rxrpc_serial_t resp_serial,
+                       ktime_t send_time, ktime_t resp_time)
+{
+       struct rxrpc_peer *peer = call->peer;
+       s64 rtt_us;
+
+       rtt_us = ktime_to_us(ktime_sub(resp_time, send_time));
+       if (rtt_us < 0)
+               return;
+
+       spin_lock(&peer->rtt_input_lock);
+       rxrpc_ack_update_rtt(peer, rtt_us);
+       if (peer->rtt_count < 3)
+               peer->rtt_count++;
+       spin_unlock(&peer->rtt_input_lock);
+
+       trace_rxrpc_rtt_rx(call, why, send_serial, resp_serial,
+                          peer->srtt_us >> 3, peer->rto_j);
+}
+
+/*
+ * Get the retransmission timeout to set in jiffies, backing it off each time
+ * we retransmit.
+ */
+unsigned long rxrpc_get_rto_backoff(struct rxrpc_peer *peer, bool retrans)
+{
+       u64 timo_j;
+       u8 backoff = READ_ONCE(peer->backoff);
+
+       timo_j = peer->rto_j;
+       timo_j <<= backoff;
+       if (retrans && timo_j * 2 <= RXRPC_RTO_MAX)
+               WRITE_ONCE(peer->backoff, backoff + 1);
+
+       if (timo_j < 1)
+               timo_j = 1;
+
+       return timo_j;
+}
+
+void rxrpc_peer_init_rtt(struct rxrpc_peer *peer)
+{
+       peer->rto_j     = RXRPC_TIMEOUT_INIT;
+       peer->mdev_us   = jiffies_to_usecs(RXRPC_TIMEOUT_INIT);
+       peer->backoff   = 0;
+       //minmax_reset(&peer->rtt_min, rxrpc_jiffies32, ~0U);
+}
index 098f1f9..52a24d4 100644 (file)
@@ -1148,7 +1148,7 @@ static int rxkad_verify_response(struct rxrpc_connection *conn,
        ret = rxkad_decrypt_ticket(conn, skb, ticket, ticket_len, &session_key,
                                   &expiry, _abort_code);
        if (ret < 0)
-               goto temporary_error_free_resp;
+               goto temporary_error_free_ticket;
 
        /* use the session key from inside the ticket to decrypt the
         * response */
@@ -1230,7 +1230,6 @@ protocol_error:
 
 temporary_error_free_ticket:
        kfree(ticket);
-temporary_error_free_resp:
        kfree(response);
 temporary_error:
        /* Ignore the response packet if we got a temporary error such as
index 0fcf157..5e9c43d 100644 (file)
@@ -66,15 +66,14 @@ static int rxrpc_wait_for_tx_window_waitall(struct rxrpc_sock *rx,
                                            struct rxrpc_call *call)
 {
        rxrpc_seq_t tx_start, tx_win;
-       signed long rtt2, timeout;
-       u64 rtt;
+       signed long rtt, timeout;
 
-       rtt = READ_ONCE(call->peer->rtt);
-       rtt2 = nsecs_to_jiffies64(rtt) * 2;
-       if (rtt2 < 2)
-               rtt2 = 2;
+       rtt = READ_ONCE(call->peer->srtt_us) >> 3;
+       rtt = usecs_to_jiffies(rtt) * 2;
+       if (rtt < 2)
+               rtt = 2;
 
-       timeout = rtt2;
+       timeout = rtt;
        tx_start = READ_ONCE(call->tx_hard_ack);
 
        for (;;) {
@@ -92,7 +91,7 @@ static int rxrpc_wait_for_tx_window_waitall(struct rxrpc_sock *rx,
                        return -EINTR;
 
                if (tx_win != tx_start) {
-                       timeout = rtt2;
+                       timeout = rtt;
                        tx_start = tx_win;
                }
 
@@ -271,16 +270,9 @@ static int rxrpc_queue_packet(struct rxrpc_sock *rx, struct rxrpc_call *call,
                _debug("need instant resend %d", ret);
                rxrpc_instant_resend(call, ix);
        } else {
-               unsigned long now = jiffies, resend_at;
+               unsigned long now = jiffies;
+               unsigned long resend_at = now + call->peer->rto_j;
 
-               if (call->peer->rtt_usage > 1)
-                       resend_at = nsecs_to_jiffies(call->peer->rtt * 3 / 2);
-               else
-                       resend_at = rxrpc_resend_timeout;
-               if (resend_at < 1)
-                       resend_at = 1;
-
-               resend_at += now;
                WRITE_ONCE(call->resend_at, resend_at);
                rxrpc_reduce_call_timer(call, resend_at, now,
                                        rxrpc_timer_set_for_send);
index 2bbb381..18dade4 100644 (file)
@@ -71,15 +71,6 @@ static struct ctl_table rxrpc_sysctl_table[] = {
                .extra1         = (void *)&one_jiffy,
                .extra2         = (void *)&max_jiffies,
        },
-       {
-               .procname       = "resend_timeout",
-               .data           = &rxrpc_resend_timeout,
-               .maxlen         = sizeof(unsigned long),
-               .mode           = 0644,
-               .proc_handler   = proc_doulongvec_ms_jiffies_minmax,
-               .extra1         = (void *)&one_jiffy,
-               .extra2         = (void *)&max_jiffies,
-       },
 
        /* Non-time values */
        {
index 2bc2946..9f36fe9 100644 (file)
@@ -1523,9 +1523,17 @@ static int sctp_cmd_interpreter(enum sctp_event_type event_type,
                        timeout = asoc->timeouts[cmd->obj.to];
                        BUG_ON(!timeout);
 
-                       timer->expires = jiffies + timeout;
-                       sctp_association_hold(asoc);
-                       add_timer(timer);
+                       /*
+                        * SCTP has a hard time with timer starts.  Because we process
+                        * timer starts as side effects, it can be hard to tell if we
+                        * have already started a timer or not, which leads to BUG
+                        * halts when we call add_timer. So here, instead of just starting
+                        * a timer, if the timer is already started, and just mod
+                        * the timer with the shorter of the two expiration times
+                        */
+                       if (!timer_pending(timer))
+                               sctp_association_hold(asoc);
+                       timer_reduce(timer, jiffies + timeout);
                        break;
 
                case SCTP_CMD_TIMER_RESTART:
index 26788f4..e86620f 100644 (file)
@@ -1856,12 +1856,13 @@ static enum sctp_disposition sctp_sf_do_dupcook_a(
        /* Update the content of current association. */
        sctp_add_cmd_sf(commands, SCTP_CMD_UPDATE_ASSOC, SCTP_ASOC(new_asoc));
        sctp_add_cmd_sf(commands, SCTP_CMD_EVENT_ULP, SCTP_ULPEVENT(ev));
-       if (sctp_state(asoc, SHUTDOWN_PENDING) &&
+       if ((sctp_state(asoc, SHUTDOWN_PENDING) ||
+            sctp_state(asoc, SHUTDOWN_SENT)) &&
            (sctp_sstate(asoc->base.sk, CLOSING) ||
             sock_flag(asoc->base.sk, SOCK_DEAD))) {
-               /* if were currently in SHUTDOWN_PENDING, but the socket
-                * has been closed by user, don't transition to ESTABLISHED.
-                * Instead trigger SHUTDOWN bundled with COOKIE_ACK.
+               /* If the socket has been closed by user, don't
+                * transition to ESTABLISHED. Instead trigger SHUTDOWN
+                * bundled with COOKIE_ACK.
                 */
                sctp_add_cmd_sf(commands, SCTP_CMD_REPLY, SCTP_CHUNK(repl));
                return sctp_sf_do_9_2_start_shutdown(net, ep, asoc,
index d6620ad..28a283f 100644 (file)
@@ -161,9 +161,11 @@ static int tipc_udp_xmit(struct net *net, struct sk_buff *skb,
                         struct udp_bearer *ub, struct udp_media_addr *src,
                         struct udp_media_addr *dst, struct dst_cache *cache)
 {
-       struct dst_entry *ndst = dst_cache_get(cache);
+       struct dst_entry *ndst;
        int ttl, err = 0;
 
+       local_bh_disable();
+       ndst = dst_cache_get(cache);
        if (dst->proto == htons(ETH_P_IP)) {
                struct rtable *rt = (struct rtable *)ndst;
 
@@ -210,9 +212,11 @@ static int tipc_udp_xmit(struct net *net, struct sk_buff *skb,
                                           src->port, dst->port, false);
 #endif
        }
+       local_bh_enable();
        return err;
 
 tx_error:
+       local_bh_enable();
        kfree_skb(skb);
        return err;
 }
index e23f94a..2d399b6 100644 (file)
@@ -780,7 +780,7 @@ static int tls_push_record(struct sock *sk, int flags,
 
 static int bpf_exec_tx_verdict(struct sk_msg *msg, struct sock *sk,
                               bool full_record, u8 record_type,
-                              size_t *copied, int flags)
+                              ssize_t *copied, int flags)
 {
        struct tls_context *tls_ctx = tls_get_ctx(sk);
        struct tls_sw_context_tx *ctx = tls_sw_ctx_tx(tls_ctx);
@@ -796,9 +796,10 @@ static int bpf_exec_tx_verdict(struct sk_msg *msg, struct sock *sk,
        psock = sk_psock_get(sk);
        if (!psock || !policy) {
                err = tls_push_record(sk, flags, record_type);
-               if (err && err != -EINPROGRESS) {
+               if (err && sk->sk_err == EBADMSG) {
                        *copied -= sk_msg_free(sk, msg);
                        tls_free_open_rec(sk);
+                       err = -sk->sk_err;
                }
                if (psock)
                        sk_psock_put(sk, psock);
@@ -824,9 +825,10 @@ more_data:
        switch (psock->eval) {
        case __SK_PASS:
                err = tls_push_record(sk, flags, record_type);
-               if (err && err != -EINPROGRESS) {
+               if (err && sk->sk_err == EBADMSG) {
                        *copied -= sk_msg_free(sk, msg);
                        tls_free_open_rec(sk);
+                       err = -sk->sk_err;
                        goto out_err;
                }
                break;
@@ -916,7 +918,8 @@ int tls_sw_sendmsg(struct sock *sk, struct msghdr *msg, size_t size)
        unsigned char record_type = TLS_RECORD_TYPE_DATA;
        bool is_kvec = iov_iter_is_kvec(&msg->msg_iter);
        bool eor = !(msg->msg_flags & MSG_MORE);
-       size_t try_to_copy, copied = 0;
+       size_t try_to_copy;
+       ssize_t copied = 0;
        struct sk_msg *msg_pl, *msg_en;
        struct tls_rec *rec;
        int required_size;
@@ -1118,7 +1121,7 @@ send_end:
 
        release_sock(sk);
        mutex_unlock(&tls_ctx->tx_lock);
-       return copied ? copied : ret;
+       return copied > 0 ? copied : ret;
 }
 
 static int tls_sw_do_sendpage(struct sock *sk, struct page *page,
@@ -1132,7 +1135,7 @@ static int tls_sw_do_sendpage(struct sock *sk, struct page *page,
        struct sk_msg *msg_pl;
        struct tls_rec *rec;
        int num_async = 0;
-       size_t copied = 0;
+       ssize_t copied = 0;
        bool full_record;
        int record_room;
        int ret = 0;
@@ -1234,7 +1237,7 @@ wait_for_memory:
        }
 sendpage_end:
        ret = sk_stream_error(sk, flags, ret);
-       return copied ? copied : ret;
+       return copied > 0 ? copied : ret;
 }
 
 int tls_sw_sendpage_locked(struct sock *sk, struct page *page,
index 280741f..f6a3ecf 100644 (file)
@@ -454,7 +454,7 @@ static ssize_t policy_update(u32 mask, const char __user *buf, size_t size,
         */
        error = aa_may_manage_policy(label, ns, mask);
        if (error)
-               return error;
+               goto end_section;
 
        data = aa_simple_write_to_buffer(buf, size, size, pos);
        error = PTR_ERR(data);
@@ -462,6 +462,7 @@ static ssize_t policy_update(u32 mask, const char __user *buf, size_t size,
                error = aa_replace_profiles(ns, label, mask, data);
                aa_put_loaddata(data);
        }
+end_section:
        end_current_label_crit_section(label);
 
        return error;
index 5a98661..5977325 100644 (file)
@@ -197,8 +197,9 @@ int aa_audit_rule_init(u32 field, u32 op, char *rulestr, void **vrule)
        rule->label = aa_label_parse(&root_ns->unconfined->label, rulestr,
                                     GFP_KERNEL, true, false);
        if (IS_ERR(rule->label)) {
+               int err = PTR_ERR(rule->label);
                aa_audit_rule_free(rule);
-               return PTR_ERR(rule->label);
+               return err;
        }
 
        *vrule = rule;
index 6ceb74e..a84ef03 100644 (file)
@@ -1328,6 +1328,7 @@ int aa_change_profile(const char *fqname, int flags)
                ctx->nnp = aa_get_label(label);
 
        if (!fqname || !*fqname) {
+               aa_put_label(label);
                AA_DEBUG("no profile name");
                return -EINVAL;
        }
@@ -1346,8 +1347,6 @@ int aa_change_profile(const char *fqname, int flags)
                        op = OP_CHANGE_PROFILE;
        }
 
-       label = aa_get_current_label();
-
        if (*fqname == '&') {
                stack = true;
                /* don't have label_parse() do stacking */
index 3568285..764b896 100644 (file)
@@ -73,7 +73,7 @@ static struct shash_desc *init_desc(char type, uint8_t hash_algo)
 {
        long rc;
        const char *algo;
-       struct crypto_shash **tfm;
+       struct crypto_shash **tfm, *tmp_tfm;
        struct shash_desc *desc;
 
        if (type == EVM_XATTR_HMAC) {
@@ -91,31 +91,31 @@ static struct shash_desc *init_desc(char type, uint8_t hash_algo)
                algo = hash_algo_name[hash_algo];
        }
 
-       if (*tfm == NULL) {
-               mutex_lock(&mutex);
-               if (*tfm)
-                       goto out;
-               *tfm = crypto_alloc_shash(algo, 0, CRYPTO_NOLOAD);
-               if (IS_ERR(*tfm)) {
-                       rc = PTR_ERR(*tfm);
-                       pr_err("Can not allocate %s (reason: %ld)\n", algo, rc);
-                       *tfm = NULL;
+       if (*tfm)
+               goto alloc;
+       mutex_lock(&mutex);
+       if (*tfm)
+               goto unlock;
+
+       tmp_tfm = crypto_alloc_shash(algo, 0, CRYPTO_NOLOAD);
+       if (IS_ERR(tmp_tfm)) {
+               pr_err("Can not allocate %s (reason: %ld)\n", algo,
+                      PTR_ERR(tmp_tfm));
+               mutex_unlock(&mutex);
+               return ERR_CAST(tmp_tfm);
+       }
+       if (type == EVM_XATTR_HMAC) {
+               rc = crypto_shash_setkey(tmp_tfm, evmkey, evmkey_len);
+               if (rc) {
+                       crypto_free_shash(tmp_tfm);
                        mutex_unlock(&mutex);
                        return ERR_PTR(rc);
                }
-               if (type == EVM_XATTR_HMAC) {
-                       rc = crypto_shash_setkey(*tfm, evmkey, evmkey_len);
-                       if (rc) {
-                               crypto_free_shash(*tfm);
-                               *tfm = NULL;
-                               mutex_unlock(&mutex);
-                               return ERR_PTR(rc);
-                       }
-               }
-out:
-               mutex_unlock(&mutex);
        }
-
+       *tfm = tmp_tfm;
+unlock:
+       mutex_unlock(&mutex);
+alloc:
        desc = kmalloc(sizeof(*desc) + crypto_shash_descsize(*tfm),
                        GFP_KERNEL);
        if (!desc)
@@ -207,7 +207,7 @@ static int evm_calc_hmac_or_hash(struct dentry *dentry,
        data->hdr.length = crypto_shash_digestsize(desc->tfm);
 
        error = -ENODATA;
-       list_for_each_entry_rcu(xattr, &evm_config_xattrnames, list) {
+       list_for_each_entry_lockless(xattr, &evm_config_xattrnames, list) {
                bool is_ima = false;
 
                if (strcmp(xattr->name, XATTR_NAME_IMA) == 0)
index d361d7f..0d36259 100644 (file)
@@ -97,7 +97,7 @@ static int evm_find_protected_xattrs(struct dentry *dentry)
        if (!(inode->i_opflags & IOP_XATTR))
                return -EOPNOTSUPP;
 
-       list_for_each_entry_rcu(xattr, &evm_config_xattrnames, list) {
+       list_for_each_entry_lockless(xattr, &evm_config_xattrnames, list) {
                error = __vfs_getxattr(dentry, inode, xattr->name, NULL, 0);
                if (error < 0) {
                        if (error == -ENODATA)
@@ -228,7 +228,7 @@ static int evm_protected_xattr(const char *req_xattr_name)
        struct xattr_list *xattr;
 
        namelen = strlen(req_xattr_name);
-       list_for_each_entry_rcu(xattr, &evm_config_xattrnames, list) {
+       list_for_each_entry_lockless(xattr, &evm_config_xattrnames, list) {
                if ((strlen(xattr->name) == namelen)
                    && (strncmp(req_xattr_name, xattr->name, namelen) == 0)) {
                        found = 1;
index 39ad103..cfc3075 100644 (file)
@@ -232,7 +232,14 @@ static ssize_t evm_write_xattrs(struct file *file, const char __user *buf,
                goto out;
        }
 
-       /* Guard against races in evm_read_xattrs */
+       /*
+        * xattr_list_mutex guards against races in evm_read_xattrs().
+        * Entries are only added to the evm_config_xattrnames list
+        * and never deleted. Therefore, the list is traversed
+        * using list_for_each_entry_lockless() without holding
+        * the mutex in evm_calc_hmac_or_hash(), evm_find_protected_xattrs()
+        * and evm_protected_xattr().
+        */
        mutex_lock(&xattr_list_mutex);
        list_for_each_entry(tmp, &evm_config_xattrnames, list) {
                if (strcmp(xattr->name, tmp->name) == 0) {
index 423c84f..88b5e28 100644 (file)
@@ -411,7 +411,7 @@ int ima_calc_file_hash(struct file *file, struct ima_digest_data *hash)
        loff_t i_size;
        int rc;
        struct file *f = file;
-       bool new_file_instance = false, modified_flags = false;
+       bool new_file_instance = false, modified_mode = false;
 
        /*
         * For consistency, fail file's opened with the O_DIRECT flag on
@@ -431,13 +431,13 @@ int ima_calc_file_hash(struct file *file, struct ima_digest_data *hash)
                f = dentry_open(&file->f_path, flags, file->f_cred);
                if (IS_ERR(f)) {
                        /*
-                        * Cannot open the file again, lets modify f_flags
+                        * Cannot open the file again, lets modify f_mode
                         * of original and continue
                         */
                        pr_info_ratelimited("Unable to reopen file for reading.\n");
                        f = file;
-                       f->f_flags |= FMODE_READ;
-                       modified_flags = true;
+                       f->f_mode |= FMODE_READ;
+                       modified_mode = true;
                } else {
                        new_file_instance = true;
                }
@@ -455,8 +455,8 @@ int ima_calc_file_hash(struct file *file, struct ima_digest_data *hash)
 out:
        if (new_file_instance)
                fput(f);
-       else if (modified_flags)
-               f->f_flags &= ~FMODE_READ;
+       else if (modified_mode)
+               f->f_mode &= ~FMODE_READ;
        return rc;
 }
 
index a71e822..3efc830 100644 (file)
@@ -338,8 +338,7 @@ static ssize_t ima_write_policy(struct file *file, const char __user *buf,
                integrity_audit_msg(AUDIT_INTEGRITY_STATUS, NULL, NULL,
                                    "policy_update", "signed policy required",
                                    1, 0);
-               if (ima_appraise & IMA_APPRAISE_ENFORCE)
-                       result = -EACCES;
+               result = -EACCES;
        } else {
                result = ima_parse_add_rule(data);
        }
index 7fed24b..51de970 100644 (file)
@@ -1965,8 +1965,20 @@ EXPORT_SYMBOL(security_ismaclabel);
 
 int security_secid_to_secctx(u32 secid, char **secdata, u32 *seclen)
 {
-       return call_int_hook(secid_to_secctx, -EOPNOTSUPP, secid, secdata,
-                               seclen);
+       struct security_hook_list *hp;
+       int rc;
+
+       /*
+        * Currently, only one LSM can implement secid_to_secctx (i.e this
+        * LSM hook is not "stackable").
+        */
+       hlist_for_each_entry(hp, &security_hook_heads.secid_to_secctx, list) {
+               rc = hp->hook.secid_to_secctx(secid, secdata, seclen);
+               if (rc != LSM_RET_DEFAULT(secid_to_secctx))
+                       return rc;
+       }
+
+       return LSM_RET_DEFAULT(secid_to_secctx);
 }
 EXPORT_SYMBOL(security_secid_to_secctx);
 
index 872a852..d531e1b 100644 (file)
@@ -433,6 +433,7 @@ static int snd_pcm_update_hw_ptr0(struct snd_pcm_substream *substream,
 
  no_delta_check:
        if (runtime->status->hw_ptr == new_hw_ptr) {
+               runtime->hw_ptr_jiffies = curr_jiffies;
                update_audio_tstamp(substream, &curr_tstamp, &audio_tstamp);
                return 0;
        }
index dc23021..041d2a3 100644 (file)
@@ -2457,6 +2457,7 @@ static const struct snd_pci_quirk alc882_fixup_tbl[] = {
        SND_PCI_QUIRK(0x1458, 0xa002, "Gigabyte EP45-DS3/Z87X-UD3H", ALC889_FIXUP_FRONT_HP_NO_PRESENCE),
        SND_PCI_QUIRK(0x1458, 0xa0b8, "Gigabyte AZ370-Gaming", ALC1220_FIXUP_GB_DUAL_CODECS),
        SND_PCI_QUIRK(0x1458, 0xa0cd, "Gigabyte X570 Aorus Master", ALC1220_FIXUP_CLEVO_P950),
+       SND_PCI_QUIRK(0x1458, 0xa0ce, "Gigabyte X570 Aorus Xtreme", ALC1220_FIXUP_CLEVO_P950),
        SND_PCI_QUIRK(0x1462, 0x1228, "MSI-GP63", ALC1220_FIXUP_CLEVO_P950),
        SND_PCI_QUIRK(0x1462, 0x1275, "MSI-GL63", ALC1220_FIXUP_CLEVO_P950),
        SND_PCI_QUIRK(0x1462, 0x1276, "MSI-GL73", ALC1220_FIXUP_CLEVO_P950),
@@ -2472,6 +2473,9 @@ static const struct snd_pci_quirk alc882_fixup_tbl[] = {
        SND_PCI_QUIRK(0x1558, 0x97e1, "Clevo P970[ER][CDFN]", ALC1220_FIXUP_CLEVO_P950),
        SND_PCI_QUIRK(0x1558, 0x65d1, "Clevo PB51[ER][CDF]", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
        SND_PCI_QUIRK(0x1558, 0x67d1, "Clevo PB71[ER][CDF]", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
+       SND_PCI_QUIRK(0x1558, 0x50d3, "Clevo PC50[ER][CDF]", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
+       SND_PCI_QUIRK(0x1558, 0x70d1, "Clevo PC70[ER][CDF]", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
+       SND_PCI_QUIRK(0x1558, 0x7714, "Clevo X170", ALC1220_FIXUP_CLEVO_PB51ED_PINS),
        SND_PCI_QUIRK_VENDOR(0x1558, "Clevo laptop", ALC882_FIXUP_EAPD),
        SND_PCI_QUIRK(0x161f, 0x2054, "Medion laptop", ALC883_FIXUP_EAPD),
        SND_PCI_QUIRK(0x17aa, 0x3a0d, "Lenovo Y530", ALC882_FIXUP_LENOVO_Y530),
index 884d0cd..73e1e54 100644 (file)
@@ -2332,7 +2332,8 @@ static int snd_ice1712_chip_init(struct snd_ice1712 *ice)
        pci_write_config_byte(ice->pci, 0x61, ice->eeprom.data[ICE_EEP1_ACLINK]);
        pci_write_config_byte(ice->pci, 0x62, ice->eeprom.data[ICE_EEP1_I2SID]);
        pci_write_config_byte(ice->pci, 0x63, ice->eeprom.data[ICE_EEP1_SPDIF]);
-       if (ice->eeprom.subvendor != ICE1712_SUBDEVICE_STDSP24) {
+       if (ice->eeprom.subvendor != ICE1712_SUBDEVICE_STDSP24 &&
+           ice->eeprom.subvendor != ICE1712_SUBDEVICE_STAUDIO_ADCIII) {
                ice->gpio.write_mask = ice->eeprom.gpiomask;
                ice->gpio.direction = ice->eeprom.gpiodir;
                snd_ice1712_write(ice, ICE1712_IREG_GPIO_WRITE_MASK,
index 6b9dce4..43d0b55 100644 (file)
@@ -19,7 +19,7 @@ void test_mmap(void)
        const size_t map_sz = roundup_page(sizeof(struct map_data));
        const int zero = 0, one = 1, two = 2, far = 1500;
        const long page_size = sysconf(_SC_PAGE_SIZE);
-       int err, duration = 0, i, data_map_fd, data_map_id, tmp_fd;
+       int err, duration = 0, i, data_map_fd, data_map_id, tmp_fd, rdmap_fd;
        struct bpf_map *data_map, *bss_map;
        void *bss_mmaped = NULL, *map_mmaped = NULL, *tmp1, *tmp2;
        struct test_mmap__bss *bss_data;
@@ -37,6 +37,17 @@ void test_mmap(void)
        data_map = skel->maps.data_map;
        data_map_fd = bpf_map__fd(data_map);
 
+       rdmap_fd = bpf_map__fd(skel->maps.rdonly_map);
+       tmp1 = mmap(NULL, 4096, PROT_READ | PROT_WRITE, MAP_SHARED, rdmap_fd, 0);
+       if (CHECK(tmp1 != MAP_FAILED, "rdonly_write_mmap", "unexpected success\n")) {
+               munmap(tmp1, 4096);
+               goto cleanup;
+       }
+       /* now double-check if it's mmap()'able at all */
+       tmp1 = mmap(NULL, 4096, PROT_READ, MAP_SHARED, rdmap_fd, 0);
+       if (CHECK(tmp1 == MAP_FAILED, "rdonly_read_mmap", "failed: %d\n", errno))
+               goto cleanup;
+
        /* get map's ID */
        memset(&map_info, 0, map_info_sz);
        err = bpf_obj_get_info_by_fd(data_map_fd, &map_info, &map_info_sz);
index 6239596..4eb42cf 100644 (file)
@@ -7,6 +7,14 @@
 
 char _license[] SEC("license") = "GPL";
 
+struct {
+       __uint(type, BPF_MAP_TYPE_ARRAY);
+       __uint(max_entries, 4096);
+       __uint(map_flags, BPF_F_MMAPABLE | BPF_F_RDONLY_PROG);
+       __type(key, __u32);
+       __type(value, char);
+} rdonly_map SEC(".maps");
+
 struct {
        __uint(type, BPF_MAP_TYPE_ARRAY);
        __uint(max_entries, 512 * 4); /* at least 4 pages of data */
index 24dd8ed..b025dae 100755 (executable)
@@ -300,7 +300,7 @@ test_uc_aware()
        local i
 
        for ((i = 0; i < attempts; ++i)); do
-               if $ARPING -c 1 -I $h1 -b 192.0.2.66 -q -w 0.1; then
+               if $ARPING -c 1 -I $h1 -b 192.0.2.66 -q -w 1; then
                        ((passes++))
                fi
 
index dbd1e01..da49ad2 100755 (executable)
@@ -264,6 +264,8 @@ trap_policer_test()
        local packets_t0
        local packets_t1
 
+       RET=0
+
        if [ $(devlink_trap_policers_num_get) -eq 0 ]; then
                check_err 1 "Failed to dump policers"
        fi
@@ -328,6 +330,8 @@ trap_group_check_policer()
 
 trap_policer_bind_test()
 {
+       RET=0
+
        devlink trap group set $DEVLINK_DEV group l2_drops policer 1
        check_err $? "Failed to bind a valid policer"
        if [ $(devlink_trap_group_policer_get "l2_drops") -ne 1 ]; then
index 0edb6d9..ca17fe0 100644 (file)
@@ -6,6 +6,7 @@ map_populate
 thuge-gen
 compaction_test
 mlock2-tests
+mremap_dontunmap
 on-fault-limit
 transhuge-stress
 userfaultfd
index 110bc4e..6a2caba 100644 (file)
@@ -74,8 +74,6 @@ int main(int argc, char **argv)
        int write = 0;
        int reserve = 1;
 
-       unsigned long i;
-
        if (signal(SIGINT, sig_handler) == SIG_ERR)
                err(1, "\ncan't catch SIGINT\n");
 
index 90598a4..4bdd6c1 100644 (file)
@@ -44,7 +44,7 @@ endef
 $(eval $(call tar_download,MUSL,musl,1.2.0,.tar.gz,https://musl.libc.org/releases/,c6de7b191139142d3f9a7b5b702c9cae1b5ee6e7f57e582da9328629408fd4e8))
 $(eval $(call tar_download,IPERF,iperf,3.7,.tar.gz,https://downloads.es.net/pub/iperf/,d846040224317caf2f75c843d309a950a7db23f9b44b94688ccbe557d6d1710c))
 $(eval $(call tar_download,BASH,bash,5.0,.tar.gz,https://ftp.gnu.org/gnu/bash/,b4a80f2ac66170b2913efbfb9f2594f1f76c7b1afd11f799e22035d63077fb4d))
-$(eval $(call tar_download,IPROUTE2,iproute2,5.4.0,.tar.xz,https://www.kernel.org/pub/linux/utils/net/iproute2/,fe97aa60a0d4c5ac830be18937e18dc3400ca713a33a89ad896ff1e3d46086ae))
+$(eval $(call tar_download,IPROUTE2,iproute2,5.6.0,.tar.xz,https://www.kernel.org/pub/linux/utils/net/iproute2/,1b5b0e25ce6e23da7526ea1da044e814ad85ba761b10dd29c2b027c056b04692))
 $(eval $(call tar_download,IPTABLES,iptables,1.8.4,.tar.bz2,https://www.netfilter.org/projects/iptables/files/,993a3a5490a544c2cbf2ef15cf7e7ed21af1845baf228318d5c36ef8827e157c))
 $(eval $(call tar_download,NMAP,nmap,7.80,.tar.bz2,https://nmap.org/dist/,fcfa5a0e42099e12e4bf7a68ebe6fde05553383a682e816a7ec9256ab4773faa))
 $(eval $(call tar_download,IPUTILS,iputils,s20190709,.tar.gz,https://github.com/iputils/iputils/archive/s20190709.tar.gz/#,a15720dd741d7538dd2645f9f516d193636ae4300ff7dbc8bfca757bf166490a))