summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorAnthony G. Basile <blueness@gentoo.org>2017-04-01 19:40:12 -0400
committerAnthony G. Basile <blueness@gentoo.org>2017-04-01 19:40:12 -0400
commitb897bcf385333e0aaa97084370db28c16dc589f4 (patch)
tree603a21bf6485d0ebd9d7216bf3e921f9c1f3a363
parentgrsecurity-3.1-4.9.18-201703261106 (diff)
downloadhardened-patchset-20170331.tar.gz
hardened-patchset-20170331.tar.bz2
hardened-patchset-20170331.zip
grsecurity-3.1-4.9.20-20170331082320170331
-rw-r--r--4.9.18/1016_linux-4.9.17.patch6091
-rw-r--r--4.9.18/1017_linux-4.9.18.patch876
-rw-r--r--4.9.20/0000_README (renamed from 4.9.18/0000_README)10
-rw-r--r--4.9.20/4420_grsecurity-3.1-4.9.20-201703310823.patch (renamed from 4.9.18/4420_grsecurity-3.1-4.9.18-201703261106.patch)208
-rw-r--r--4.9.20/4425_grsec_remove_EI_PAX.patch (renamed from 4.9.18/4425_grsec_remove_EI_PAX.patch)0
-rw-r--r--4.9.20/4426_default_XATTR_PAX_FLAGS.patch (renamed from 4.9.18/4426_default_XATTR_PAX_FLAGS.patch)0
-rw-r--r--4.9.20/4427_force_XATTR_PAX_tmpfs.patch (renamed from 4.9.18/4427_force_XATTR_PAX_tmpfs.patch)0
-rw-r--r--4.9.20/4430_grsec-remove-localversion-grsec.patch (renamed from 4.9.18/4430_grsec-remove-localversion-grsec.patch)0
-rw-r--r--4.9.20/4435_grsec-mute-warnings.patch (renamed from 4.9.18/4435_grsec-mute-warnings.patch)0
-rw-r--r--4.9.20/4440_grsec-remove-protected-paths.patch (renamed from 4.9.18/4440_grsec-remove-protected-paths.patch)0
-rw-r--r--4.9.20/4450_grsec-kconfig-default-gids.patch (renamed from 4.9.18/4450_grsec-kconfig-default-gids.patch)0
-rw-r--r--4.9.20/4465_selinux-avc_audit-log-curr_ip.patch (renamed from 4.9.18/4465_selinux-avc_audit-log-curr_ip.patch)0
-rw-r--r--4.9.20/4470_disable-compat_vdso.patch (renamed from 4.9.18/4470_disable-compat_vdso.patch)0
-rw-r--r--4.9.20/4475_emutramp_default_on.patch (renamed from 4.9.18/4475_emutramp_default_on.patch)0
14 files changed, 143 insertions, 7042 deletions
diff --git a/4.9.18/1016_linux-4.9.17.patch b/4.9.18/1016_linux-4.9.17.patch
deleted file mode 100644
index 1a83496..0000000
--- a/4.9.18/1016_linux-4.9.17.patch
+++ /dev/null
@@ -1,6091 +0,0 @@
-diff --git a/Documentation/arm64/silicon-errata.txt b/Documentation/arm64/silicon-errata.txt
-index 405da11..d11af52 100644
---- a/Documentation/arm64/silicon-errata.txt
-+++ b/Documentation/arm64/silicon-errata.txt
-@@ -42,24 +42,26 @@ file acts as a registry of software workarounds in the Linux Kernel and
- will be updated when new workarounds are committed and backported to
- stable kernels.
-
--| Implementor | Component | Erratum ID | Kconfig |
--+----------------+-----------------+-----------------+-------------------------+
--| ARM | Cortex-A53 | #826319 | ARM64_ERRATUM_826319 |
--| ARM | Cortex-A53 | #827319 | ARM64_ERRATUM_827319 |
--| ARM | Cortex-A53 | #824069 | ARM64_ERRATUM_824069 |
--| ARM | Cortex-A53 | #819472 | ARM64_ERRATUM_819472 |
--| ARM | Cortex-A53 | #845719 | ARM64_ERRATUM_845719 |
--| ARM | Cortex-A53 | #843419 | ARM64_ERRATUM_843419 |
--| ARM | Cortex-A57 | #832075 | ARM64_ERRATUM_832075 |
--| ARM | Cortex-A57 | #852523 | N/A |
--| ARM | Cortex-A57 | #834220 | ARM64_ERRATUM_834220 |
--| ARM | Cortex-A72 | #853709 | N/A |
--| ARM | MMU-500 | #841119,#826419 | N/A |
--| | | | |
--| Cavium | ThunderX ITS | #22375, #24313 | CAVIUM_ERRATUM_22375 |
--| Cavium | ThunderX ITS | #23144 | CAVIUM_ERRATUM_23144 |
--| Cavium | ThunderX GICv3 | #23154 | CAVIUM_ERRATUM_23154 |
--| Cavium | ThunderX Core | #27456 | CAVIUM_ERRATUM_27456 |
--| Cavium | ThunderX SMMUv2 | #27704 | N/A |
--| | | | |
--| Freescale/NXP | LS2080A/LS1043A | A-008585 | FSL_ERRATUM_A008585 |
-+| Implementor | Component | Erratum ID | Kconfig |
-++----------------+-----------------+-----------------+-----------------------------+
-+| ARM | Cortex-A53 | #826319 | ARM64_ERRATUM_826319 |
-+| ARM | Cortex-A53 | #827319 | ARM64_ERRATUM_827319 |
-+| ARM | Cortex-A53 | #824069 | ARM64_ERRATUM_824069 |
-+| ARM | Cortex-A53 | #819472 | ARM64_ERRATUM_819472 |
-+| ARM | Cortex-A53 | #845719 | ARM64_ERRATUM_845719 |
-+| ARM | Cortex-A53 | #843419 | ARM64_ERRATUM_843419 |
-+| ARM | Cortex-A57 | #832075 | ARM64_ERRATUM_832075 |
-+| ARM | Cortex-A57 | #852523 | N/A |
-+| ARM | Cortex-A57 | #834220 | ARM64_ERRATUM_834220 |
-+| ARM | Cortex-A72 | #853709 | N/A |
-+| ARM | MMU-500 | #841119,#826419 | N/A |
-+| | | | |
-+| Cavium | ThunderX ITS | #22375, #24313 | CAVIUM_ERRATUM_22375 |
-+| Cavium | ThunderX ITS | #23144 | CAVIUM_ERRATUM_23144 |
-+| Cavium | ThunderX GICv3 | #23154 | CAVIUM_ERRATUM_23154 |
-+| Cavium | ThunderX Core | #27456 | CAVIUM_ERRATUM_27456 |
-+| Cavium | ThunderX SMMUv2 | #27704 | N/A |
-+| | | | |
-+| Freescale/NXP | LS2080A/LS1043A | A-008585 | FSL_ERRATUM_A008585 |
-+| | | | |
-+| Qualcomm Tech. | QDF2400 ITS | E0065 | QCOM_QDF2400_ERRATUM_0065 |
-diff --git a/Makefile b/Makefile
-index 4e0f962..004f90a 100644
---- a/Makefile
-+++ b/Makefile
-@@ -1,6 +1,6 @@
- VERSION = 4
- PATCHLEVEL = 9
--SUBLEVEL = 16
-+SUBLEVEL = 17
- EXTRAVERSION =
- NAME = Roaring Lionus
-
-diff --git a/arch/arm64/Kconfig b/arch/arm64/Kconfig
-index 969ef88..cf57a77 100644
---- a/arch/arm64/Kconfig
-+++ b/arch/arm64/Kconfig
-@@ -474,6 +474,16 @@ config CAVIUM_ERRATUM_27456
-
- If unsure, say Y.
-
-+config QCOM_QDF2400_ERRATUM_0065
-+ bool "QDF2400 E0065: Incorrect GITS_TYPER.ITT_Entry_size"
-+ default y
-+ help
-+ On Qualcomm Datacenter Technologies QDF2400 SoC, ITS hardware reports
-+ ITE size incorrectly. The GITS_TYPER.ITT_Entry_size field should have
-+ been indicated as 16Bytes (0xf), not 8Bytes (0x7).
-+
-+ If unsure, say Y.
-+
- endmenu
-
-
-diff --git a/arch/arm64/kvm/hyp/tlb.c b/arch/arm64/kvm/hyp/tlb.c
-index 88e2f2b..55889d0 100644
---- a/arch/arm64/kvm/hyp/tlb.c
-+++ b/arch/arm64/kvm/hyp/tlb.c
-@@ -17,14 +17,62 @@
-
- #include <asm/kvm_hyp.h>
-
-+static void __hyp_text __tlb_switch_to_guest_vhe(struct kvm *kvm)
-+{
-+ u64 val;
-+
-+ /*
-+ * With VHE enabled, we have HCR_EL2.{E2H,TGE} = {1,1}, and
-+ * most TLB operations target EL2/EL0. In order to affect the
-+ * guest TLBs (EL1/EL0), we need to change one of these two
-+ * bits. Changing E2H is impossible (goodbye TTBR1_EL2), so
-+ * let's flip TGE before executing the TLB operation.
-+ */
-+ write_sysreg(kvm->arch.vttbr, vttbr_el2);
-+ val = read_sysreg(hcr_el2);
-+ val &= ~HCR_TGE;
-+ write_sysreg(val, hcr_el2);
-+ isb();
-+}
-+
-+static void __hyp_text __tlb_switch_to_guest_nvhe(struct kvm *kvm)
-+{
-+ write_sysreg(kvm->arch.vttbr, vttbr_el2);
-+ isb();
-+}
-+
-+static hyp_alternate_select(__tlb_switch_to_guest,
-+ __tlb_switch_to_guest_nvhe,
-+ __tlb_switch_to_guest_vhe,
-+ ARM64_HAS_VIRT_HOST_EXTN);
-+
-+static void __hyp_text __tlb_switch_to_host_vhe(struct kvm *kvm)
-+{
-+ /*
-+ * We're done with the TLB operation, let's restore the host's
-+ * view of HCR_EL2.
-+ */
-+ write_sysreg(0, vttbr_el2);
-+ write_sysreg(HCR_HOST_VHE_FLAGS, hcr_el2);
-+}
-+
-+static void __hyp_text __tlb_switch_to_host_nvhe(struct kvm *kvm)
-+{
-+ write_sysreg(0, vttbr_el2);
-+}
-+
-+static hyp_alternate_select(__tlb_switch_to_host,
-+ __tlb_switch_to_host_nvhe,
-+ __tlb_switch_to_host_vhe,
-+ ARM64_HAS_VIRT_HOST_EXTN);
-+
- void __hyp_text __kvm_tlb_flush_vmid_ipa(struct kvm *kvm, phys_addr_t ipa)
- {
- dsb(ishst);
-
- /* Switch to requested VMID */
- kvm = kern_hyp_va(kvm);
-- write_sysreg(kvm->arch.vttbr, vttbr_el2);
-- isb();
-+ __tlb_switch_to_guest()(kvm);
-
- /*
- * We could do so much better if we had the VA as well.
-@@ -45,7 +93,7 @@ void __hyp_text __kvm_tlb_flush_vmid_ipa(struct kvm *kvm, phys_addr_t ipa)
- dsb(ish);
- isb();
-
-- write_sysreg(0, vttbr_el2);
-+ __tlb_switch_to_host()(kvm);
- }
-
- void __hyp_text __kvm_tlb_flush_vmid(struct kvm *kvm)
-@@ -54,14 +102,13 @@ void __hyp_text __kvm_tlb_flush_vmid(struct kvm *kvm)
-
- /* Switch to requested VMID */
- kvm = kern_hyp_va(kvm);
-- write_sysreg(kvm->arch.vttbr, vttbr_el2);
-- isb();
-+ __tlb_switch_to_guest()(kvm);
-
- asm volatile("tlbi vmalls12e1is" : : );
- dsb(ish);
- isb();
-
-- write_sysreg(0, vttbr_el2);
-+ __tlb_switch_to_host()(kvm);
- }
-
- void __hyp_text __kvm_tlb_flush_local_vmid(struct kvm_vcpu *vcpu)
-@@ -69,14 +116,13 @@ void __hyp_text __kvm_tlb_flush_local_vmid(struct kvm_vcpu *vcpu)
- struct kvm *kvm = kern_hyp_va(kern_hyp_va(vcpu)->kvm);
-
- /* Switch to requested VMID */
-- write_sysreg(kvm->arch.vttbr, vttbr_el2);
-- isb();
-+ __tlb_switch_to_guest()(kvm);
-
- asm volatile("tlbi vmalle1" : : );
- dsb(nsh);
- isb();
-
-- write_sysreg(0, vttbr_el2);
-+ __tlb_switch_to_host()(kvm);
- }
-
- void __hyp_text __kvm_flush_vm_context(void)
-diff --git a/arch/powerpc/crypto/crc32c-vpmsum_glue.c b/arch/powerpc/crypto/crc32c-vpmsum_glue.c
-index 9fa046d..4119945 100644
---- a/arch/powerpc/crypto/crc32c-vpmsum_glue.c
-+++ b/arch/powerpc/crypto/crc32c-vpmsum_glue.c
-@@ -52,7 +52,7 @@ static int crc32c_vpmsum_cra_init(struct crypto_tfm *tfm)
- {
- u32 *key = crypto_tfm_ctx(tfm);
-
-- *key = 0;
-+ *key = ~0;
-
- return 0;
- }
-diff --git a/arch/powerpc/include/asm/mmu_context.h b/arch/powerpc/include/asm/mmu_context.h
-index 5c45114..b9e3f0a 100644
---- a/arch/powerpc/include/asm/mmu_context.h
-+++ b/arch/powerpc/include/asm/mmu_context.h
-@@ -19,16 +19,18 @@ extern void destroy_context(struct mm_struct *mm);
- struct mm_iommu_table_group_mem_t;
-
- extern int isolate_lru_page(struct page *page); /* from internal.h */
--extern bool mm_iommu_preregistered(void);
--extern long mm_iommu_get(unsigned long ua, unsigned long entries,
-+extern bool mm_iommu_preregistered(struct mm_struct *mm);
-+extern long mm_iommu_get(struct mm_struct *mm,
-+ unsigned long ua, unsigned long entries,
- struct mm_iommu_table_group_mem_t **pmem);
--extern long mm_iommu_put(struct mm_iommu_table_group_mem_t *mem);
--extern void mm_iommu_init(mm_context_t *ctx);
--extern void mm_iommu_cleanup(mm_context_t *ctx);
--extern struct mm_iommu_table_group_mem_t *mm_iommu_lookup(unsigned long ua,
-- unsigned long size);
--extern struct mm_iommu_table_group_mem_t *mm_iommu_find(unsigned long ua,
-- unsigned long entries);
-+extern long mm_iommu_put(struct mm_struct *mm,
-+ struct mm_iommu_table_group_mem_t *mem);
-+extern void mm_iommu_init(struct mm_struct *mm);
-+extern void mm_iommu_cleanup(struct mm_struct *mm);
-+extern struct mm_iommu_table_group_mem_t *mm_iommu_lookup(struct mm_struct *mm,
-+ unsigned long ua, unsigned long size);
-+extern struct mm_iommu_table_group_mem_t *mm_iommu_find(struct mm_struct *mm,
-+ unsigned long ua, unsigned long entries);
- extern long mm_iommu_ua_to_hpa(struct mm_iommu_table_group_mem_t *mem,
- unsigned long ua, unsigned long *hpa);
- extern long mm_iommu_mapped_inc(struct mm_iommu_table_group_mem_t *mem);
-diff --git a/arch/powerpc/kernel/setup-common.c b/arch/powerpc/kernel/setup-common.c
-index 270ee30..f516ac5 100644
---- a/arch/powerpc/kernel/setup-common.c
-+++ b/arch/powerpc/kernel/setup-common.c
-@@ -915,7 +915,7 @@ void __init setup_arch(char **cmdline_p)
- init_mm.context.pte_frag = NULL;
- #endif
- #ifdef CONFIG_SPAPR_TCE_IOMMU
-- mm_iommu_init(&init_mm.context);
-+ mm_iommu_init(&init_mm);
- #endif
- irqstack_early_init();
- exc_lvl_early_init();
-diff --git a/arch/powerpc/mm/mmu_context_book3s64.c b/arch/powerpc/mm/mmu_context_book3s64.c
-index b114f8b..73bf6e1 100644
---- a/arch/powerpc/mm/mmu_context_book3s64.c
-+++ b/arch/powerpc/mm/mmu_context_book3s64.c
-@@ -115,7 +115,7 @@ int init_new_context(struct task_struct *tsk, struct mm_struct *mm)
- mm->context.pte_frag = NULL;
- #endif
- #ifdef CONFIG_SPAPR_TCE_IOMMU
-- mm_iommu_init(&mm->context);
-+ mm_iommu_init(mm);
- #endif
- return 0;
- }
-@@ -156,13 +156,11 @@ static inline void destroy_pagetable_page(struct mm_struct *mm)
- }
- #endif
-
--
- void destroy_context(struct mm_struct *mm)
- {
- #ifdef CONFIG_SPAPR_TCE_IOMMU
-- mm_iommu_cleanup(&mm->context);
-+ WARN_ON_ONCE(!list_empty(&mm->context.iommu_group_mem_list));
- #endif
--
- #ifdef CONFIG_PPC_ICSWX
- drop_cop(mm->context.acop, mm);
- kfree(mm->context.cop_lockp);
-diff --git a/arch/powerpc/mm/mmu_context_iommu.c b/arch/powerpc/mm/mmu_context_iommu.c
-index e0f1c33..7de7124 100644
---- a/arch/powerpc/mm/mmu_context_iommu.c
-+++ b/arch/powerpc/mm/mmu_context_iommu.c
-@@ -56,7 +56,7 @@ static long mm_iommu_adjust_locked_vm(struct mm_struct *mm,
- }
-
- pr_debug("[%d] RLIMIT_MEMLOCK HASH64 %c%ld %ld/%ld\n",
-- current->pid,
-+ current ? current->pid : 0,
- incr ? '+' : '-',
- npages << PAGE_SHIFT,
- mm->locked_vm << PAGE_SHIFT,
-@@ -66,12 +66,9 @@ static long mm_iommu_adjust_locked_vm(struct mm_struct *mm,
- return ret;
- }
-
--bool mm_iommu_preregistered(void)
-+bool mm_iommu_preregistered(struct mm_struct *mm)
- {
-- if (!current || !current->mm)
-- return false;
--
-- return !list_empty(&current->mm->context.iommu_group_mem_list);
-+ return !list_empty(&mm->context.iommu_group_mem_list);
- }
- EXPORT_SYMBOL_GPL(mm_iommu_preregistered);
-
-@@ -124,19 +121,16 @@ static int mm_iommu_move_page_from_cma(struct page *page)
- return 0;
- }
-
--long mm_iommu_get(unsigned long ua, unsigned long entries,
-+long mm_iommu_get(struct mm_struct *mm, unsigned long ua, unsigned long entries,
- struct mm_iommu_table_group_mem_t **pmem)
- {
- struct mm_iommu_table_group_mem_t *mem;
- long i, j, ret = 0, locked_entries = 0;
- struct page *page = NULL;
-
-- if (!current || !current->mm)
-- return -ESRCH; /* process exited */
--
- mutex_lock(&mem_list_mutex);
-
-- list_for_each_entry_rcu(mem, &current->mm->context.iommu_group_mem_list,
-+ list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list,
- next) {
- if ((mem->ua == ua) && (mem->entries == entries)) {
- ++mem->used;
-@@ -154,7 +148,7 @@ long mm_iommu_get(unsigned long ua, unsigned long entries,
-
- }
-
-- ret = mm_iommu_adjust_locked_vm(current->mm, entries, true);
-+ ret = mm_iommu_adjust_locked_vm(mm, entries, true);
- if (ret)
- goto unlock_exit;
-
-@@ -190,7 +184,7 @@ long mm_iommu_get(unsigned long ua, unsigned long entries,
- * of the CMA zone if possible. NOTE: faulting in + migration
- * can be expensive. Batching can be considered later
- */
-- if (get_pageblock_migratetype(page) == MIGRATE_CMA) {
-+ if (is_migrate_cma_page(page)) {
- if (mm_iommu_move_page_from_cma(page))
- goto populate;
- if (1 != get_user_pages_fast(ua + (i << PAGE_SHIFT),
-@@ -215,11 +209,11 @@ long mm_iommu_get(unsigned long ua, unsigned long entries,
- mem->entries = entries;
- *pmem = mem;
-
-- list_add_rcu(&mem->next, &current->mm->context.iommu_group_mem_list);
-+ list_add_rcu(&mem->next, &mm->context.iommu_group_mem_list);
-
- unlock_exit:
- if (locked_entries && ret)
-- mm_iommu_adjust_locked_vm(current->mm, locked_entries, false);
-+ mm_iommu_adjust_locked_vm(mm, locked_entries, false);
-
- mutex_unlock(&mem_list_mutex);
-
-@@ -264,17 +258,13 @@ static void mm_iommu_free(struct rcu_head *head)
- static void mm_iommu_release(struct mm_iommu_table_group_mem_t *mem)
- {
- list_del_rcu(&mem->next);
-- mm_iommu_adjust_locked_vm(current->mm, mem->entries, false);
- call_rcu(&mem->rcu, mm_iommu_free);
- }
-
--long mm_iommu_put(struct mm_iommu_table_group_mem_t *mem)
-+long mm_iommu_put(struct mm_struct *mm, struct mm_iommu_table_group_mem_t *mem)
- {
- long ret = 0;
-
-- if (!current || !current->mm)
-- return -ESRCH; /* process exited */
--
- mutex_lock(&mem_list_mutex);
-
- if (mem->used == 0) {
-@@ -297,6 +287,8 @@ long mm_iommu_put(struct mm_iommu_table_group_mem_t *mem)
- /* @mapped became 0 so now mappings are disabled, release the region */
- mm_iommu_release(mem);
-
-+ mm_iommu_adjust_locked_vm(mm, mem->entries, false);
-+
- unlock_exit:
- mutex_unlock(&mem_list_mutex);
-
-@@ -304,14 +296,12 @@ long mm_iommu_put(struct mm_iommu_table_group_mem_t *mem)
- }
- EXPORT_SYMBOL_GPL(mm_iommu_put);
-
--struct mm_iommu_table_group_mem_t *mm_iommu_lookup(unsigned long ua,
-- unsigned long size)
-+struct mm_iommu_table_group_mem_t *mm_iommu_lookup(struct mm_struct *mm,
-+ unsigned long ua, unsigned long size)
- {
- struct mm_iommu_table_group_mem_t *mem, *ret = NULL;
-
-- list_for_each_entry_rcu(mem,
-- &current->mm->context.iommu_group_mem_list,
-- next) {
-+ list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next) {
- if ((mem->ua <= ua) &&
- (ua + size <= mem->ua +
- (mem->entries << PAGE_SHIFT))) {
-@@ -324,14 +314,12 @@ struct mm_iommu_table_group_mem_t *mm_iommu_lookup(unsigned long ua,
- }
- EXPORT_SYMBOL_GPL(mm_iommu_lookup);
-
--struct mm_iommu_table_group_mem_t *mm_iommu_find(unsigned long ua,
-- unsigned long entries)
-+struct mm_iommu_table_group_mem_t *mm_iommu_find(struct mm_struct *mm,
-+ unsigned long ua, unsigned long entries)
- {
- struct mm_iommu_table_group_mem_t *mem, *ret = NULL;
-
-- list_for_each_entry_rcu(mem,
-- &current->mm->context.iommu_group_mem_list,
-- next) {
-+ list_for_each_entry_rcu(mem, &mm->context.iommu_group_mem_list, next) {
- if ((mem->ua == ua) && (mem->entries == entries)) {
- ret = mem;
- break;
-@@ -373,17 +361,7 @@ void mm_iommu_mapped_dec(struct mm_iommu_table_group_mem_t *mem)
- }
- EXPORT_SYMBOL_GPL(mm_iommu_mapped_dec);
-
--void mm_iommu_init(mm_context_t *ctx)
-+void mm_iommu_init(struct mm_struct *mm)
- {
-- INIT_LIST_HEAD_RCU(&ctx->iommu_group_mem_list);
--}
--
--void mm_iommu_cleanup(mm_context_t *ctx)
--{
-- struct mm_iommu_table_group_mem_t *mem, *tmp;
--
-- list_for_each_entry_safe(mem, tmp, &ctx->iommu_group_mem_list, next) {
-- list_del_rcu(&mem->next);
-- mm_iommu_do_free(mem);
-- }
-+ INIT_LIST_HEAD_RCU(&mm->context.iommu_group_mem_list);
- }
-diff --git a/arch/x86/events/core.c b/arch/x86/events/core.c
-index 7fe88bb..38623e2 100644
---- a/arch/x86/events/core.c
-+++ b/arch/x86/events/core.c
-@@ -2096,8 +2096,8 @@ static int x86_pmu_event_init(struct perf_event *event)
-
- static void refresh_pce(void *ignored)
- {
-- if (current->mm)
-- load_mm_cr4(current->mm);
-+ if (current->active_mm)
-+ load_mm_cr4(current->active_mm);
- }
-
- static void x86_pmu_event_mapped(struct perf_event *event)
-diff --git a/arch/x86/kernel/cpu/mshyperv.c b/arch/x86/kernel/cpu/mshyperv.c
-index 8f44c5a..f228f74 100644
---- a/arch/x86/kernel/cpu/mshyperv.c
-+++ b/arch/x86/kernel/cpu/mshyperv.c
-@@ -31,6 +31,7 @@
- #include <asm/apic.h>
- #include <asm/timer.h>
- #include <asm/reboot.h>
-+#include <asm/nmi.h>
-
- struct ms_hyperv_info ms_hyperv;
- EXPORT_SYMBOL_GPL(ms_hyperv);
-@@ -158,6 +159,26 @@ static unsigned char hv_get_nmi_reason(void)
- return 0;
- }
-
-+#ifdef CONFIG_X86_LOCAL_APIC
-+/*
-+ * Prior to WS2016 Debug-VM sends NMIs to all CPUs which makes
-+ * it dificult to process CHANNELMSG_UNLOAD in case of crash. Handle
-+ * unknown NMI on the first CPU which gets it.
-+ */
-+static int hv_nmi_unknown(unsigned int val, struct pt_regs *regs)
-+{
-+ static atomic_t nmi_cpu = ATOMIC_INIT(-1);
-+
-+ if (!unknown_nmi_panic)
-+ return NMI_DONE;
-+
-+ if (atomic_cmpxchg(&nmi_cpu, -1, raw_smp_processor_id()) != -1)
-+ return NMI_HANDLED;
-+
-+ return NMI_DONE;
-+}
-+#endif
-+
- static void __init ms_hyperv_init_platform(void)
- {
- /*
-@@ -183,6 +204,9 @@ static void __init ms_hyperv_init_platform(void)
- pr_info("HyperV: LAPIC Timer Frequency: %#x\n",
- lapic_timer_frequency);
- }
-+
-+ register_nmi_handler(NMI_UNKNOWN, hv_nmi_unknown, NMI_FLAG_FIRST,
-+ "hv_nmi_unknown");
- #endif
-
- if (ms_hyperv.features & HV_X64_MSR_TIME_REF_COUNT_AVAILABLE)
-diff --git a/arch/x86/kernel/head64.c b/arch/x86/kernel/head64.c
-index 54a2372..b5785c1 100644
---- a/arch/x86/kernel/head64.c
-+++ b/arch/x86/kernel/head64.c
-@@ -4,6 +4,7 @@
- * Copyright (C) 2000 Andrea Arcangeli <andrea@suse.de> SuSE
- */
-
-+#define DISABLE_BRANCH_PROFILING
- #include <linux/init.h>
- #include <linux/linkage.h>
- #include <linux/types.h>
-diff --git a/arch/x86/kernel/tsc.c b/arch/x86/kernel/tsc.c
-index 46b2f41..eea88fe 100644
---- a/arch/x86/kernel/tsc.c
-+++ b/arch/x86/kernel/tsc.c
-@@ -1287,6 +1287,8 @@ static int __init init_tsc_clocksource(void)
- * exporting a reliable TSC.
- */
- if (boot_cpu_has(X86_FEATURE_TSC_RELIABLE)) {
-+ if (boot_cpu_has(X86_FEATURE_ART))
-+ art_related_clocksource = &clocksource_tsc;
- clocksource_register_khz(&clocksource_tsc, tsc_khz);
- return 0;
- }
-diff --git a/arch/x86/mm/kasan_init_64.c b/arch/x86/mm/kasan_init_64.c
-index 0493c17..333362f 100644
---- a/arch/x86/mm/kasan_init_64.c
-+++ b/arch/x86/mm/kasan_init_64.c
-@@ -1,3 +1,4 @@
-+#define DISABLE_BRANCH_PROFILING
- #define pr_fmt(fmt) "kasan: " fmt
- #include <linux/bootmem.h>
- #include <linux/kasan.h>
-diff --git a/arch/x86/pci/xen.c b/arch/x86/pci/xen.c
-index bedfab9..a00a6c0 100644
---- a/arch/x86/pci/xen.c
-+++ b/arch/x86/pci/xen.c
-@@ -234,23 +234,14 @@ static int xen_hvm_setup_msi_irqs(struct pci_dev *dev, int nvec, int type)
- return 1;
-
- for_each_pci_msi_entry(msidesc, dev) {
-- __pci_read_msi_msg(msidesc, &msg);
-- pirq = MSI_ADDR_EXT_DEST_ID(msg.address_hi) |
-- ((msg.address_lo >> MSI_ADDR_DEST_ID_SHIFT) & 0xff);
-- if (msg.data != XEN_PIRQ_MSI_DATA ||
-- xen_irq_from_pirq(pirq) < 0) {
-- pirq = xen_allocate_pirq_msi(dev, msidesc);
-- if (pirq < 0) {
-- irq = -ENODEV;
-- goto error;
-- }
-- xen_msi_compose_msg(dev, pirq, &msg);
-- __pci_write_msi_msg(msidesc, &msg);
-- dev_dbg(&dev->dev, "xen: msi bound to pirq=%d\n", pirq);
-- } else {
-- dev_dbg(&dev->dev,
-- "xen: msi already bound to pirq=%d\n", pirq);
-+ pirq = xen_allocate_pirq_msi(dev, msidesc);
-+ if (pirq < 0) {
-+ irq = -ENODEV;
-+ goto error;
- }
-+ xen_msi_compose_msg(dev, pirq, &msg);
-+ __pci_write_msi_msg(msidesc, &msg);
-+ dev_dbg(&dev->dev, "xen: msi bound to pirq=%d\n", pirq);
- irq = xen_bind_pirq_msi_to_irq(dev, msidesc, pirq,
- (type == PCI_CAP_ID_MSI) ? nvec : 1,
- (type == PCI_CAP_ID_MSIX) ?
-diff --git a/block/scsi_ioctl.c b/block/scsi_ioctl.c
-index 0774799..c6fee74 100644
---- a/block/scsi_ioctl.c
-+++ b/block/scsi_ioctl.c
-@@ -182,6 +182,9 @@ static void blk_set_cmd_filter_defaults(struct blk_cmd_filter *filter)
- __set_bit(WRITE_16, filter->write_ok);
- __set_bit(WRITE_LONG, filter->write_ok);
- __set_bit(WRITE_LONG_2, filter->write_ok);
-+ __set_bit(WRITE_SAME, filter->write_ok);
-+ __set_bit(WRITE_SAME_16, filter->write_ok);
-+ __set_bit(WRITE_SAME_32, filter->write_ok);
- __set_bit(ERASE, filter->write_ok);
- __set_bit(GPCMD_MODE_SELECT_10, filter->write_ok);
- __set_bit(MODE_SELECT, filter->write_ok);
-diff --git a/drivers/acpi/blacklist.c b/drivers/acpi/blacklist.c
-index bdc67ba..4421f7c 100644
---- a/drivers/acpi/blacklist.c
-+++ b/drivers/acpi/blacklist.c
-@@ -160,6 +160,34 @@ static struct dmi_system_id acpi_rev_dmi_table[] __initdata = {
- DMI_MATCH(DMI_PRODUCT_NAME, "XPS 13 9343"),
- },
- },
-+ {
-+ .callback = dmi_enable_rev_override,
-+ .ident = "DELL Precision 5520",
-+ .matches = {
-+ DMI_MATCH(DMI_SYS_VENDOR, "Dell Inc."),
-+ DMI_MATCH(DMI_PRODUCT_NAME, "Precision 5520"),
-+ },
-+ },
-+ {
-+ .callback = dmi_enable_rev_override,
-+ .ident = "DELL Precision 3520",
-+ .matches = {
-+ DMI_MATCH(DMI_SYS_VENDOR, "Dell Inc."),
-+ DMI_MATCH(DMI_PRODUCT_NAME, "Precision 3520"),
-+ },
-+ },
-+ /*
-+ * Resolves a quirk with the Dell Latitude 3350 that
-+ * causes the ethernet adapter to not function.
-+ */
-+ {
-+ .callback = dmi_enable_rev_override,
-+ .ident = "DELL Latitude 3350",
-+ .matches = {
-+ DMI_MATCH(DMI_SYS_VENDOR, "Dell Inc."),
-+ DMI_MATCH(DMI_PRODUCT_NAME, "Latitude 3350"),
-+ },
-+ },
- #endif
- {}
- };
-diff --git a/drivers/clk/bcm/clk-bcm2835.c b/drivers/clk/bcm/clk-bcm2835.c
-index 3bbd2a5..2acaa77 100644
---- a/drivers/clk/bcm/clk-bcm2835.c
-+++ b/drivers/clk/bcm/clk-bcm2835.c
-@@ -1598,7 +1598,7 @@ static const struct bcm2835_clk_desc clk_desc_array[] = {
- .a2w_reg = A2W_PLLH_AUX,
- .load_mask = CM_PLLH_LOADAUX,
- .hold_mask = 0,
-- .fixed_divider = 10),
-+ .fixed_divider = 1),
- [BCM2835_PLLH_PIX] = REGISTER_PLL_DIV(
- .name = "pllh_pix",
- .source_pll = "pllh",
-diff --git a/drivers/dma/ioat/init.c b/drivers/dma/ioat/init.c
-index 015f711..d235fbe 100644
---- a/drivers/dma/ioat/init.c
-+++ b/drivers/dma/ioat/init.c
-@@ -691,7 +691,7 @@ static int ioat_alloc_chan_resources(struct dma_chan *c)
- /* doing 2 32bit writes to mmio since 1 64b write doesn't work */
- ioat_chan->completion =
- dma_pool_zalloc(ioat_chan->ioat_dma->completion_pool,
-- GFP_KERNEL, &ioat_chan->completion_dma);
-+ GFP_NOWAIT, &ioat_chan->completion_dma);
- if (!ioat_chan->completion)
- return -ENOMEM;
-
-@@ -701,7 +701,7 @@ static int ioat_alloc_chan_resources(struct dma_chan *c)
- ioat_chan->reg_base + IOAT_CHANCMP_OFFSET_HIGH);
-
- order = IOAT_MAX_ORDER;
-- ring = ioat_alloc_ring(c, order, GFP_KERNEL);
-+ ring = ioat_alloc_ring(c, order, GFP_NOWAIT);
- if (!ring)
- return -ENOMEM;
-
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/Kbuild b/drivers/gpu/drm/nouveau/nvkm/engine/disp/Kbuild
-index 77a52b5..70f0344 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/Kbuild
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/Kbuild
-@@ -95,9 +95,11 @@ nvkm-y += nvkm/engine/disp/cursg84.o
- nvkm-y += nvkm/engine/disp/cursgt215.o
- nvkm-y += nvkm/engine/disp/cursgf119.o
- nvkm-y += nvkm/engine/disp/cursgk104.o
-+nvkm-y += nvkm/engine/disp/cursgp102.o
-
- nvkm-y += nvkm/engine/disp/oimmnv50.o
- nvkm-y += nvkm/engine/disp/oimmg84.o
- nvkm-y += nvkm/engine/disp/oimmgt215.o
- nvkm-y += nvkm/engine/disp/oimmgf119.o
- nvkm-y += nvkm/engine/disp/oimmgk104.o
-+nvkm-y += nvkm/engine/disp/oimmgp102.o
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c
-index dd2953b..9d90d8b 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c
-@@ -82,7 +82,7 @@ nv50_disp_chan_mthd(struct nv50_disp_chan *chan, int debug)
-
- if (mthd->addr) {
- snprintf(cname_, sizeof(cname_), "%s %d",
-- mthd->name, chan->chid);
-+ mthd->name, chan->chid.user);
- cname = cname_;
- }
-
-@@ -139,7 +139,7 @@ nv50_disp_chan_uevent_ctor(struct nvkm_object *object, void *data, u32 size,
- if (!(ret = nvif_unvers(ret, &data, &size, args->none))) {
- notify->size = sizeof(struct nvif_notify_uevent_rep);
- notify->types = 1;
-- notify->index = chan->chid;
-+ notify->index = chan->chid.user;
- return 0;
- }
-
-@@ -159,7 +159,7 @@ nv50_disp_chan_rd32(struct nvkm_object *object, u64 addr, u32 *data)
- struct nv50_disp_chan *chan = nv50_disp_chan(object);
- struct nv50_disp *disp = chan->root->disp;
- struct nvkm_device *device = disp->base.engine.subdev.device;
-- *data = nvkm_rd32(device, 0x640000 + (chan->chid * 0x1000) + addr);
-+ *data = nvkm_rd32(device, 0x640000 + (chan->chid.user * 0x1000) + addr);
- return 0;
- }
-
-@@ -169,7 +169,7 @@ nv50_disp_chan_wr32(struct nvkm_object *object, u64 addr, u32 data)
- struct nv50_disp_chan *chan = nv50_disp_chan(object);
- struct nv50_disp *disp = chan->root->disp;
- struct nvkm_device *device = disp->base.engine.subdev.device;
-- nvkm_wr32(device, 0x640000 + (chan->chid * 0x1000) + addr, data);
-+ nvkm_wr32(device, 0x640000 + (chan->chid.user * 0x1000) + addr, data);
- return 0;
- }
-
-@@ -196,7 +196,7 @@ nv50_disp_chan_map(struct nvkm_object *object, u64 *addr, u32 *size)
- struct nv50_disp *disp = chan->root->disp;
- struct nvkm_device *device = disp->base.engine.subdev.device;
- *addr = device->func->resource_addr(device, 0) +
-- 0x640000 + (chan->chid * 0x1000);
-+ 0x640000 + (chan->chid.user * 0x1000);
- *size = 0x001000;
- return 0;
- }
-@@ -243,8 +243,8 @@ nv50_disp_chan_dtor(struct nvkm_object *object)
- {
- struct nv50_disp_chan *chan = nv50_disp_chan(object);
- struct nv50_disp *disp = chan->root->disp;
-- if (chan->chid >= 0)
-- disp->chan[chan->chid] = NULL;
-+ if (chan->chid.user >= 0)
-+ disp->chan[chan->chid.user] = NULL;
- return chan->func->dtor ? chan->func->dtor(chan) : chan;
- }
-
-@@ -263,7 +263,7 @@ nv50_disp_chan = {
- int
- nv50_disp_chan_ctor(const struct nv50_disp_chan_func *func,
- const struct nv50_disp_chan_mthd *mthd,
-- struct nv50_disp_root *root, int chid, int head,
-+ struct nv50_disp_root *root, int ctrl, int user, int head,
- const struct nvkm_oclass *oclass,
- struct nv50_disp_chan *chan)
- {
-@@ -273,21 +273,22 @@ nv50_disp_chan_ctor(const struct nv50_disp_chan_func *func,
- chan->func = func;
- chan->mthd = mthd;
- chan->root = root;
-- chan->chid = chid;
-+ chan->chid.ctrl = ctrl;
-+ chan->chid.user = user;
- chan->head = head;
-
-- if (disp->chan[chan->chid]) {
-- chan->chid = -1;
-+ if (disp->chan[chan->chid.user]) {
-+ chan->chid.user = -1;
- return -EBUSY;
- }
-- disp->chan[chan->chid] = chan;
-+ disp->chan[chan->chid.user] = chan;
- return 0;
- }
-
- int
- nv50_disp_chan_new_(const struct nv50_disp_chan_func *func,
- const struct nv50_disp_chan_mthd *mthd,
-- struct nv50_disp_root *root, int chid, int head,
-+ struct nv50_disp_root *root, int ctrl, int user, int head,
- const struct nvkm_oclass *oclass,
- struct nvkm_object **pobject)
- {
-@@ -297,5 +298,6 @@ nv50_disp_chan_new_(const struct nv50_disp_chan_func *func,
- return -ENOMEM;
- *pobject = &chan->object;
-
-- return nv50_disp_chan_ctor(func, mthd, root, chid, head, oclass, chan);
-+ return nv50_disp_chan_ctor(func, mthd, root, ctrl, user,
-+ head, oclass, chan);
- }
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.h b/drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.h
-index f5f683d..737b38f 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.h
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.h
-@@ -7,7 +7,11 @@ struct nv50_disp_chan {
- const struct nv50_disp_chan_func *func;
- const struct nv50_disp_chan_mthd *mthd;
- struct nv50_disp_root *root;
-- int chid;
-+
-+ struct {
-+ int ctrl;
-+ int user;
-+ } chid;
- int head;
-
- struct nvkm_object object;
-@@ -25,11 +29,11 @@ struct nv50_disp_chan_func {
-
- int nv50_disp_chan_ctor(const struct nv50_disp_chan_func *,
- const struct nv50_disp_chan_mthd *,
-- struct nv50_disp_root *, int chid, int head,
-+ struct nv50_disp_root *, int ctrl, int user, int head,
- const struct nvkm_oclass *, struct nv50_disp_chan *);
- int nv50_disp_chan_new_(const struct nv50_disp_chan_func *,
- const struct nv50_disp_chan_mthd *,
-- struct nv50_disp_root *, int chid, int head,
-+ struct nv50_disp_root *, int ctrl, int user, int head,
- const struct nvkm_oclass *, struct nvkm_object **);
-
- extern const struct nv50_disp_chan_func nv50_disp_pioc_func;
-@@ -90,13 +94,16 @@ extern const struct nv50_disp_chan_mthd gk104_disp_ovly_chan_mthd;
- struct nv50_disp_pioc_oclass {
- int (*ctor)(const struct nv50_disp_chan_func *,
- const struct nv50_disp_chan_mthd *,
-- struct nv50_disp_root *, int chid,
-+ struct nv50_disp_root *, int ctrl, int user,
- const struct nvkm_oclass *, void *data, u32 size,
- struct nvkm_object **);
- struct nvkm_sclass base;
- const struct nv50_disp_chan_func *func;
- const struct nv50_disp_chan_mthd *mthd;
-- int chid;
-+ struct {
-+ int ctrl;
-+ int user;
-+ } chid;
- };
-
- extern const struct nv50_disp_pioc_oclass nv50_disp_oimm_oclass;
-@@ -114,15 +121,17 @@ extern const struct nv50_disp_pioc_oclass gf119_disp_curs_oclass;
- extern const struct nv50_disp_pioc_oclass gk104_disp_oimm_oclass;
- extern const struct nv50_disp_pioc_oclass gk104_disp_curs_oclass;
-
-+extern const struct nv50_disp_pioc_oclass gp102_disp_oimm_oclass;
-+extern const struct nv50_disp_pioc_oclass gp102_disp_curs_oclass;
-
- int nv50_disp_curs_new(const struct nv50_disp_chan_func *,
- const struct nv50_disp_chan_mthd *,
-- struct nv50_disp_root *, int chid,
-+ struct nv50_disp_root *, int ctrl, int user,
- const struct nvkm_oclass *, void *data, u32 size,
- struct nvkm_object **);
- int nv50_disp_oimm_new(const struct nv50_disp_chan_func *,
- const struct nv50_disp_chan_mthd *,
-- struct nv50_disp_root *, int chid,
-+ struct nv50_disp_root *, int ctrl, int user,
- const struct nvkm_oclass *, void *data, u32 size,
- struct nvkm_object **);
- #endif
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursg84.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursg84.c
-index dd99fc7..fa781b5 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursg84.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursg84.c
-@@ -33,5 +33,5 @@ g84_disp_curs_oclass = {
- .base.maxver = 0,
- .ctor = nv50_disp_curs_new,
- .func = &nv50_disp_pioc_func,
-- .chid = 7,
-+ .chid = { 7, 7 },
- };
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgf119.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgf119.c
-index 2a1574e..2be6fb0 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgf119.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgf119.c
-@@ -33,5 +33,5 @@ gf119_disp_curs_oclass = {
- .base.maxver = 0,
- .ctor = nv50_disp_curs_new,
- .func = &gf119_disp_pioc_func,
-- .chid = 13,
-+ .chid = { 13, 13 },
- };
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgk104.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgk104.c
-index 28e8f06..2a99db4 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgk104.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgk104.c
-@@ -33,5 +33,5 @@ gk104_disp_curs_oclass = {
- .base.maxver = 0,
- .ctor = nv50_disp_curs_new,
- .func = &gf119_disp_pioc_func,
-- .chid = 13,
-+ .chid = { 13, 13 },
- };
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgp102.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgp102.c
-new file mode 100644
-index 0000000..e958210
---- /dev/null
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgp102.c
-@@ -0,0 +1,37 @@
-+/*
-+ * Copyright 2016 Red Hat Inc.
-+ *
-+ * Permission is hereby granted, free of charge, to any person obtaining a
-+ * copy of this software and associated documentation files (the "Software"),
-+ * to deal in the Software without restriction, including without limitation
-+ * the rights to use, copy, modify, merge, publish, distribute, sublicense,
-+ * and/or sell copies of the Software, and to permit persons to whom the
-+ * Software is furnished to do so, subject to the following conditions:
-+ *
-+ * The above copyright notice and this permission notice shall be included in
-+ * all copies or substantial portions of the Software.
-+ *
-+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
-+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
-+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
-+ * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
-+ * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
-+ * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-+ * OTHER DEALINGS IN THE SOFTWARE.
-+ *
-+ * Authors: Ben Skeggs <bskeggs@redhat.com>
-+ */
-+#include "channv50.h"
-+#include "rootnv50.h"
-+
-+#include <nvif/class.h>
-+
-+const struct nv50_disp_pioc_oclass
-+gp102_disp_curs_oclass = {
-+ .base.oclass = GK104_DISP_CURSOR,
-+ .base.minver = 0,
-+ .base.maxver = 0,
-+ .ctor = nv50_disp_curs_new,
-+ .func = &gf119_disp_pioc_func,
-+ .chid = { 13, 17 },
-+};
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgt215.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgt215.c
-index d8a4b9c..00a7f35 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgt215.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursgt215.c
-@@ -33,5 +33,5 @@ gt215_disp_curs_oclass = {
- .base.maxver = 0,
- .ctor = nv50_disp_curs_new,
- .func = &nv50_disp_pioc_func,
-- .chid = 7,
-+ .chid = { 7, 7 },
- };
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursnv50.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursnv50.c
-index 8b13204..82ff82d 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursnv50.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/cursnv50.c
-@@ -33,7 +33,7 @@
- int
- nv50_disp_curs_new(const struct nv50_disp_chan_func *func,
- const struct nv50_disp_chan_mthd *mthd,
-- struct nv50_disp_root *root, int chid,
-+ struct nv50_disp_root *root, int ctrl, int user,
- const struct nvkm_oclass *oclass, void *data, u32 size,
- struct nvkm_object **pobject)
- {
-@@ -54,7 +54,7 @@ nv50_disp_curs_new(const struct nv50_disp_chan_func *func,
- } else
- return ret;
-
-- return nv50_disp_chan_new_(func, mthd, root, chid + head,
-+ return nv50_disp_chan_new_(func, mthd, root, ctrl + head, user + head,
- head, oclass, pobject);
- }
-
-@@ -65,5 +65,5 @@ nv50_disp_curs_oclass = {
- .base.maxver = 0,
- .ctor = nv50_disp_curs_new,
- .func = &nv50_disp_pioc_func,
-- .chid = 7,
-+ .chid = { 7, 7 },
- };
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacgf119.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacgf119.c
-index a57f7ce..ce7cd74 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacgf119.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacgf119.c
-@@ -32,8 +32,8 @@ gf119_disp_dmac_bind(struct nv50_disp_dmac *chan,
- struct nvkm_object *object, u32 handle)
- {
- return nvkm_ramht_insert(chan->base.root->ramht, object,
-- chan->base.chid, -9, handle,
-- chan->base.chid << 27 | 0x00000001);
-+ chan->base.chid.user, -9, handle,
-+ chan->base.chid.user << 27 | 0x00000001);
- }
-
- void
-@@ -42,22 +42,23 @@ gf119_disp_dmac_fini(struct nv50_disp_dmac *chan)
- struct nv50_disp *disp = chan->base.root->disp;
- struct nvkm_subdev *subdev = &disp->base.engine.subdev;
- struct nvkm_device *device = subdev->device;
-- int chid = chan->base.chid;
-+ int ctrl = chan->base.chid.ctrl;
-+ int user = chan->base.chid.user;
-
- /* deactivate channel */
-- nvkm_mask(device, 0x610490 + (chid * 0x0010), 0x00001010, 0x00001000);
-- nvkm_mask(device, 0x610490 + (chid * 0x0010), 0x00000003, 0x00000000);
-+ nvkm_mask(device, 0x610490 + (ctrl * 0x0010), 0x00001010, 0x00001000);
-+ nvkm_mask(device, 0x610490 + (ctrl * 0x0010), 0x00000003, 0x00000000);
- if (nvkm_msec(device, 2000,
-- if (!(nvkm_rd32(device, 0x610490 + (chid * 0x10)) & 0x001e0000))
-+ if (!(nvkm_rd32(device, 0x610490 + (ctrl * 0x10)) & 0x001e0000))
- break;
- ) < 0) {
-- nvkm_error(subdev, "ch %d fini: %08x\n", chid,
-- nvkm_rd32(device, 0x610490 + (chid * 0x10)));
-+ nvkm_error(subdev, "ch %d fini: %08x\n", user,
-+ nvkm_rd32(device, 0x610490 + (ctrl * 0x10)));
- }
-
- /* disable error reporting and completion notification */
-- nvkm_mask(device, 0x610090, 0x00000001 << chid, 0x00000000);
-- nvkm_mask(device, 0x6100a0, 0x00000001 << chid, 0x00000000);
-+ nvkm_mask(device, 0x610090, 0x00000001 << user, 0x00000000);
-+ nvkm_mask(device, 0x6100a0, 0x00000001 << user, 0x00000000);
- }
-
- static int
-@@ -66,26 +67,27 @@ gf119_disp_dmac_init(struct nv50_disp_dmac *chan)
- struct nv50_disp *disp = chan->base.root->disp;
- struct nvkm_subdev *subdev = &disp->base.engine.subdev;
- struct nvkm_device *device = subdev->device;
-- int chid = chan->base.chid;
-+ int ctrl = chan->base.chid.ctrl;
-+ int user = chan->base.chid.user;
-
- /* enable error reporting */
-- nvkm_mask(device, 0x6100a0, 0x00000001 << chid, 0x00000001 << chid);
-+ nvkm_mask(device, 0x6100a0, 0x00000001 << user, 0x00000001 << user);
-
- /* initialise channel for dma command submission */
-- nvkm_wr32(device, 0x610494 + (chid * 0x0010), chan->push);
-- nvkm_wr32(device, 0x610498 + (chid * 0x0010), 0x00010000);
-- nvkm_wr32(device, 0x61049c + (chid * 0x0010), 0x00000001);
-- nvkm_mask(device, 0x610490 + (chid * 0x0010), 0x00000010, 0x00000010);
-- nvkm_wr32(device, 0x640000 + (chid * 0x1000), 0x00000000);
-- nvkm_wr32(device, 0x610490 + (chid * 0x0010), 0x00000013);
-+ nvkm_wr32(device, 0x610494 + (ctrl * 0x0010), chan->push);
-+ nvkm_wr32(device, 0x610498 + (ctrl * 0x0010), 0x00010000);
-+ nvkm_wr32(device, 0x61049c + (ctrl * 0x0010), 0x00000001);
-+ nvkm_mask(device, 0x610490 + (ctrl * 0x0010), 0x00000010, 0x00000010);
-+ nvkm_wr32(device, 0x640000 + (ctrl * 0x1000), 0x00000000);
-+ nvkm_wr32(device, 0x610490 + (ctrl * 0x0010), 0x00000013);
-
- /* wait for it to go inactive */
- if (nvkm_msec(device, 2000,
-- if (!(nvkm_rd32(device, 0x610490 + (chid * 0x10)) & 0x80000000))
-+ if (!(nvkm_rd32(device, 0x610490 + (ctrl * 0x10)) & 0x80000000))
- break;
- ) < 0) {
-- nvkm_error(subdev, "ch %d init: %08x\n", chid,
-- nvkm_rd32(device, 0x610490 + (chid * 0x10)));
-+ nvkm_error(subdev, "ch %d init: %08x\n", user,
-+ nvkm_rd32(device, 0x610490 + (ctrl * 0x10)));
- return -EBUSY;
- }
-
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacgp104.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacgp104.c
-index ad24c2c..d26d3b4 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacgp104.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacgp104.c
-@@ -32,26 +32,27 @@ gp104_disp_dmac_init(struct nv50_disp_dmac *chan)
- struct nv50_disp *disp = chan->base.root->disp;
- struct nvkm_subdev *subdev = &disp->base.engine.subdev;
- struct nvkm_device *device = subdev->device;
-- int chid = chan->base.chid;
-+ int ctrl = chan->base.chid.ctrl;
-+ int user = chan->base.chid.user;
-
- /* enable error reporting */
-- nvkm_mask(device, 0x6100a0, 0x00000001 << chid, 0x00000001 << chid);
-+ nvkm_mask(device, 0x6100a0, 0x00000001 << user, 0x00000001 << user);
-
- /* initialise channel for dma command submission */
-- nvkm_wr32(device, 0x611494 + (chid * 0x0010), chan->push);
-- nvkm_wr32(device, 0x611498 + (chid * 0x0010), 0x00010000);
-- nvkm_wr32(device, 0x61149c + (chid * 0x0010), 0x00000001);
-- nvkm_mask(device, 0x610490 + (chid * 0x0010), 0x00000010, 0x00000010);
-- nvkm_wr32(device, 0x640000 + (chid * 0x1000), 0x00000000);
-- nvkm_wr32(device, 0x610490 + (chid * 0x0010), 0x00000013);
-+ nvkm_wr32(device, 0x611494 + (ctrl * 0x0010), chan->push);
-+ nvkm_wr32(device, 0x611498 + (ctrl * 0x0010), 0x00010000);
-+ nvkm_wr32(device, 0x61149c + (ctrl * 0x0010), 0x00000001);
-+ nvkm_mask(device, 0x610490 + (ctrl * 0x0010), 0x00000010, 0x00000010);
-+ nvkm_wr32(device, 0x640000 + (ctrl * 0x1000), 0x00000000);
-+ nvkm_wr32(device, 0x610490 + (ctrl * 0x0010), 0x00000013);
-
- /* wait for it to go inactive */
- if (nvkm_msec(device, 2000,
-- if (!(nvkm_rd32(device, 0x610490 + (chid * 0x10)) & 0x80000000))
-+ if (!(nvkm_rd32(device, 0x610490 + (ctrl * 0x10)) & 0x80000000))
- break;
- ) < 0) {
-- nvkm_error(subdev, "ch %d init: %08x\n", chid,
-- nvkm_rd32(device, 0x610490 + (chid * 0x10)));
-+ nvkm_error(subdev, "ch %d init: %08x\n", user,
-+ nvkm_rd32(device, 0x610490 + (ctrl * 0x10)));
- return -EBUSY;
- }
-
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacnv50.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacnv50.c
-index 9c6645a..0a1381a 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacnv50.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacnv50.c
-@@ -149,7 +149,7 @@ nv50_disp_dmac_new_(const struct nv50_disp_dmac_func *func,
- chan->func = func;
-
- ret = nv50_disp_chan_ctor(&nv50_disp_dmac_func_, mthd, root,
-- chid, head, oclass, &chan->base);
-+ chid, chid, head, oclass, &chan->base);
- if (ret)
- return ret;
-
-@@ -179,9 +179,9 @@ nv50_disp_dmac_bind(struct nv50_disp_dmac *chan,
- struct nvkm_object *object, u32 handle)
- {
- return nvkm_ramht_insert(chan->base.root->ramht, object,
-- chan->base.chid, -10, handle,
-- chan->base.chid << 28 |
-- chan->base.chid);
-+ chan->base.chid.user, -10, handle,
-+ chan->base.chid.user << 28 |
-+ chan->base.chid.user);
- }
-
- static void
-@@ -190,21 +190,22 @@ nv50_disp_dmac_fini(struct nv50_disp_dmac *chan)
- struct nv50_disp *disp = chan->base.root->disp;
- struct nvkm_subdev *subdev = &disp->base.engine.subdev;
- struct nvkm_device *device = subdev->device;
-- int chid = chan->base.chid;
-+ int ctrl = chan->base.chid.ctrl;
-+ int user = chan->base.chid.user;
-
- /* deactivate channel */
-- nvkm_mask(device, 0x610200 + (chid * 0x0010), 0x00001010, 0x00001000);
-- nvkm_mask(device, 0x610200 + (chid * 0x0010), 0x00000003, 0x00000000);
-+ nvkm_mask(device, 0x610200 + (ctrl * 0x0010), 0x00001010, 0x00001000);
-+ nvkm_mask(device, 0x610200 + (ctrl * 0x0010), 0x00000003, 0x00000000);
- if (nvkm_msec(device, 2000,
-- if (!(nvkm_rd32(device, 0x610200 + (chid * 0x10)) & 0x001e0000))
-+ if (!(nvkm_rd32(device, 0x610200 + (ctrl * 0x10)) & 0x001e0000))
- break;
- ) < 0) {
-- nvkm_error(subdev, "ch %d fini timeout, %08x\n", chid,
-- nvkm_rd32(device, 0x610200 + (chid * 0x10)));
-+ nvkm_error(subdev, "ch %d fini timeout, %08x\n", user,
-+ nvkm_rd32(device, 0x610200 + (ctrl * 0x10)));
- }
-
- /* disable error reporting and completion notifications */
-- nvkm_mask(device, 0x610028, 0x00010001 << chid, 0x00000000 << chid);
-+ nvkm_mask(device, 0x610028, 0x00010001 << user, 0x00000000 << user);
- }
-
- static int
-@@ -213,26 +214,27 @@ nv50_disp_dmac_init(struct nv50_disp_dmac *chan)
- struct nv50_disp *disp = chan->base.root->disp;
- struct nvkm_subdev *subdev = &disp->base.engine.subdev;
- struct nvkm_device *device = subdev->device;
-- int chid = chan->base.chid;
-+ int ctrl = chan->base.chid.ctrl;
-+ int user = chan->base.chid.user;
-
- /* enable error reporting */
-- nvkm_mask(device, 0x610028, 0x00010000 << chid, 0x00010000 << chid);
-+ nvkm_mask(device, 0x610028, 0x00010000 << user, 0x00010000 << user);
-
- /* initialise channel for dma command submission */
-- nvkm_wr32(device, 0x610204 + (chid * 0x0010), chan->push);
-- nvkm_wr32(device, 0x610208 + (chid * 0x0010), 0x00010000);
-- nvkm_wr32(device, 0x61020c + (chid * 0x0010), chid);
-- nvkm_mask(device, 0x610200 + (chid * 0x0010), 0x00000010, 0x00000010);
-- nvkm_wr32(device, 0x640000 + (chid * 0x1000), 0x00000000);
-- nvkm_wr32(device, 0x610200 + (chid * 0x0010), 0x00000013);
-+ nvkm_wr32(device, 0x610204 + (ctrl * 0x0010), chan->push);
-+ nvkm_wr32(device, 0x610208 + (ctrl * 0x0010), 0x00010000);
-+ nvkm_wr32(device, 0x61020c + (ctrl * 0x0010), ctrl);
-+ nvkm_mask(device, 0x610200 + (ctrl * 0x0010), 0x00000010, 0x00000010);
-+ nvkm_wr32(device, 0x640000 + (ctrl * 0x1000), 0x00000000);
-+ nvkm_wr32(device, 0x610200 + (ctrl * 0x0010), 0x00000013);
-
- /* wait for it to go inactive */
- if (nvkm_msec(device, 2000,
-- if (!(nvkm_rd32(device, 0x610200 + (chid * 0x10)) & 0x80000000))
-+ if (!(nvkm_rd32(device, 0x610200 + (ctrl * 0x10)) & 0x80000000))
- break;
- ) < 0) {
-- nvkm_error(subdev, "ch %d init timeout, %08x\n", chid,
-- nvkm_rd32(device, 0x610200 + (chid * 0x10)));
-+ nvkm_error(subdev, "ch %d init timeout, %08x\n", user,
-+ nvkm_rd32(device, 0x610200 + (ctrl * 0x10)));
- return -EBUSY;
- }
-
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmg84.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmg84.c
-index 54a4ae8..5ad5d0f 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmg84.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmg84.c
-@@ -33,5 +33,5 @@ g84_disp_oimm_oclass = {
- .base.maxver = 0,
- .ctor = nv50_disp_oimm_new,
- .func = &nv50_disp_pioc_func,
-- .chid = 5,
-+ .chid = { 5, 5 },
- };
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgf119.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgf119.c
-index c658db5..1f9fd34 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgf119.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgf119.c
-@@ -33,5 +33,5 @@ gf119_disp_oimm_oclass = {
- .base.maxver = 0,
- .ctor = nv50_disp_oimm_new,
- .func = &gf119_disp_pioc_func,
-- .chid = 9,
-+ .chid = { 9, 9 },
- };
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgk104.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgk104.c
-index b1fde8c..0c09fe8 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgk104.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgk104.c
-@@ -33,5 +33,5 @@ gk104_disp_oimm_oclass = {
- .base.maxver = 0,
- .ctor = nv50_disp_oimm_new,
- .func = &gf119_disp_pioc_func,
-- .chid = 9,
-+ .chid = { 9, 9 },
- };
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgp102.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgp102.c
-new file mode 100644
-index 0000000..abf8236
---- /dev/null
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgp102.c
-@@ -0,0 +1,37 @@
-+/*
-+ * Copyright 2016 Red Hat Inc.
-+ *
-+ * Permission is hereby granted, free of charge, to any person obtaining a
-+ * copy of this software and associated documentation files (the "Software"),
-+ * to deal in the Software without restriction, including without limitation
-+ * the rights to use, copy, modify, merge, publish, distribute, sublicense,
-+ * and/or sell copies of the Software, and to permit persons to whom the
-+ * Software is furnished to do so, subject to the following conditions:
-+ *
-+ * The above copyright notice and this permission notice shall be included in
-+ * all copies or substantial portions of the Software.
-+ *
-+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
-+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
-+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
-+ * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
-+ * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
-+ * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
-+ * OTHER DEALINGS IN THE SOFTWARE.
-+ *
-+ * Authors: Ben Skeggs <bskeggs@redhat.com>
-+ */
-+#include "channv50.h"
-+#include "rootnv50.h"
-+
-+#include <nvif/class.h>
-+
-+const struct nv50_disp_pioc_oclass
-+gp102_disp_oimm_oclass = {
-+ .base.oclass = GK104_DISP_OVERLAY,
-+ .base.minver = 0,
-+ .base.maxver = 0,
-+ .ctor = nv50_disp_oimm_new,
-+ .func = &gf119_disp_pioc_func,
-+ .chid = { 9, 13 },
-+};
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgt215.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgt215.c
-index f4e7eb3..1281db2 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgt215.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmgt215.c
-@@ -33,5 +33,5 @@ gt215_disp_oimm_oclass = {
- .base.maxver = 0,
- .ctor = nv50_disp_oimm_new,
- .func = &nv50_disp_pioc_func,
-- .chid = 5,
-+ .chid = { 5, 5 },
- };
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmnv50.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmnv50.c
-index 3940b9c..07540f3 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmnv50.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/oimmnv50.c
-@@ -33,7 +33,7 @@
- int
- nv50_disp_oimm_new(const struct nv50_disp_chan_func *func,
- const struct nv50_disp_chan_mthd *mthd,
-- struct nv50_disp_root *root, int chid,
-+ struct nv50_disp_root *root, int ctrl, int user,
- const struct nvkm_oclass *oclass, void *data, u32 size,
- struct nvkm_object **pobject)
- {
-@@ -54,7 +54,7 @@ nv50_disp_oimm_new(const struct nv50_disp_chan_func *func,
- } else
- return ret;
-
-- return nv50_disp_chan_new_(func, mthd, root, chid + head,
-+ return nv50_disp_chan_new_(func, mthd, root, ctrl + head, user + head,
- head, oclass, pobject);
- }
-
-@@ -65,5 +65,5 @@ nv50_disp_oimm_oclass = {
- .base.maxver = 0,
- .ctor = nv50_disp_oimm_new,
- .func = &nv50_disp_pioc_func,
-- .chid = 5,
-+ .chid = { 5, 5 },
- };
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/piocgf119.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/piocgf119.c
-index a625a98..0abaa64 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/piocgf119.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/piocgf119.c
-@@ -32,20 +32,21 @@ gf119_disp_pioc_fini(struct nv50_disp_chan *chan)
- struct nv50_disp *disp = chan->root->disp;
- struct nvkm_subdev *subdev = &disp->base.engine.subdev;
- struct nvkm_device *device = subdev->device;
-- int chid = chan->chid;
-+ int ctrl = chan->chid.ctrl;
-+ int user = chan->chid.user;
-
-- nvkm_mask(device, 0x610490 + (chid * 0x10), 0x00000001, 0x00000000);
-+ nvkm_mask(device, 0x610490 + (ctrl * 0x10), 0x00000001, 0x00000000);
- if (nvkm_msec(device, 2000,
-- if (!(nvkm_rd32(device, 0x610490 + (chid * 0x10)) & 0x00030000))
-+ if (!(nvkm_rd32(device, 0x610490 + (ctrl * 0x10)) & 0x00030000))
- break;
- ) < 0) {
-- nvkm_error(subdev, "ch %d fini: %08x\n", chid,
-- nvkm_rd32(device, 0x610490 + (chid * 0x10)));
-+ nvkm_error(subdev, "ch %d fini: %08x\n", user,
-+ nvkm_rd32(device, 0x610490 + (ctrl * 0x10)));
- }
-
- /* disable error reporting and completion notification */
-- nvkm_mask(device, 0x610090, 0x00000001 << chid, 0x00000000);
-- nvkm_mask(device, 0x6100a0, 0x00000001 << chid, 0x00000000);
-+ nvkm_mask(device, 0x610090, 0x00000001 << user, 0x00000000);
-+ nvkm_mask(device, 0x6100a0, 0x00000001 << user, 0x00000000);
- }
-
- static int
-@@ -54,20 +55,21 @@ gf119_disp_pioc_init(struct nv50_disp_chan *chan)
- struct nv50_disp *disp = chan->root->disp;
- struct nvkm_subdev *subdev = &disp->base.engine.subdev;
- struct nvkm_device *device = subdev->device;
-- int chid = chan->chid;
-+ int ctrl = chan->chid.ctrl;
-+ int user = chan->chid.user;
-
- /* enable error reporting */
-- nvkm_mask(device, 0x6100a0, 0x00000001 << chid, 0x00000001 << chid);
-+ nvkm_mask(device, 0x6100a0, 0x00000001 << user, 0x00000001 << user);
-
- /* activate channel */
-- nvkm_wr32(device, 0x610490 + (chid * 0x10), 0x00000001);
-+ nvkm_wr32(device, 0x610490 + (ctrl * 0x10), 0x00000001);
- if (nvkm_msec(device, 2000,
-- u32 tmp = nvkm_rd32(device, 0x610490 + (chid * 0x10));
-+ u32 tmp = nvkm_rd32(device, 0x610490 + (ctrl * 0x10));
- if ((tmp & 0x00030000) == 0x00010000)
- break;
- ) < 0) {
-- nvkm_error(subdev, "ch %d init: %08x\n", chid,
-- nvkm_rd32(device, 0x610490 + (chid * 0x10)));
-+ nvkm_error(subdev, "ch %d init: %08x\n", user,
-+ nvkm_rd32(device, 0x610490 + (ctrl * 0x10)));
- return -EBUSY;
- }
-
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/piocnv50.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/piocnv50.c
-index 9d2618d..0211e0e 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/piocnv50.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/piocnv50.c
-@@ -32,15 +32,16 @@ nv50_disp_pioc_fini(struct nv50_disp_chan *chan)
- struct nv50_disp *disp = chan->root->disp;
- struct nvkm_subdev *subdev = &disp->base.engine.subdev;
- struct nvkm_device *device = subdev->device;
-- int chid = chan->chid;
-+ int ctrl = chan->chid.ctrl;
-+ int user = chan->chid.user;
-
-- nvkm_mask(device, 0x610200 + (chid * 0x10), 0x00000001, 0x00000000);
-+ nvkm_mask(device, 0x610200 + (ctrl * 0x10), 0x00000001, 0x00000000);
- if (nvkm_msec(device, 2000,
-- if (!(nvkm_rd32(device, 0x610200 + (chid * 0x10)) & 0x00030000))
-+ if (!(nvkm_rd32(device, 0x610200 + (ctrl * 0x10)) & 0x00030000))
- break;
- ) < 0) {
-- nvkm_error(subdev, "ch %d timeout: %08x\n", chid,
-- nvkm_rd32(device, 0x610200 + (chid * 0x10)));
-+ nvkm_error(subdev, "ch %d timeout: %08x\n", user,
-+ nvkm_rd32(device, 0x610200 + (ctrl * 0x10)));
- }
- }
-
-@@ -50,26 +51,27 @@ nv50_disp_pioc_init(struct nv50_disp_chan *chan)
- struct nv50_disp *disp = chan->root->disp;
- struct nvkm_subdev *subdev = &disp->base.engine.subdev;
- struct nvkm_device *device = subdev->device;
-- int chid = chan->chid;
-+ int ctrl = chan->chid.ctrl;
-+ int user = chan->chid.user;
-
-- nvkm_wr32(device, 0x610200 + (chid * 0x10), 0x00002000);
-+ nvkm_wr32(device, 0x610200 + (ctrl * 0x10), 0x00002000);
- if (nvkm_msec(device, 2000,
-- if (!(nvkm_rd32(device, 0x610200 + (chid * 0x10)) & 0x00030000))
-+ if (!(nvkm_rd32(device, 0x610200 + (ctrl * 0x10)) & 0x00030000))
- break;
- ) < 0) {
-- nvkm_error(subdev, "ch %d timeout0: %08x\n", chid,
-- nvkm_rd32(device, 0x610200 + (chid * 0x10)));
-+ nvkm_error(subdev, "ch %d timeout0: %08x\n", user,
-+ nvkm_rd32(device, 0x610200 + (ctrl * 0x10)));
- return -EBUSY;
- }
-
-- nvkm_wr32(device, 0x610200 + (chid * 0x10), 0x00000001);
-+ nvkm_wr32(device, 0x610200 + (ctrl * 0x10), 0x00000001);
- if (nvkm_msec(device, 2000,
-- u32 tmp = nvkm_rd32(device, 0x610200 + (chid * 0x10));
-+ u32 tmp = nvkm_rd32(device, 0x610200 + (ctrl * 0x10));
- if ((tmp & 0x00030000) == 0x00010000)
- break;
- ) < 0) {
-- nvkm_error(subdev, "ch %d timeout1: %08x\n", chid,
-- nvkm_rd32(device, 0x610200 + (chid * 0x10)));
-+ nvkm_error(subdev, "ch %d timeout1: %08x\n", user,
-+ nvkm_rd32(device, 0x610200 + (ctrl * 0x10)));
- return -EBUSY;
- }
-
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgp104.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgp104.c
-index 8443e04..b053b29 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgp104.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgp104.c
-@@ -36,8 +36,8 @@ gp104_disp_root = {
- &gp104_disp_ovly_oclass,
- },
- .pioc = {
-- &gk104_disp_oimm_oclass,
-- &gk104_disp_curs_oclass,
-+ &gp102_disp_oimm_oclass,
-+ &gp102_disp_curs_oclass,
- },
- };
-
-diff --git a/drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c b/drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c
-index 2f9cecd..05c829a 100644
---- a/drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c
-+++ b/drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c
-@@ -207,8 +207,8 @@ nv50_disp_root_pioc_new_(const struct nvkm_oclass *oclass,
- {
- const struct nv50_disp_pioc_oclass *sclass = oclass->priv;
- struct nv50_disp_root *root = nv50_disp_root(oclass->parent);
-- return sclass->ctor(sclass->func, sclass->mthd, root, sclass->chid,
-- oclass, data, size, pobject);
-+ return sclass->ctor(sclass->func, sclass->mthd, root, sclass->chid.ctrl,
-+ sclass->chid.user, oclass, data, size, pobject);
- }
-
- static int
-diff --git a/drivers/gpu/drm/vc4/vc4_crtc.c b/drivers/gpu/drm/vc4/vc4_crtc.c
-index d544ff9..7aadce1 100644
---- a/drivers/gpu/drm/vc4/vc4_crtc.c
-+++ b/drivers/gpu/drm/vc4/vc4_crtc.c
-@@ -83,8 +83,7 @@ struct vc4_crtc_data {
- /* Which channel of the HVS this pixelvalve sources from. */
- int hvs_channel;
-
-- enum vc4_encoder_type encoder0_type;
-- enum vc4_encoder_type encoder1_type;
-+ enum vc4_encoder_type encoder_types[4];
- };
-
- #define CRTC_WRITE(offset, val) writel(val, vc4_crtc->regs + (offset))
-@@ -669,6 +668,14 @@ void vc4_disable_vblank(struct drm_device *dev, unsigned int crtc_id)
- CRTC_WRITE(PV_INTEN, 0);
- }
-
-+/* Must be called with the event lock held */
-+bool vc4_event_pending(struct drm_crtc *crtc)
-+{
-+ struct vc4_crtc *vc4_crtc = to_vc4_crtc(crtc);
-+
-+ return !!vc4_crtc->event;
-+}
-+
- static void vc4_crtc_handle_page_flip(struct vc4_crtc *vc4_crtc)
- {
- struct drm_crtc *crtc = &vc4_crtc->base;
-@@ -859,20 +866,26 @@ static const struct drm_crtc_helper_funcs vc4_crtc_helper_funcs = {
-
- static const struct vc4_crtc_data pv0_data = {
- .hvs_channel = 0,
-- .encoder0_type = VC4_ENCODER_TYPE_DSI0,
-- .encoder1_type = VC4_ENCODER_TYPE_DPI,
-+ .encoder_types = {
-+ [PV_CONTROL_CLK_SELECT_DSI] = VC4_ENCODER_TYPE_DSI0,
-+ [PV_CONTROL_CLK_SELECT_DPI_SMI_HDMI] = VC4_ENCODER_TYPE_DPI,
-+ },
- };
-
- static const struct vc4_crtc_data pv1_data = {
- .hvs_channel = 2,
-- .encoder0_type = VC4_ENCODER_TYPE_DSI1,
-- .encoder1_type = VC4_ENCODER_TYPE_SMI,
-+ .encoder_types = {
-+ [PV_CONTROL_CLK_SELECT_DSI] = VC4_ENCODER_TYPE_DSI1,
-+ [PV_CONTROL_CLK_SELECT_DPI_SMI_HDMI] = VC4_ENCODER_TYPE_SMI,
-+ },
- };
-
- static const struct vc4_crtc_data pv2_data = {
- .hvs_channel = 1,
-- .encoder0_type = VC4_ENCODER_TYPE_VEC,
-- .encoder1_type = VC4_ENCODER_TYPE_HDMI,
-+ .encoder_types = {
-+ [PV_CONTROL_CLK_SELECT_DPI_SMI_HDMI] = VC4_ENCODER_TYPE_HDMI,
-+ [PV_CONTROL_CLK_SELECT_VEC] = VC4_ENCODER_TYPE_VEC,
-+ },
- };
-
- static const struct of_device_id vc4_crtc_dt_match[] = {
-@@ -886,17 +899,20 @@ static void vc4_set_crtc_possible_masks(struct drm_device *drm,
- struct drm_crtc *crtc)
- {
- struct vc4_crtc *vc4_crtc = to_vc4_crtc(crtc);
-+ const struct vc4_crtc_data *crtc_data = vc4_crtc->data;
-+ const enum vc4_encoder_type *encoder_types = crtc_data->encoder_types;
- struct drm_encoder *encoder;
-
- drm_for_each_encoder(encoder, drm) {
- struct vc4_encoder *vc4_encoder = to_vc4_encoder(encoder);
--
-- if (vc4_encoder->type == vc4_crtc->data->encoder0_type) {
-- vc4_encoder->clock_select = 0;
-- encoder->possible_crtcs |= drm_crtc_mask(crtc);
-- } else if (vc4_encoder->type == vc4_crtc->data->encoder1_type) {
-- vc4_encoder->clock_select = 1;
-- encoder->possible_crtcs |= drm_crtc_mask(crtc);
-+ int i;
-+
-+ for (i = 0; i < ARRAY_SIZE(crtc_data->encoder_types); i++) {
-+ if (vc4_encoder->type == encoder_types[i]) {
-+ vc4_encoder->clock_select = i;
-+ encoder->possible_crtcs |= drm_crtc_mask(crtc);
-+ break;
-+ }
- }
- }
- }
-diff --git a/drivers/gpu/drm/vc4/vc4_drv.h b/drivers/gpu/drm/vc4/vc4_drv.h
-index 7c1e4d9..50a55ef 100644
---- a/drivers/gpu/drm/vc4/vc4_drv.h
-+++ b/drivers/gpu/drm/vc4/vc4_drv.h
-@@ -194,6 +194,7 @@ to_vc4_plane(struct drm_plane *plane)
- }
-
- enum vc4_encoder_type {
-+ VC4_ENCODER_TYPE_NONE,
- VC4_ENCODER_TYPE_HDMI,
- VC4_ENCODER_TYPE_VEC,
- VC4_ENCODER_TYPE_DSI0,
-@@ -440,6 +441,7 @@ int vc4_bo_stats_debugfs(struct seq_file *m, void *arg);
- extern struct platform_driver vc4_crtc_driver;
- int vc4_enable_vblank(struct drm_device *dev, unsigned int crtc_id);
- void vc4_disable_vblank(struct drm_device *dev, unsigned int crtc_id);
-+bool vc4_event_pending(struct drm_crtc *crtc);
- int vc4_crtc_debugfs_regs(struct seq_file *m, void *arg);
- int vc4_crtc_get_scanoutpos(struct drm_device *dev, unsigned int crtc_id,
- unsigned int flags, int *vpos, int *hpos,
-diff --git a/drivers/gpu/drm/vc4/vc4_kms.c b/drivers/gpu/drm/vc4/vc4_kms.c
-index c1f65c6..67af2af 100644
---- a/drivers/gpu/drm/vc4/vc4_kms.c
-+++ b/drivers/gpu/drm/vc4/vc4_kms.c
-@@ -119,17 +119,34 @@ static int vc4_atomic_commit(struct drm_device *dev,
-
- /* Make sure that any outstanding modesets have finished. */
- if (nonblock) {
-- ret = down_trylock(&vc4->async_modeset);
-- if (ret) {
-+ struct drm_crtc *crtc;
-+ struct drm_crtc_state *crtc_state;
-+ unsigned long flags;
-+ bool busy = false;
-+
-+ /*
-+ * If there's an undispatched event to send then we're
-+ * obviously still busy. If there isn't, then we can
-+ * unconditionally wait for the semaphore because it
-+ * shouldn't be contended (for long).
-+ *
-+ * This is to prevent a race where queuing a new flip
-+ * from userspace immediately on receipt of an event
-+ * beats our clean-up and returns EBUSY.
-+ */
-+ spin_lock_irqsave(&dev->event_lock, flags);
-+ for_each_crtc_in_state(state, crtc, crtc_state, i)
-+ busy |= vc4_event_pending(crtc);
-+ spin_unlock_irqrestore(&dev->event_lock, flags);
-+ if (busy) {
- kfree(c);
- return -EBUSY;
- }
-- } else {
-- ret = down_interruptible(&vc4->async_modeset);
-- if (ret) {
-- kfree(c);
-- return ret;
-- }
-+ }
-+ ret = down_interruptible(&vc4->async_modeset);
-+ if (ret) {
-+ kfree(c);
-+ return ret;
- }
-
- ret = drm_atomic_helper_prepare_planes(dev, state);
-diff --git a/drivers/gpu/drm/vc4/vc4_regs.h b/drivers/gpu/drm/vc4/vc4_regs.h
-index 1aa44c2..39f6886 100644
---- a/drivers/gpu/drm/vc4/vc4_regs.h
-+++ b/drivers/gpu/drm/vc4/vc4_regs.h
-@@ -177,8 +177,9 @@
- # define PV_CONTROL_WAIT_HSTART BIT(12)
- # define PV_CONTROL_PIXEL_REP_MASK VC4_MASK(5, 4)
- # define PV_CONTROL_PIXEL_REP_SHIFT 4
--# define PV_CONTROL_CLK_SELECT_DSI_VEC 0
-+# define PV_CONTROL_CLK_SELECT_DSI 0
- # define PV_CONTROL_CLK_SELECT_DPI_SMI_HDMI 1
-+# define PV_CONTROL_CLK_SELECT_VEC 2
- # define PV_CONTROL_CLK_SELECT_MASK VC4_MASK(3, 2)
- # define PV_CONTROL_CLK_SELECT_SHIFT 2
- # define PV_CONTROL_FIFO_CLR BIT(1)
-diff --git a/drivers/irqchip/irq-gic-v3-its.c b/drivers/irqchip/irq-gic-v3-its.c
-index c5dee30..acb9d25 100644
---- a/drivers/irqchip/irq-gic-v3-its.c
-+++ b/drivers/irqchip/irq-gic-v3-its.c
-@@ -1598,6 +1598,14 @@ static void __maybe_unused its_enable_quirk_cavium_23144(void *data)
- its->flags |= ITS_FLAGS_WORKAROUND_CAVIUM_23144;
- }
-
-+static void __maybe_unused its_enable_quirk_qdf2400_e0065(void *data)
-+{
-+ struct its_node *its = data;
-+
-+ /* On QDF2400, the size of the ITE is 16Bytes */
-+ its->ite_size = 16;
-+}
-+
- static const struct gic_quirk its_quirks[] = {
- #ifdef CONFIG_CAVIUM_ERRATUM_22375
- {
-@@ -1615,6 +1623,14 @@ static const struct gic_quirk its_quirks[] = {
- .init = its_enable_quirk_cavium_23144,
- },
- #endif
-+#ifdef CONFIG_QCOM_QDF2400_ERRATUM_0065
-+ {
-+ .desc = "ITS: QDF2400 erratum 0065",
-+ .iidr = 0x00001070, /* QDF2400 ITS rev 1.x */
-+ .mask = 0xffffffff,
-+ .init = its_enable_quirk_qdf2400_e0065,
-+ },
-+#endif
- {
- }
- };
-diff --git a/drivers/media/usb/uvc/uvc_driver.c b/drivers/media/usb/uvc/uvc_driver.c
-index 302e284..cde43b6 100644
---- a/drivers/media/usb/uvc/uvc_driver.c
-+++ b/drivers/media/usb/uvc/uvc_driver.c
-@@ -1595,6 +1595,114 @@ static const char *uvc_print_chain(struct uvc_video_chain *chain)
- return buffer;
- }
-
-+static struct uvc_video_chain *uvc_alloc_chain(struct uvc_device *dev)
-+{
-+ struct uvc_video_chain *chain;
-+
-+ chain = kzalloc(sizeof(*chain), GFP_KERNEL);
-+ if (chain == NULL)
-+ return NULL;
-+
-+ INIT_LIST_HEAD(&chain->entities);
-+ mutex_init(&chain->ctrl_mutex);
-+ chain->dev = dev;
-+ v4l2_prio_init(&chain->prio);
-+
-+ return chain;
-+}
-+
-+/*
-+ * Fallback heuristic for devices that don't connect units and terminals in a
-+ * valid chain.
-+ *
-+ * Some devices have invalid baSourceID references, causing uvc_scan_chain()
-+ * to fail, but if we just take the entities we can find and put them together
-+ * in the most sensible chain we can think of, turns out they do work anyway.
-+ * Note: This heuristic assumes there is a single chain.
-+ *
-+ * At the time of writing, devices known to have such a broken chain are
-+ * - Acer Integrated Camera (5986:055a)
-+ * - Realtek rtl157a7 (0bda:57a7)
-+ */
-+static int uvc_scan_fallback(struct uvc_device *dev)
-+{
-+ struct uvc_video_chain *chain;
-+ struct uvc_entity *iterm = NULL;
-+ struct uvc_entity *oterm = NULL;
-+ struct uvc_entity *entity;
-+ struct uvc_entity *prev;
-+
-+ /*
-+ * Start by locating the input and output terminals. We only support
-+ * devices with exactly one of each for now.
-+ */
-+ list_for_each_entry(entity, &dev->entities, list) {
-+ if (UVC_ENTITY_IS_ITERM(entity)) {
-+ if (iterm)
-+ return -EINVAL;
-+ iterm = entity;
-+ }
-+
-+ if (UVC_ENTITY_IS_OTERM(entity)) {
-+ if (oterm)
-+ return -EINVAL;
-+ oterm = entity;
-+ }
-+ }
-+
-+ if (iterm == NULL || oterm == NULL)
-+ return -EINVAL;
-+
-+ /* Allocate the chain and fill it. */
-+ chain = uvc_alloc_chain(dev);
-+ if (chain == NULL)
-+ return -ENOMEM;
-+
-+ if (uvc_scan_chain_entity(chain, oterm) < 0)
-+ goto error;
-+
-+ prev = oterm;
-+
-+ /*
-+ * Add all Processing and Extension Units with two pads. The order
-+ * doesn't matter much, use reverse list traversal to connect units in
-+ * UVC descriptor order as we build the chain from output to input. This
-+ * leads to units appearing in the order meant by the manufacturer for
-+ * the cameras known to require this heuristic.
-+ */
-+ list_for_each_entry_reverse(entity, &dev->entities, list) {
-+ if (entity->type != UVC_VC_PROCESSING_UNIT &&
-+ entity->type != UVC_VC_EXTENSION_UNIT)
-+ continue;
-+
-+ if (entity->num_pads != 2)
-+ continue;
-+
-+ if (uvc_scan_chain_entity(chain, entity) < 0)
-+ goto error;
-+
-+ prev->baSourceID[0] = entity->id;
-+ prev = entity;
-+ }
-+
-+ if (uvc_scan_chain_entity(chain, iterm) < 0)
-+ goto error;
-+
-+ prev->baSourceID[0] = iterm->id;
-+
-+ list_add_tail(&chain->list, &dev->chains);
-+
-+ uvc_trace(UVC_TRACE_PROBE,
-+ "Found a video chain by fallback heuristic (%s).\n",
-+ uvc_print_chain(chain));
-+
-+ return 0;
-+
-+error:
-+ kfree(chain);
-+ return -EINVAL;
-+}
-+
- /*
- * Scan the device for video chains and register video devices.
- *
-@@ -1617,15 +1725,10 @@ static int uvc_scan_device(struct uvc_device *dev)
- if (term->chain.next || term->chain.prev)
- continue;
-
-- chain = kzalloc(sizeof(*chain), GFP_KERNEL);
-+ chain = uvc_alloc_chain(dev);
- if (chain == NULL)
- return -ENOMEM;
-
-- INIT_LIST_HEAD(&chain->entities);
-- mutex_init(&chain->ctrl_mutex);
-- chain->dev = dev;
-- v4l2_prio_init(&chain->prio);
--
- term->flags |= UVC_ENTITY_FLAG_DEFAULT;
-
- if (uvc_scan_chain(chain, term) < 0) {
-@@ -1639,6 +1742,9 @@ static int uvc_scan_device(struct uvc_device *dev)
- list_add_tail(&chain->list, &dev->chains);
- }
-
-+ if (list_empty(&dev->chains))
-+ uvc_scan_fallback(dev);
-+
- if (list_empty(&dev->chains)) {
- uvc_printk(KERN_INFO, "No valid video chain found.\n");
- return -1;
-diff --git a/drivers/net/ethernet/ibm/ibmveth.c b/drivers/net/ethernet/ibm/ibmveth.c
-index a36022b..03dca73 100644
---- a/drivers/net/ethernet/ibm/ibmveth.c
-+++ b/drivers/net/ethernet/ibm/ibmveth.c
-@@ -1181,7 +1181,9 @@ static netdev_tx_t ibmveth_start_xmit(struct sk_buff *skb,
-
- static void ibmveth_rx_mss_helper(struct sk_buff *skb, u16 mss, int lrg_pkt)
- {
-+ struct tcphdr *tcph;
- int offset = 0;
-+ int hdr_len;
-
- /* only TCP packets will be aggregated */
- if (skb->protocol == htons(ETH_P_IP)) {
-@@ -1208,14 +1210,20 @@ static void ibmveth_rx_mss_helper(struct sk_buff *skb, u16 mss, int lrg_pkt)
- /* if mss is not set through Large Packet bit/mss in rx buffer,
- * expect that the mss will be written to the tcp header checksum.
- */
-+ tcph = (struct tcphdr *)(skb->data + offset);
- if (lrg_pkt) {
- skb_shinfo(skb)->gso_size = mss;
- } else if (offset) {
-- struct tcphdr *tcph = (struct tcphdr *)(skb->data + offset);
--
- skb_shinfo(skb)->gso_size = ntohs(tcph->check);
- tcph->check = 0;
- }
-+
-+ if (skb_shinfo(skb)->gso_size) {
-+ hdr_len = offset + tcph->doff * 4;
-+ skb_shinfo(skb)->gso_segs =
-+ DIV_ROUND_UP(skb->len - hdr_len,
-+ skb_shinfo(skb)->gso_size);
-+ }
- }
-
- static int ibmveth_poll(struct napi_struct *napi, int budget)
-diff --git a/drivers/net/ethernet/intel/igb/e1000_phy.c b/drivers/net/ethernet/intel/igb/e1000_phy.c
-index 5b54254..2788a54 100644
---- a/drivers/net/ethernet/intel/igb/e1000_phy.c
-+++ b/drivers/net/ethernet/intel/igb/e1000_phy.c
-@@ -77,6 +77,10 @@ s32 igb_get_phy_id(struct e1000_hw *hw)
- s32 ret_val = 0;
- u16 phy_id;
-
-+ /* ensure PHY page selection to fix misconfigured i210 */
-+ if ((hw->mac.type == e1000_i210) || (hw->mac.type == e1000_i211))
-+ phy->ops.write_reg(hw, I347AT4_PAGE_SELECT, 0);
-+
- ret_val = phy->ops.read_reg(hw, PHY_ID1, &phy_id);
- if (ret_val)
- goto out;
-diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_main.c b/drivers/net/ethernet/mellanox/mlx5/core/en_main.c
-index b3067137..d4fa851 100644
---- a/drivers/net/ethernet/mellanox/mlx5/core/en_main.c
-+++ b/drivers/net/ethernet/mellanox/mlx5/core/en_main.c
-@@ -81,6 +81,7 @@ static bool mlx5e_check_fragmented_striding_rq_cap(struct mlx5_core_dev *mdev)
- static void mlx5e_set_rq_type_params(struct mlx5e_priv *priv, u8 rq_type)
- {
- priv->params.rq_wq_type = rq_type;
-+ priv->params.lro_wqe_sz = MLX5E_PARAMS_DEFAULT_LRO_WQE_SZ;
- switch (priv->params.rq_wq_type) {
- case MLX5_WQ_TYPE_LINKED_LIST_STRIDING_RQ:
- priv->params.log_rq_size = MLX5E_PARAMS_DEFAULT_LOG_RQ_SIZE_MPW;
-@@ -92,6 +93,10 @@ static void mlx5e_set_rq_type_params(struct mlx5e_priv *priv, u8 rq_type)
- break;
- default: /* MLX5_WQ_TYPE_LINKED_LIST */
- priv->params.log_rq_size = MLX5E_PARAMS_DEFAULT_LOG_RQ_SIZE;
-+
-+ /* Extra room needed for build_skb */
-+ priv->params.lro_wqe_sz -= MLX5_RX_HEADROOM +
-+ SKB_DATA_ALIGN(sizeof(struct skb_shared_info));
- }
- priv->params.min_rx_wqes = mlx5_min_rx_wqes(priv->params.rq_wq_type,
- BIT(priv->params.log_rq_size));
-@@ -3473,12 +3478,6 @@ static void mlx5e_build_nic_netdev_priv(struct mlx5_core_dev *mdev,
- mlx5e_build_default_indir_rqt(mdev, priv->params.indirection_rqt,
- MLX5E_INDIR_RQT_SIZE, profile->max_nch(mdev));
-
-- priv->params.lro_wqe_sz =
-- MLX5E_PARAMS_DEFAULT_LRO_WQE_SZ -
-- /* Extra room needed for build_skb */
-- MLX5_RX_HEADROOM -
-- SKB_DATA_ALIGN(sizeof(struct skb_shared_info));
--
- /* Initialize pflags */
- MLX5E_SET_PRIV_FLAG(priv, MLX5E_PFLAG_RX_CQE_BASED_MODER,
- priv->params.rx_cq_period_mode == MLX5_CQ_PERIOD_MODE_START_FROM_CQE);
-@@ -3936,6 +3935,19 @@ static void mlx5e_register_vport_rep(struct mlx5_core_dev *mdev)
- }
- }
-
-+static void mlx5e_unregister_vport_rep(struct mlx5_core_dev *mdev)
-+{
-+ struct mlx5_eswitch *esw = mdev->priv.eswitch;
-+ int total_vfs = MLX5_TOTAL_VPORTS(mdev);
-+ int vport;
-+
-+ if (!MLX5_CAP_GEN(mdev, vport_group_manager))
-+ return;
-+
-+ for (vport = 1; vport < total_vfs; vport++)
-+ mlx5_eswitch_unregister_vport_rep(esw, vport);
-+}
-+
- void mlx5e_detach_netdev(struct mlx5_core_dev *mdev, struct net_device *netdev)
- {
- struct mlx5e_priv *priv = netdev_priv(netdev);
-@@ -3983,6 +3995,7 @@ static int mlx5e_attach(struct mlx5_core_dev *mdev, void *vpriv)
- return err;
- }
-
-+ mlx5e_register_vport_rep(mdev);
- return 0;
- }
-
-@@ -3994,6 +4007,7 @@ static void mlx5e_detach(struct mlx5_core_dev *mdev, void *vpriv)
- if (!netif_device_present(netdev))
- return;
-
-+ mlx5e_unregister_vport_rep(mdev);
- mlx5e_detach_netdev(mdev, netdev);
- mlx5e_destroy_mdev_resources(mdev);
- }
-@@ -4012,8 +4026,6 @@ static void *mlx5e_add(struct mlx5_core_dev *mdev)
- if (err)
- return NULL;
-
-- mlx5e_register_vport_rep(mdev);
--
- if (MLX5_CAP_GEN(mdev, vport_group_manager))
- ppriv = &esw->offloads.vport_reps[0];
-
-@@ -4065,13 +4077,7 @@ void mlx5e_destroy_netdev(struct mlx5_core_dev *mdev, struct mlx5e_priv *priv)
-
- static void mlx5e_remove(struct mlx5_core_dev *mdev, void *vpriv)
- {
-- struct mlx5_eswitch *esw = mdev->priv.eswitch;
-- int total_vfs = MLX5_TOTAL_VPORTS(mdev);
- struct mlx5e_priv *priv = vpriv;
-- int vport;
--
-- for (vport = 1; vport < total_vfs; vport++)
-- mlx5_eswitch_unregister_vport_rep(esw, vport);
-
- unregister_netdev(priv->netdev);
- mlx5e_detach(mdev, vpriv);
-diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c b/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
-index e7b2158..796bdf0 100644
---- a/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
-+++ b/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
-@@ -92,19 +92,18 @@ static inline void mlx5e_cqes_update_owner(struct mlx5e_cq *cq, u32 cqcc, int n)
- static inline void mlx5e_decompress_cqe(struct mlx5e_rq *rq,
- struct mlx5e_cq *cq, u32 cqcc)
- {
-- u16 wqe_cnt_step;
--
- cq->title.byte_cnt = cq->mini_arr[cq->mini_arr_idx].byte_cnt;
- cq->title.check_sum = cq->mini_arr[cq->mini_arr_idx].checksum;
- cq->title.op_own &= 0xf0;
- cq->title.op_own |= 0x01 & (cqcc >> cq->wq.log_sz);
- cq->title.wqe_counter = cpu_to_be16(cq->decmprs_wqe_counter);
-
-- wqe_cnt_step =
-- rq->wq_type == MLX5_WQ_TYPE_LINKED_LIST_STRIDING_RQ ?
-- mpwrq_get_cqe_consumed_strides(&cq->title) : 1;
-- cq->decmprs_wqe_counter =
-- (cq->decmprs_wqe_counter + wqe_cnt_step) & rq->wq.sz_m1;
-+ if (rq->wq_type == MLX5_WQ_TYPE_LINKED_LIST_STRIDING_RQ)
-+ cq->decmprs_wqe_counter +=
-+ mpwrq_get_cqe_consumed_strides(&cq->title);
-+ else
-+ cq->decmprs_wqe_counter =
-+ (cq->decmprs_wqe_counter + 1) & rq->wq.sz_m1;
- }
-
- static inline void mlx5e_decompress_cqe_no_hash(struct mlx5e_rq *rq,
-diff --git a/drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c b/drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
-index e83072d..6905630 100644
---- a/drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
-+++ b/drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
-@@ -500,30 +500,40 @@ static int
- mlxsw_sp_vr_lpm_tree_check(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_vr *vr,
- struct mlxsw_sp_prefix_usage *req_prefix_usage)
- {
-- struct mlxsw_sp_lpm_tree *lpm_tree;
-+ struct mlxsw_sp_lpm_tree *lpm_tree = vr->lpm_tree;
-+ struct mlxsw_sp_lpm_tree *new_tree;
-+ int err;
-
-- if (mlxsw_sp_prefix_usage_eq(req_prefix_usage,
-- &vr->lpm_tree->prefix_usage))
-+ if (mlxsw_sp_prefix_usage_eq(req_prefix_usage, &lpm_tree->prefix_usage))
- return 0;
-
-- lpm_tree = mlxsw_sp_lpm_tree_get(mlxsw_sp, req_prefix_usage,
-+ new_tree = mlxsw_sp_lpm_tree_get(mlxsw_sp, req_prefix_usage,
- vr->proto, false);
-- if (IS_ERR(lpm_tree)) {
-+ if (IS_ERR(new_tree)) {
- /* We failed to get a tree according to the required
- * prefix usage. However, the current tree might be still good
- * for us if our requirement is subset of the prefixes used
- * in the tree.
- */
- if (mlxsw_sp_prefix_usage_subset(req_prefix_usage,
-- &vr->lpm_tree->prefix_usage))
-+ &lpm_tree->prefix_usage))
- return 0;
-- return PTR_ERR(lpm_tree);
-+ return PTR_ERR(new_tree);
- }
-
-- mlxsw_sp_vr_lpm_tree_unbind(mlxsw_sp, vr);
-- mlxsw_sp_lpm_tree_put(mlxsw_sp, vr->lpm_tree);
-+ /* Prevent packet loss by overwriting existing binding */
-+ vr->lpm_tree = new_tree;
-+ err = mlxsw_sp_vr_lpm_tree_bind(mlxsw_sp, vr);
-+ if (err)
-+ goto err_tree_bind;
-+ mlxsw_sp_lpm_tree_put(mlxsw_sp, lpm_tree);
-+
-+ return 0;
-+
-+err_tree_bind:
- vr->lpm_tree = lpm_tree;
-- return mlxsw_sp_vr_lpm_tree_bind(mlxsw_sp, vr);
-+ mlxsw_sp_lpm_tree_put(mlxsw_sp, new_tree);
-+ return err;
- }
-
- static struct mlxsw_sp_vr *mlxsw_sp_vr_get(struct mlxsw_sp *mlxsw_sp,
-diff --git a/drivers/net/geneve.c b/drivers/net/geneve.c
-index 8b4822a..3c1f89a 100644
---- a/drivers/net/geneve.c
-+++ b/drivers/net/geneve.c
-@@ -1039,16 +1039,22 @@ static netdev_tx_t geneve_xmit(struct sk_buff *skb, struct net_device *dev)
- {
- struct geneve_dev *geneve = netdev_priv(dev);
- struct ip_tunnel_info *info = NULL;
-+ int err;
-
- if (geneve->collect_md)
- info = skb_tunnel_info(skb);
-
-+ rcu_read_lock();
- #if IS_ENABLED(CONFIG_IPV6)
- if ((info && ip_tunnel_info_af(info) == AF_INET6) ||
- (!info && geneve->remote.sa.sa_family == AF_INET6))
-- return geneve6_xmit_skb(skb, dev, info);
-+ err = geneve6_xmit_skb(skb, dev, info);
-+ else
- #endif
-- return geneve_xmit_skb(skb, dev, info);
-+ err = geneve_xmit_skb(skb, dev, info);
-+ rcu_read_unlock();
-+
-+ return err;
- }
-
- static int __geneve_change_mtu(struct net_device *dev, int new_mtu, bool strict)
-diff --git a/drivers/net/phy/phy.c b/drivers/net/phy/phy.c
-index f424b86..201ffa5 100644
---- a/drivers/net/phy/phy.c
-+++ b/drivers/net/phy/phy.c
-@@ -611,14 +611,18 @@ void phy_start_machine(struct phy_device *phydev)
- * phy_trigger_machine - trigger the state machine to run
- *
- * @phydev: the phy_device struct
-+ * @sync: indicate whether we should wait for the workqueue cancelation
- *
- * Description: There has been a change in state which requires that the
- * state machine runs.
- */
-
--static void phy_trigger_machine(struct phy_device *phydev)
-+static void phy_trigger_machine(struct phy_device *phydev, bool sync)
- {
-- cancel_delayed_work_sync(&phydev->state_queue);
-+ if (sync)
-+ cancel_delayed_work_sync(&phydev->state_queue);
-+ else
-+ cancel_delayed_work(&phydev->state_queue);
- queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, 0);
- }
-
-@@ -655,7 +659,7 @@ static void phy_error(struct phy_device *phydev)
- phydev->state = PHY_HALTED;
- mutex_unlock(&phydev->lock);
-
-- phy_trigger_machine(phydev);
-+ phy_trigger_machine(phydev, false);
- }
-
- /**
-@@ -817,7 +821,7 @@ void phy_change(struct work_struct *work)
- }
-
- /* reschedule state queue work to run as soon as possible */
-- phy_trigger_machine(phydev);
-+ phy_trigger_machine(phydev, true);
- return;
-
- ignore:
-@@ -907,7 +911,7 @@ void phy_start(struct phy_device *phydev)
- if (do_resume)
- phy_resume(phydev);
-
-- phy_trigger_machine(phydev);
-+ phy_trigger_machine(phydev, true);
- }
- EXPORT_SYMBOL(phy_start);
-
-diff --git a/drivers/net/tun.c b/drivers/net/tun.c
-index b31aca8..a931b73 100644
---- a/drivers/net/tun.c
-+++ b/drivers/net/tun.c
-@@ -819,7 +819,18 @@ static void tun_net_uninit(struct net_device *dev)
- /* Net device open. */
- static int tun_net_open(struct net_device *dev)
- {
-+ struct tun_struct *tun = netdev_priv(dev);
-+ int i;
-+
- netif_tx_start_all_queues(dev);
-+
-+ for (i = 0; i < tun->numqueues; i++) {
-+ struct tun_file *tfile;
-+
-+ tfile = rtnl_dereference(tun->tfiles[i]);
-+ tfile->socket.sk->sk_write_space(tfile->socket.sk);
-+ }
-+
- return 0;
- }
-
-@@ -1116,9 +1127,10 @@ static unsigned int tun_chr_poll(struct file *file, poll_table *wait)
- if (!skb_array_empty(&tfile->tx_array))
- mask |= POLLIN | POLLRDNORM;
-
-- if (sock_writeable(sk) ||
-- (!test_and_set_bit(SOCKWQ_ASYNC_NOSPACE, &sk->sk_socket->flags) &&
-- sock_writeable(sk)))
-+ if (tun->dev->flags & IFF_UP &&
-+ (sock_writeable(sk) ||
-+ (!test_and_set_bit(SOCKWQ_ASYNC_NOSPACE, &sk->sk_socket->flags) &&
-+ sock_writeable(sk))))
- mask |= POLLOUT | POLLWRNORM;
-
- if (tun->dev->reg_state != NETREG_REGISTERED)
-diff --git a/drivers/net/vrf.c b/drivers/net/vrf.c
-index 95cf1d8..bc744ac 100644
---- a/drivers/net/vrf.c
-+++ b/drivers/net/vrf.c
-@@ -346,6 +346,7 @@ static netdev_tx_t is_ip_tx_frame(struct sk_buff *skb, struct net_device *dev)
-
- static netdev_tx_t vrf_xmit(struct sk_buff *skb, struct net_device *dev)
- {
-+ int len = skb->len;
- netdev_tx_t ret = is_ip_tx_frame(skb, dev);
-
- if (likely(ret == NET_XMIT_SUCCESS || ret == NET_XMIT_CN)) {
-@@ -353,7 +354,7 @@ static netdev_tx_t vrf_xmit(struct sk_buff *skb, struct net_device *dev)
-
- u64_stats_update_begin(&dstats->syncp);
- dstats->tx_pkts++;
-- dstats->tx_bytes += skb->len;
-+ dstats->tx_bytes += len;
- u64_stats_update_end(&dstats->syncp);
- } else {
- this_cpu_inc(dev->dstats->tx_drps);
-diff --git a/drivers/net/vxlan.c b/drivers/net/vxlan.c
-index d4f495b..3c4c2cf 100644
---- a/drivers/net/vxlan.c
-+++ b/drivers/net/vxlan.c
-@@ -1942,7 +1942,6 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
- const struct iphdr *old_iph;
- union vxlan_addr *dst;
- union vxlan_addr remote_ip, local_ip;
-- union vxlan_addr *src;
- struct vxlan_metadata _md;
- struct vxlan_metadata *md = &_md;
- __be16 src_port = 0, dst_port;
-@@ -1956,11 +1955,12 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
-
- info = skb_tunnel_info(skb);
-
-+ rcu_read_lock();
- if (rdst) {
- dst_port = rdst->remote_port ? rdst->remote_port : vxlan->cfg.dst_port;
- vni = rdst->remote_vni;
- dst = &rdst->remote_ip;
-- src = &vxlan->cfg.saddr;
-+ local_ip = vxlan->cfg.saddr;
- dst_cache = &rdst->dst_cache;
- } else {
- if (!info) {
-@@ -1979,7 +1979,6 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
- local_ip.sin6.sin6_addr = info->key.u.ipv6.src;
- }
- dst = &remote_ip;
-- src = &local_ip;
- dst_cache = &info->dst_cache;
- }
-
-@@ -1987,7 +1986,7 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
- if (did_rsc) {
- /* short-circuited back to local bridge */
- vxlan_encap_bypass(skb, vxlan, vxlan);
-- return;
-+ goto out_unlock;
- }
- goto drop;
- }
-@@ -2028,7 +2027,7 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
- rt = vxlan_get_route(vxlan, skb,
- rdst ? rdst->remote_ifindex : 0, tos,
- dst->sin.sin_addr.s_addr,
-- &src->sin.sin_addr.s_addr,
-+ &local_ip.sin.sin_addr.s_addr,
- dst_cache, info);
- if (IS_ERR(rt)) {
- netdev_dbg(dev, "no route to %pI4\n",
-@@ -2056,7 +2055,7 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
- if (!dst_vxlan)
- goto tx_error;
- vxlan_encap_bypass(skb, vxlan, dst_vxlan);
-- return;
-+ goto out_unlock;
- }
-
- if (!info)
-@@ -2071,7 +2070,7 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
- if (err < 0)
- goto xmit_tx_error;
-
-- udp_tunnel_xmit_skb(rt, sk, skb, src->sin.sin_addr.s_addr,
-+ udp_tunnel_xmit_skb(rt, sk, skb, local_ip.sin.sin_addr.s_addr,
- dst->sin.sin_addr.s_addr, tos, ttl, df,
- src_port, dst_port, xnet, !udp_sum);
- #if IS_ENABLED(CONFIG_IPV6)
-@@ -2087,7 +2086,7 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
- ndst = vxlan6_get_route(vxlan, skb,
- rdst ? rdst->remote_ifindex : 0, tos,
- label, &dst->sin6.sin6_addr,
-- &src->sin6.sin6_addr,
-+ &local_ip.sin6.sin6_addr,
- dst_cache, info);
- if (IS_ERR(ndst)) {
- netdev_dbg(dev, "no route to %pI6\n",
-@@ -2117,7 +2116,7 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
- if (!dst_vxlan)
- goto tx_error;
- vxlan_encap_bypass(skb, vxlan, dst_vxlan);
-- return;
-+ goto out_unlock;
- }
-
- if (!info)
-@@ -2131,15 +2130,16 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
- if (err < 0) {
- dst_release(ndst);
- dev->stats.tx_errors++;
-- return;
-+ goto out_unlock;
- }
- udp_tunnel6_xmit_skb(ndst, sk, skb, dev,
-- &src->sin6.sin6_addr,
-+ &local_ip.sin6.sin6_addr,
- &dst->sin6.sin6_addr, tos, ttl,
- label, src_port, dst_port, !udp_sum);
- #endif
- }
--
-+out_unlock:
-+ rcu_read_unlock();
- return;
-
- drop:
-@@ -2155,6 +2155,7 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev,
- dev->stats.tx_errors++;
- tx_free:
- dev_kfree_skb(skb);
-+ rcu_read_unlock();
- }
-
- /* Transmit local packets over Vxlan
-@@ -2637,7 +2638,7 @@ static int vxlan_validate(struct nlattr *tb[], struct nlattr *data[])
-
- if (data[IFLA_VXLAN_ID]) {
- __u32 id = nla_get_u32(data[IFLA_VXLAN_ID]);
-- if (id >= VXLAN_VID_MASK)
-+ if (id >= VXLAN_N_VID)
- return -ERANGE;
- }
-
-diff --git a/drivers/pci/iov.c b/drivers/pci/iov.c
-index e30f05c..4722782 100644
---- a/drivers/pci/iov.c
-+++ b/drivers/pci/iov.c
-@@ -306,13 +306,6 @@ static int sriov_enable(struct pci_dev *dev, int nr_virtfn)
- return rc;
- }
-
-- pci_iov_set_numvfs(dev, nr_virtfn);
-- iov->ctrl |= PCI_SRIOV_CTRL_VFE | PCI_SRIOV_CTRL_MSE;
-- pci_cfg_access_lock(dev);
-- pci_write_config_word(dev, iov->pos + PCI_SRIOV_CTRL, iov->ctrl);
-- msleep(100);
-- pci_cfg_access_unlock(dev);
--
- iov->initial_VFs = initial;
- if (nr_virtfn < initial)
- initial = nr_virtfn;
-@@ -323,6 +316,13 @@ static int sriov_enable(struct pci_dev *dev, int nr_virtfn)
- goto err_pcibios;
- }
-
-+ pci_iov_set_numvfs(dev, nr_virtfn);
-+ iov->ctrl |= PCI_SRIOV_CTRL_VFE | PCI_SRIOV_CTRL_MSE;
-+ pci_cfg_access_lock(dev);
-+ pci_write_config_word(dev, iov->pos + PCI_SRIOV_CTRL, iov->ctrl);
-+ msleep(100);
-+ pci_cfg_access_unlock(dev);
-+
- for (i = 0; i < initial; i++) {
- rc = pci_iov_add_virtfn(dev, i, 0);
- if (rc)
-@@ -554,21 +554,61 @@ void pci_iov_release(struct pci_dev *dev)
- }
-
- /**
-- * pci_iov_resource_bar - get position of the SR-IOV BAR
-+ * pci_iov_update_resource - update a VF BAR
- * @dev: the PCI device
- * @resno: the resource number
- *
-- * Returns position of the BAR encapsulated in the SR-IOV capability.
-+ * Update a VF BAR in the SR-IOV capability of a PF.
- */
--int pci_iov_resource_bar(struct pci_dev *dev, int resno)
-+void pci_iov_update_resource(struct pci_dev *dev, int resno)
- {
-- if (resno < PCI_IOV_RESOURCES || resno > PCI_IOV_RESOURCE_END)
-- return 0;
-+ struct pci_sriov *iov = dev->is_physfn ? dev->sriov : NULL;
-+ struct resource *res = dev->resource + resno;
-+ int vf_bar = resno - PCI_IOV_RESOURCES;
-+ struct pci_bus_region region;
-+ u16 cmd;
-+ u32 new;
-+ int reg;
-+
-+ /*
-+ * The generic pci_restore_bars() path calls this for all devices,
-+ * including VFs and non-SR-IOV devices. If this is not a PF, we
-+ * have nothing to do.
-+ */
-+ if (!iov)
-+ return;
-+
-+ pci_read_config_word(dev, iov->pos + PCI_SRIOV_CTRL, &cmd);
-+ if ((cmd & PCI_SRIOV_CTRL_VFE) && (cmd & PCI_SRIOV_CTRL_MSE)) {
-+ dev_WARN(&dev->dev, "can't update enabled VF BAR%d %pR\n",
-+ vf_bar, res);
-+ return;
-+ }
-+
-+ /*
-+ * Ignore unimplemented BARs, unused resource slots for 64-bit
-+ * BARs, and non-movable resources, e.g., those described via
-+ * Enhanced Allocation.
-+ */
-+ if (!res->flags)
-+ return;
-+
-+ if (res->flags & IORESOURCE_UNSET)
-+ return;
-+
-+ if (res->flags & IORESOURCE_PCI_FIXED)
-+ return;
-
-- BUG_ON(!dev->is_physfn);
-+ pcibios_resource_to_bus(dev->bus, &region, res);
-+ new = region.start;
-+ new |= res->flags & ~PCI_BASE_ADDRESS_MEM_MASK;
-
-- return dev->sriov->pos + PCI_SRIOV_BAR +
-- 4 * (resno - PCI_IOV_RESOURCES);
-+ reg = iov->pos + PCI_SRIOV_BAR + 4 * vf_bar;
-+ pci_write_config_dword(dev, reg, new);
-+ if (res->flags & IORESOURCE_MEM_64) {
-+ new = region.start >> 16 >> 16;
-+ pci_write_config_dword(dev, reg + 4, new);
-+ }
- }
-
- resource_size_t __weak pcibios_iov_resource_alignment(struct pci_dev *dev,
-diff --git a/drivers/pci/pci.c b/drivers/pci/pci.c
-index eda6a7c..6922964 100644
---- a/drivers/pci/pci.c
-+++ b/drivers/pci/pci.c
-@@ -564,10 +564,6 @@ static void pci_restore_bars(struct pci_dev *dev)
- {
- int i;
-
-- /* Per SR-IOV spec 3.4.1.11, VF BARs are RO zero */
-- if (dev->is_virtfn)
-- return;
--
- for (i = 0; i < PCI_BRIDGE_RESOURCES; i++)
- pci_update_resource(dev, i);
- }
-@@ -4835,36 +4831,6 @@ int pci_select_bars(struct pci_dev *dev, unsigned long flags)
- }
- EXPORT_SYMBOL(pci_select_bars);
-
--/**
-- * pci_resource_bar - get position of the BAR associated with a resource
-- * @dev: the PCI device
-- * @resno: the resource number
-- * @type: the BAR type to be filled in
-- *
-- * Returns BAR position in config space, or 0 if the BAR is invalid.
-- */
--int pci_resource_bar(struct pci_dev *dev, int resno, enum pci_bar_type *type)
--{
-- int reg;
--
-- if (resno < PCI_ROM_RESOURCE) {
-- *type = pci_bar_unknown;
-- return PCI_BASE_ADDRESS_0 + 4 * resno;
-- } else if (resno == PCI_ROM_RESOURCE) {
-- *type = pci_bar_mem32;
-- return dev->rom_base_reg;
-- } else if (resno < PCI_BRIDGE_RESOURCES) {
-- /* device specific resource */
-- *type = pci_bar_unknown;
-- reg = pci_iov_resource_bar(dev, resno);
-- if (reg)
-- return reg;
-- }
--
-- dev_err(&dev->dev, "BAR %d: invalid resource\n", resno);
-- return 0;
--}
--
- /* Some architectures require additional programming to enable VGA */
- static arch_set_vga_state_t arch_set_vga_state;
-
-diff --git a/drivers/pci/pci.h b/drivers/pci/pci.h
-index 4518562..a5d37f6 100644
---- a/drivers/pci/pci.h
-+++ b/drivers/pci/pci.h
-@@ -245,7 +245,6 @@ bool pci_bus_read_dev_vendor_id(struct pci_bus *bus, int devfn, u32 *pl,
- int pci_setup_device(struct pci_dev *dev);
- int __pci_read_base(struct pci_dev *dev, enum pci_bar_type type,
- struct resource *res, unsigned int reg);
--int pci_resource_bar(struct pci_dev *dev, int resno, enum pci_bar_type *type);
- void pci_configure_ari(struct pci_dev *dev);
- void __pci_bus_size_bridges(struct pci_bus *bus,
- struct list_head *realloc_head);
-@@ -289,7 +288,7 @@ static inline void pci_restore_ats_state(struct pci_dev *dev)
- #ifdef CONFIG_PCI_IOV
- int pci_iov_init(struct pci_dev *dev);
- void pci_iov_release(struct pci_dev *dev);
--int pci_iov_resource_bar(struct pci_dev *dev, int resno);
-+void pci_iov_update_resource(struct pci_dev *dev, int resno);
- resource_size_t pci_sriov_resource_alignment(struct pci_dev *dev, int resno);
- void pci_restore_iov_state(struct pci_dev *dev);
- int pci_iov_bus_range(struct pci_bus *bus);
-@@ -303,10 +302,6 @@ static inline void pci_iov_release(struct pci_dev *dev)
-
- {
- }
--static inline int pci_iov_resource_bar(struct pci_dev *dev, int resno)
--{
-- return 0;
--}
- static inline void pci_restore_iov_state(struct pci_dev *dev)
- {
- }
-diff --git a/drivers/pci/probe.c b/drivers/pci/probe.c
-index 300770c..d266d80 100644
---- a/drivers/pci/probe.c
-+++ b/drivers/pci/probe.c
-@@ -227,7 +227,8 @@ int __pci_read_base(struct pci_dev *dev, enum pci_bar_type type,
- mask64 = (u32)PCI_BASE_ADDRESS_MEM_MASK;
- }
- } else {
-- res->flags |= (l & IORESOURCE_ROM_ENABLE);
-+ if (l & PCI_ROM_ADDRESS_ENABLE)
-+ res->flags |= IORESOURCE_ROM_ENABLE;
- l64 = l & PCI_ROM_ADDRESS_MASK;
- sz64 = sz & PCI_ROM_ADDRESS_MASK;
- mask64 = (u32)PCI_ROM_ADDRESS_MASK;
-diff --git a/drivers/pci/rom.c b/drivers/pci/rom.c
-index 06663d3..b6edb18 100644
---- a/drivers/pci/rom.c
-+++ b/drivers/pci/rom.c
-@@ -35,6 +35,11 @@ int pci_enable_rom(struct pci_dev *pdev)
- if (res->flags & IORESOURCE_ROM_SHADOW)
- return 0;
-
-+ /*
-+ * Ideally pci_update_resource() would update the ROM BAR address,
-+ * and we would only set the enable bit here. But apparently some
-+ * devices have buggy ROM BARs that read as zero when disabled.
-+ */
- pcibios_resource_to_bus(pdev->bus, &region, res);
- pci_read_config_dword(pdev, pdev->rom_base_reg, &rom_addr);
- rom_addr &= ~PCI_ROM_ADDRESS_MASK;
-diff --git a/drivers/pci/setup-res.c b/drivers/pci/setup-res.c
-index 9526e34..4bc589e 100644
---- a/drivers/pci/setup-res.c
-+++ b/drivers/pci/setup-res.c
-@@ -25,21 +25,18 @@
- #include <linux/slab.h>
- #include "pci.h"
-
--
--void pci_update_resource(struct pci_dev *dev, int resno)
-+static void pci_std_update_resource(struct pci_dev *dev, int resno)
- {
- struct pci_bus_region region;
- bool disable;
- u16 cmd;
- u32 new, check, mask;
- int reg;
-- enum pci_bar_type type;
- struct resource *res = dev->resource + resno;
-
-- if (dev->is_virtfn) {
-- dev_warn(&dev->dev, "can't update VF BAR%d\n", resno);
-+ /* Per SR-IOV spec 3.4.1.11, VF BARs are RO zero */
-+ if (dev->is_virtfn)
- return;
-- }
-
- /*
- * Ignore resources for unimplemented BARs and unused resource slots
-@@ -60,21 +57,34 @@ void pci_update_resource(struct pci_dev *dev, int resno)
- return;
-
- pcibios_resource_to_bus(dev->bus, &region, res);
-+ new = region.start;
-
-- new = region.start | (res->flags & PCI_REGION_FLAG_MASK);
-- if (res->flags & IORESOURCE_IO)
-+ if (res->flags & IORESOURCE_IO) {
- mask = (u32)PCI_BASE_ADDRESS_IO_MASK;
-- else
-+ new |= res->flags & ~PCI_BASE_ADDRESS_IO_MASK;
-+ } else if (resno == PCI_ROM_RESOURCE) {
-+ mask = (u32)PCI_ROM_ADDRESS_MASK;
-+ } else {
- mask = (u32)PCI_BASE_ADDRESS_MEM_MASK;
-+ new |= res->flags & ~PCI_BASE_ADDRESS_MEM_MASK;
-+ }
-
-- reg = pci_resource_bar(dev, resno, &type);
-- if (!reg)
-- return;
-- if (type != pci_bar_unknown) {
-+ if (resno < PCI_ROM_RESOURCE) {
-+ reg = PCI_BASE_ADDRESS_0 + 4 * resno;
-+ } else if (resno == PCI_ROM_RESOURCE) {
-+
-+ /*
-+ * Apparently some Matrox devices have ROM BARs that read
-+ * as zero when disabled, so don't update ROM BARs unless
-+ * they're enabled. See https://lkml.org/lkml/2005/8/30/138.
-+ */
- if (!(res->flags & IORESOURCE_ROM_ENABLE))
- return;
-+
-+ reg = dev->rom_base_reg;
- new |= PCI_ROM_ADDRESS_ENABLE;
-- }
-+ } else
-+ return;
-
- /*
- * We can't update a 64-bit BAR atomically, so when possible,
-@@ -110,6 +120,16 @@ void pci_update_resource(struct pci_dev *dev, int resno)
- pci_write_config_word(dev, PCI_COMMAND, cmd);
- }
-
-+void pci_update_resource(struct pci_dev *dev, int resno)
-+{
-+ if (resno <= PCI_ROM_RESOURCE)
-+ pci_std_update_resource(dev, resno);
-+#ifdef CONFIG_PCI_IOV
-+ else if (resno >= PCI_IOV_RESOURCES && resno <= PCI_IOV_RESOURCE_END)
-+ pci_iov_update_resource(dev, resno);
-+#endif
-+}
-+
- int pci_claim_resource(struct pci_dev *dev, int resource)
- {
- struct resource *res = &dev->resource[resource];
-diff --git a/drivers/s390/crypto/ap_bus.c b/drivers/s390/crypto/ap_bus.c
-index ed92fb0..76b802c 100644
---- a/drivers/s390/crypto/ap_bus.c
-+++ b/drivers/s390/crypto/ap_bus.c
-@@ -1712,6 +1712,9 @@ static void ap_scan_bus(struct work_struct *unused)
- ap_dev->queue_depth = queue_depth;
- ap_dev->raw_hwtype = device_type;
- ap_dev->device_type = device_type;
-+ /* CEX6 toleration: map to CEX5 */
-+ if (device_type == AP_DEVICE_TYPE_CEX6)
-+ ap_dev->device_type = AP_DEVICE_TYPE_CEX5;
- ap_dev->functions = device_functions;
- spin_lock_init(&ap_dev->lock);
- INIT_LIST_HEAD(&ap_dev->pendingq);
-diff --git a/drivers/s390/crypto/ap_bus.h b/drivers/s390/crypto/ap_bus.h
-index d7fdf5c..fd66d2c 100644
---- a/drivers/s390/crypto/ap_bus.h
-+++ b/drivers/s390/crypto/ap_bus.h
-@@ -105,6 +105,7 @@ static inline int ap_test_bit(unsigned int *ptr, unsigned int nr)
- #define AP_DEVICE_TYPE_CEX3C 9
- #define AP_DEVICE_TYPE_CEX4 10
- #define AP_DEVICE_TYPE_CEX5 11
-+#define AP_DEVICE_TYPE_CEX6 12
-
- /*
- * Known function facilities
-diff --git a/drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c b/drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c
-index 91dfd58..c4fe95a 100644
---- a/drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c
-+++ b/drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c
-@@ -22,7 +22,7 @@
- *
- ****************************************************************************/
-
--#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
-+#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
-
- #include <linux/module.h>
- #include <linux/kernel.h>
-@@ -82,7 +82,7 @@ static void ibmvscsis_determine_resid(struct se_cmd *se_cmd,
- }
- } else if (se_cmd->se_cmd_flags & SCF_OVERFLOW_BIT) {
- if (se_cmd->data_direction == DMA_TO_DEVICE) {
-- /* residual data from an overflow write */
-+ /* residual data from an overflow write */
- rsp->flags = SRP_RSP_FLAG_DOOVER;
- rsp->data_out_res_cnt = cpu_to_be32(residual_count);
- } else if (se_cmd->data_direction == DMA_FROM_DEVICE) {
-@@ -102,7 +102,7 @@ static void ibmvscsis_determine_resid(struct se_cmd *se_cmd,
- * and the function returns TRUE.
- *
- * EXECUTION ENVIRONMENT:
-- * Interrupt or Process environment
-+ * Interrupt or Process environment
- */
- static bool connection_broken(struct scsi_info *vscsi)
- {
-@@ -325,7 +325,7 @@ static struct viosrp_crq *ibmvscsis_cmd_q_dequeue(uint mask,
- }
-
- /**
-- * ibmvscsis_send_init_message() - send initialize message to the client
-+ * ibmvscsis_send_init_message() - send initialize message to the client
- * @vscsi: Pointer to our adapter structure
- * @format: Which Init Message format to send
- *
-@@ -383,13 +383,13 @@ static long ibmvscsis_check_init_msg(struct scsi_info *vscsi, uint *format)
- vscsi->cmd_q.base_addr);
- if (crq) {
- *format = (uint)(crq->format);
-- rc = ERROR;
-+ rc = ERROR;
- crq->valid = INVALIDATE_CMD_RESP_EL;
- dma_rmb();
- }
- } else {
- *format = (uint)(crq->format);
-- rc = ERROR;
-+ rc = ERROR;
- crq->valid = INVALIDATE_CMD_RESP_EL;
- dma_rmb();
- }
-@@ -398,166 +398,6 @@ static long ibmvscsis_check_init_msg(struct scsi_info *vscsi, uint *format)
- }
-
- /**
-- * ibmvscsis_establish_new_q() - Establish new CRQ queue
-- * @vscsi: Pointer to our adapter structure
-- * @new_state: New state being established after resetting the queue
-- *
-- * Must be called with interrupt lock held.
-- */
--static long ibmvscsis_establish_new_q(struct scsi_info *vscsi, uint new_state)
--{
-- long rc = ADAPT_SUCCESS;
-- uint format;
--
-- vscsi->flags &= PRESERVE_FLAG_FIELDS;
-- vscsi->rsp_q_timer.timer_pops = 0;
-- vscsi->debit = 0;
-- vscsi->credit = 0;
--
-- rc = vio_enable_interrupts(vscsi->dma_dev);
-- if (rc) {
-- pr_warn("reset_queue: failed to enable interrupts, rc %ld\n",
-- rc);
-- return rc;
-- }
--
-- rc = ibmvscsis_check_init_msg(vscsi, &format);
-- if (rc) {
-- dev_err(&vscsi->dev, "reset_queue: check_init_msg failed, rc %ld\n",
-- rc);
-- return rc;
-- }
--
-- if (format == UNUSED_FORMAT && new_state == WAIT_CONNECTION) {
-- rc = ibmvscsis_send_init_message(vscsi, INIT_MSG);
-- switch (rc) {
-- case H_SUCCESS:
-- case H_DROPPED:
-- case H_CLOSED:
-- rc = ADAPT_SUCCESS;
-- break;
--
-- case H_PARAMETER:
-- case H_HARDWARE:
-- break;
--
-- default:
-- vscsi->state = UNDEFINED;
-- rc = H_HARDWARE;
-- break;
-- }
-- }
--
-- return rc;
--}
--
--/**
-- * ibmvscsis_reset_queue() - Reset CRQ Queue
-- * @vscsi: Pointer to our adapter structure
-- * @new_state: New state to establish after resetting the queue
-- *
-- * This function calls h_free_q and then calls h_reg_q and does all
-- * of the bookkeeping to get us back to where we can communicate.
-- *
-- * Actually, we don't always call h_free_crq. A problem was discovered
-- * where one partition would close and reopen his queue, which would
-- * cause his partner to get a transport event, which would cause him to
-- * close and reopen his queue, which would cause the original partition
-- * to get a transport event, etc., etc. To prevent this, we don't
-- * actually close our queue if the client initiated the reset, (i.e.
-- * either we got a transport event or we have detected that the client's
-- * queue is gone)
-- *
-- * EXECUTION ENVIRONMENT:
-- * Process environment, called with interrupt lock held
-- */
--static void ibmvscsis_reset_queue(struct scsi_info *vscsi, uint new_state)
--{
-- int bytes;
-- long rc = ADAPT_SUCCESS;
--
-- pr_debug("reset_queue: flags 0x%x\n", vscsi->flags);
--
-- /* don't reset, the client did it for us */
-- if (vscsi->flags & (CLIENT_FAILED | TRANS_EVENT)) {
-- vscsi->flags &= PRESERVE_FLAG_FIELDS;
-- vscsi->rsp_q_timer.timer_pops = 0;
-- vscsi->debit = 0;
-- vscsi->credit = 0;
-- vscsi->state = new_state;
-- vio_enable_interrupts(vscsi->dma_dev);
-- } else {
-- rc = ibmvscsis_free_command_q(vscsi);
-- if (rc == ADAPT_SUCCESS) {
-- vscsi->state = new_state;
--
-- bytes = vscsi->cmd_q.size * PAGE_SIZE;
-- rc = h_reg_crq(vscsi->dds.unit_id,
-- vscsi->cmd_q.crq_token, bytes);
-- if (rc == H_CLOSED || rc == H_SUCCESS) {
-- rc = ibmvscsis_establish_new_q(vscsi,
-- new_state);
-- }
--
-- if (rc != ADAPT_SUCCESS) {
-- pr_debug("reset_queue: reg_crq rc %ld\n", rc);
--
-- vscsi->state = ERR_DISCONNECTED;
-- vscsi->flags |= RESPONSE_Q_DOWN;
-- ibmvscsis_free_command_q(vscsi);
-- }
-- } else {
-- vscsi->state = ERR_DISCONNECTED;
-- vscsi->flags |= RESPONSE_Q_DOWN;
-- }
-- }
--}
--
--/**
-- * ibmvscsis_free_cmd_resources() - Free command resources
-- * @vscsi: Pointer to our adapter structure
-- * @cmd: Command which is not longer in use
-- *
-- * Must be called with interrupt lock held.
-- */
--static void ibmvscsis_free_cmd_resources(struct scsi_info *vscsi,
-- struct ibmvscsis_cmd *cmd)
--{
-- struct iu_entry *iue = cmd->iue;
--
-- switch (cmd->type) {
-- case TASK_MANAGEMENT:
-- case SCSI_CDB:
-- /*
-- * When the queue goes down this value is cleared, so it
-- * cannot be cleared in this general purpose function.
-- */
-- if (vscsi->debit)
-- vscsi->debit -= 1;
-- break;
-- case ADAPTER_MAD:
-- vscsi->flags &= ~PROCESSING_MAD;
-- break;
-- case UNSET_TYPE:
-- break;
-- default:
-- dev_err(&vscsi->dev, "free_cmd_resources unknown type %d\n",
-- cmd->type);
-- break;
-- }
--
-- cmd->iue = NULL;
-- list_add_tail(&cmd->list, &vscsi->free_cmd);
-- srp_iu_put(iue);
--
-- if (list_empty(&vscsi->active_q) && list_empty(&vscsi->schedule_q) &&
-- list_empty(&vscsi->waiting_rsp) && (vscsi->flags & WAIT_FOR_IDLE)) {
-- vscsi->flags &= ~WAIT_FOR_IDLE;
-- complete(&vscsi->wait_idle);
-- }
--}
--
--/**
- * ibmvscsis_disconnect() - Helper function to disconnect
- * @work: Pointer to work_struct, gives access to our adapter structure
- *
-@@ -576,7 +416,6 @@ static void ibmvscsis_disconnect(struct work_struct *work)
- proc_work);
- u16 new_state;
- bool wait_idle = false;
-- long rc = ADAPT_SUCCESS;
-
- spin_lock_bh(&vscsi->intr_lock);
- new_state = vscsi->new_state;
-@@ -590,7 +429,7 @@ static void ibmvscsis_disconnect(struct work_struct *work)
- * should transitition to the new state
- */
- switch (vscsi->state) {
-- /* Should never be called while in this state. */
-+ /* Should never be called while in this state. */
- case NO_QUEUE:
- /*
- * Can never transition from this state;
-@@ -629,30 +468,24 @@ static void ibmvscsis_disconnect(struct work_struct *work)
- vscsi->state = new_state;
- break;
-
-- /*
-- * If this is a transition into an error state.
-- * a client is attempting to establish a connection
-- * and has violated the RPA protocol.
-- * There can be nothing pending on the adapter although
-- * there can be requests in the command queue.
-- */
- case WAIT_ENABLED:
-- case PART_UP_WAIT_ENAB:
- switch (new_state) {
-- case ERR_DISCONNECT:
-- vscsi->flags |= RESPONSE_Q_DOWN;
-+ case UNCONFIGURING:
- vscsi->state = new_state;
-+ vscsi->flags |= RESPONSE_Q_DOWN;
- vscsi->flags &= ~(SCHEDULE_DISCONNECT |
- DISCONNECT_SCHEDULED);
-- ibmvscsis_free_command_q(vscsi);
-- break;
-- case ERR_DISCONNECT_RECONNECT:
-- ibmvscsis_reset_queue(vscsi, WAIT_ENABLED);
-+ dma_rmb();
-+ if (vscsi->flags & CFG_SLEEPING) {
-+ vscsi->flags &= ~CFG_SLEEPING;
-+ complete(&vscsi->unconfig);
-+ }
- break;
-
- /* should never happen */
-+ case ERR_DISCONNECT:
-+ case ERR_DISCONNECT_RECONNECT:
- case WAIT_IDLE:
-- rc = ERROR;
- dev_err(&vscsi->dev, "disconnect: invalid state %d for WAIT_IDLE\n",
- vscsi->state);
- break;
-@@ -661,6 +494,13 @@ static void ibmvscsis_disconnect(struct work_struct *work)
-
- case WAIT_IDLE:
- switch (new_state) {
-+ case UNCONFIGURING:
-+ vscsi->flags |= RESPONSE_Q_DOWN;
-+ vscsi->state = new_state;
-+ vscsi->flags &= ~(SCHEDULE_DISCONNECT |
-+ DISCONNECT_SCHEDULED);
-+ ibmvscsis_free_command_q(vscsi);
-+ break;
- case ERR_DISCONNECT:
- case ERR_DISCONNECT_RECONNECT:
- vscsi->state = new_state;
-@@ -765,45 +605,348 @@ static void ibmvscsis_post_disconnect(struct scsi_info *vscsi, uint new_state,
- else
- state = vscsi->state;
-
-- switch (state) {
-- case NO_QUEUE:
-- case UNCONFIGURING:
-- break;
-+ switch (state) {
-+ case NO_QUEUE:
-+ case UNCONFIGURING:
-+ break;
-+
-+ case ERR_DISCONNECTED:
-+ case ERR_DISCONNECT:
-+ case UNDEFINED:
-+ if (new_state == UNCONFIGURING)
-+ vscsi->new_state = new_state;
-+ break;
-+
-+ case ERR_DISCONNECT_RECONNECT:
-+ switch (new_state) {
-+ case UNCONFIGURING:
-+ case ERR_DISCONNECT:
-+ vscsi->new_state = new_state;
-+ break;
-+ default:
-+ break;
-+ }
-+ break;
-+
-+ case WAIT_ENABLED:
-+ case WAIT_IDLE:
-+ case WAIT_CONNECTION:
-+ case CONNECTED:
-+ case SRP_PROCESSING:
-+ vscsi->new_state = new_state;
-+ break;
-+
-+ default:
-+ break;
-+ }
-+ }
-+
-+ pr_debug("Leaving post_disconnect: flags 0x%x, new_state 0x%x\n",
-+ vscsi->flags, vscsi->new_state);
-+}
-+
-+/**
-+ * ibmvscsis_handle_init_compl_msg() - Respond to an Init Complete Message
-+ * @vscsi: Pointer to our adapter structure
-+ *
-+ * Must be called with interrupt lock held.
-+ */
-+static long ibmvscsis_handle_init_compl_msg(struct scsi_info *vscsi)
-+{
-+ long rc = ADAPT_SUCCESS;
-+
-+ switch (vscsi->state) {
-+ case NO_QUEUE:
-+ case ERR_DISCONNECT:
-+ case ERR_DISCONNECT_RECONNECT:
-+ case ERR_DISCONNECTED:
-+ case UNCONFIGURING:
-+ case UNDEFINED:
-+ rc = ERROR;
-+ break;
-+
-+ case WAIT_CONNECTION:
-+ vscsi->state = CONNECTED;
-+ break;
-+
-+ case WAIT_IDLE:
-+ case SRP_PROCESSING:
-+ case CONNECTED:
-+ case WAIT_ENABLED:
-+ default:
-+ rc = ERROR;
-+ dev_err(&vscsi->dev, "init_msg: invalid state %d to get init compl msg\n",
-+ vscsi->state);
-+ ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT_RECONNECT, 0);
-+ break;
-+ }
-+
-+ return rc;
-+}
-+
-+/**
-+ * ibmvscsis_handle_init_msg() - Respond to an Init Message
-+ * @vscsi: Pointer to our adapter structure
-+ *
-+ * Must be called with interrupt lock held.
-+ */
-+static long ibmvscsis_handle_init_msg(struct scsi_info *vscsi)
-+{
-+ long rc = ADAPT_SUCCESS;
-+
-+ switch (vscsi->state) {
-+ case WAIT_CONNECTION:
-+ rc = ibmvscsis_send_init_message(vscsi, INIT_COMPLETE_MSG);
-+ switch (rc) {
-+ case H_SUCCESS:
-+ vscsi->state = CONNECTED;
-+ break;
-+
-+ case H_PARAMETER:
-+ dev_err(&vscsi->dev, "init_msg: failed to send, rc %ld\n",
-+ rc);
-+ ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT, 0);
-+ break;
-+
-+ case H_DROPPED:
-+ dev_err(&vscsi->dev, "init_msg: failed to send, rc %ld\n",
-+ rc);
-+ rc = ERROR;
-+ ibmvscsis_post_disconnect(vscsi,
-+ ERR_DISCONNECT_RECONNECT, 0);
-+ break;
-+
-+ case H_CLOSED:
-+ pr_warn("init_msg: failed to send, rc %ld\n", rc);
-+ rc = 0;
-+ break;
-+ }
-+ break;
-+
-+ case UNDEFINED:
-+ rc = ERROR;
-+ break;
-+
-+ case UNCONFIGURING:
-+ break;
-+
-+ case WAIT_ENABLED:
-+ case CONNECTED:
-+ case SRP_PROCESSING:
-+ case WAIT_IDLE:
-+ case NO_QUEUE:
-+ case ERR_DISCONNECT:
-+ case ERR_DISCONNECT_RECONNECT:
-+ case ERR_DISCONNECTED:
-+ default:
-+ rc = ERROR;
-+ dev_err(&vscsi->dev, "init_msg: invalid state %d to get init msg\n",
-+ vscsi->state);
-+ ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT_RECONNECT, 0);
-+ break;
-+ }
-+
-+ return rc;
-+}
-+
-+/**
-+ * ibmvscsis_init_msg() - Respond to an init message
-+ * @vscsi: Pointer to our adapter structure
-+ * @crq: Pointer to CRQ element containing the Init Message
-+ *
-+ * EXECUTION ENVIRONMENT:
-+ * Interrupt, interrupt lock held
-+ */
-+static long ibmvscsis_init_msg(struct scsi_info *vscsi, struct viosrp_crq *crq)
-+{
-+ long rc = ADAPT_SUCCESS;
-+
-+ pr_debug("init_msg: state 0x%hx\n", vscsi->state);
-+
-+ rc = h_vioctl(vscsi->dds.unit_id, H_GET_PARTNER_INFO,
-+ (u64)vscsi->map_ioba | ((u64)PAGE_SIZE << 32), 0, 0, 0,
-+ 0);
-+ if (rc == H_SUCCESS) {
-+ vscsi->client_data.partition_number =
-+ be64_to_cpu(*(u64 *)vscsi->map_buf);
-+ pr_debug("init_msg, part num %d\n",
-+ vscsi->client_data.partition_number);
-+ } else {
-+ pr_debug("init_msg h_vioctl rc %ld\n", rc);
-+ rc = ADAPT_SUCCESS;
-+ }
-+
-+ if (crq->format == INIT_MSG) {
-+ rc = ibmvscsis_handle_init_msg(vscsi);
-+ } else if (crq->format == INIT_COMPLETE_MSG) {
-+ rc = ibmvscsis_handle_init_compl_msg(vscsi);
-+ } else {
-+ rc = ERROR;
-+ dev_err(&vscsi->dev, "init_msg: invalid format %d\n",
-+ (uint)crq->format);
-+ ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT_RECONNECT, 0);
-+ }
-+
-+ return rc;
-+}
-+
-+/**
-+ * ibmvscsis_establish_new_q() - Establish new CRQ queue
-+ * @vscsi: Pointer to our adapter structure
-+ *
-+ * Must be called with interrupt lock held.
-+ */
-+static long ibmvscsis_establish_new_q(struct scsi_info *vscsi)
-+{
-+ long rc = ADAPT_SUCCESS;
-+ uint format;
-+
-+ vscsi->flags &= PRESERVE_FLAG_FIELDS;
-+ vscsi->rsp_q_timer.timer_pops = 0;
-+ vscsi->debit = 0;
-+ vscsi->credit = 0;
-+
-+ rc = vio_enable_interrupts(vscsi->dma_dev);
-+ if (rc) {
-+ pr_warn("establish_new_q: failed to enable interrupts, rc %ld\n",
-+ rc);
-+ return rc;
-+ }
-+
-+ rc = ibmvscsis_check_init_msg(vscsi, &format);
-+ if (rc) {
-+ dev_err(&vscsi->dev, "establish_new_q: check_init_msg failed, rc %ld\n",
-+ rc);
-+ return rc;
-+ }
-+
-+ if (format == UNUSED_FORMAT) {
-+ rc = ibmvscsis_send_init_message(vscsi, INIT_MSG);
-+ switch (rc) {
-+ case H_SUCCESS:
-+ case H_DROPPED:
-+ case H_CLOSED:
-+ rc = ADAPT_SUCCESS;
-+ break;
-+
-+ case H_PARAMETER:
-+ case H_HARDWARE:
-+ break;
-+
-+ default:
-+ vscsi->state = UNDEFINED;
-+ rc = H_HARDWARE;
-+ break;
-+ }
-+ } else if (format == INIT_MSG) {
-+ rc = ibmvscsis_handle_init_msg(vscsi);
-+ }
-+
-+ return rc;
-+}
-+
-+/**
-+ * ibmvscsis_reset_queue() - Reset CRQ Queue
-+ * @vscsi: Pointer to our adapter structure
-+ *
-+ * This function calls h_free_q and then calls h_reg_q and does all
-+ * of the bookkeeping to get us back to where we can communicate.
-+ *
-+ * Actually, we don't always call h_free_crq. A problem was discovered
-+ * where one partition would close and reopen his queue, which would
-+ * cause his partner to get a transport event, which would cause him to
-+ * close and reopen his queue, which would cause the original partition
-+ * to get a transport event, etc., etc. To prevent this, we don't
-+ * actually close our queue if the client initiated the reset, (i.e.
-+ * either we got a transport event or we have detected that the client's
-+ * queue is gone)
-+ *
-+ * EXECUTION ENVIRONMENT:
-+ * Process environment, called with interrupt lock held
-+ */
-+static void ibmvscsis_reset_queue(struct scsi_info *vscsi)
-+{
-+ int bytes;
-+ long rc = ADAPT_SUCCESS;
-+
-+ pr_debug("reset_queue: flags 0x%x\n", vscsi->flags);
-+
-+ /* don't reset, the client did it for us */
-+ if (vscsi->flags & (CLIENT_FAILED | TRANS_EVENT)) {
-+ vscsi->flags &= PRESERVE_FLAG_FIELDS;
-+ vscsi->rsp_q_timer.timer_pops = 0;
-+ vscsi->debit = 0;
-+ vscsi->credit = 0;
-+ vscsi->state = WAIT_CONNECTION;
-+ vio_enable_interrupts(vscsi->dma_dev);
-+ } else {
-+ rc = ibmvscsis_free_command_q(vscsi);
-+ if (rc == ADAPT_SUCCESS) {
-+ vscsi->state = WAIT_CONNECTION;
-+
-+ bytes = vscsi->cmd_q.size * PAGE_SIZE;
-+ rc = h_reg_crq(vscsi->dds.unit_id,
-+ vscsi->cmd_q.crq_token, bytes);
-+ if (rc == H_CLOSED || rc == H_SUCCESS) {
-+ rc = ibmvscsis_establish_new_q(vscsi);
-+ }
-
-- case ERR_DISCONNECTED:
-- case ERR_DISCONNECT:
-- case UNDEFINED:
-- if (new_state == UNCONFIGURING)
-- vscsi->new_state = new_state;
-- break;
-+ if (rc != ADAPT_SUCCESS) {
-+ pr_debug("reset_queue: reg_crq rc %ld\n", rc);
-
-- case ERR_DISCONNECT_RECONNECT:
-- switch (new_state) {
-- case UNCONFIGURING:
-- case ERR_DISCONNECT:
-- vscsi->new_state = new_state;
-- break;
-- default:
-- break;
-+ vscsi->state = ERR_DISCONNECTED;
-+ vscsi->flags |= RESPONSE_Q_DOWN;
-+ ibmvscsis_free_command_q(vscsi);
- }
-- break;
-+ } else {
-+ vscsi->state = ERR_DISCONNECTED;
-+ vscsi->flags |= RESPONSE_Q_DOWN;
-+ }
-+ }
-+}
-
-- case WAIT_ENABLED:
-- case PART_UP_WAIT_ENAB:
-- case WAIT_IDLE:
-- case WAIT_CONNECTION:
-- case CONNECTED:
-- case SRP_PROCESSING:
-- vscsi->new_state = new_state;
-- break;
-+/**
-+ * ibmvscsis_free_cmd_resources() - Free command resources
-+ * @vscsi: Pointer to our adapter structure
-+ * @cmd: Command which is not longer in use
-+ *
-+ * Must be called with interrupt lock held.
-+ */
-+static void ibmvscsis_free_cmd_resources(struct scsi_info *vscsi,
-+ struct ibmvscsis_cmd *cmd)
-+{
-+ struct iu_entry *iue = cmd->iue;
-
-- default:
-- break;
-- }
-+ switch (cmd->type) {
-+ case TASK_MANAGEMENT:
-+ case SCSI_CDB:
-+ /*
-+ * When the queue goes down this value is cleared, so it
-+ * cannot be cleared in this general purpose function.
-+ */
-+ if (vscsi->debit)
-+ vscsi->debit -= 1;
-+ break;
-+ case ADAPTER_MAD:
-+ vscsi->flags &= ~PROCESSING_MAD;
-+ break;
-+ case UNSET_TYPE:
-+ break;
-+ default:
-+ dev_err(&vscsi->dev, "free_cmd_resources unknown type %d\n",
-+ cmd->type);
-+ break;
- }
-
-- pr_debug("Leaving post_disconnect: flags 0x%x, new_state 0x%x\n",
-- vscsi->flags, vscsi->new_state);
-+ cmd->iue = NULL;
-+ list_add_tail(&cmd->list, &vscsi->free_cmd);
-+ srp_iu_put(iue);
-+
-+ if (list_empty(&vscsi->active_q) && list_empty(&vscsi->schedule_q) &&
-+ list_empty(&vscsi->waiting_rsp) && (vscsi->flags & WAIT_FOR_IDLE)) {
-+ vscsi->flags &= ~WAIT_FOR_IDLE;
-+ complete(&vscsi->wait_idle);
-+ }
- }
-
- /**
-@@ -864,10 +1007,6 @@ static long ibmvscsis_trans_event(struct scsi_info *vscsi,
- TRANS_EVENT));
- break;
-
-- case PART_UP_WAIT_ENAB:
-- vscsi->state = WAIT_ENABLED;
-- break;
--
- case SRP_PROCESSING:
- if ((vscsi->debit > 0) ||
- !list_empty(&vscsi->schedule_q) ||
-@@ -896,7 +1035,7 @@ static long ibmvscsis_trans_event(struct scsi_info *vscsi,
- }
- }
-
-- rc = vscsi->flags & SCHEDULE_DISCONNECT;
-+ rc = vscsi->flags & SCHEDULE_DISCONNECT;
-
- pr_debug("Leaving trans_event: flags 0x%x, state 0x%hx, rc %ld\n",
- vscsi->flags, vscsi->state, rc);
-@@ -1067,16 +1206,28 @@ static void ibmvscsis_adapter_idle(struct scsi_info *vscsi)
- free_qs = true;
-
- switch (vscsi->state) {
-+ case UNCONFIGURING:
-+ ibmvscsis_free_command_q(vscsi);
-+ dma_rmb();
-+ isync();
-+ if (vscsi->flags & CFG_SLEEPING) {
-+ vscsi->flags &= ~CFG_SLEEPING;
-+ complete(&vscsi->unconfig);
-+ }
-+ break;
- case ERR_DISCONNECT_RECONNECT:
-- ibmvscsis_reset_queue(vscsi, WAIT_CONNECTION);
-+ ibmvscsis_reset_queue(vscsi);
- pr_debug("adapter_idle, disc_rec: flags 0x%x\n", vscsi->flags);
- break;
-
- case ERR_DISCONNECT:
- ibmvscsis_free_command_q(vscsi);
-- vscsi->flags &= ~DISCONNECT_SCHEDULED;
-+ vscsi->flags &= ~(SCHEDULE_DISCONNECT | DISCONNECT_SCHEDULED);
- vscsi->flags |= RESPONSE_Q_DOWN;
-- vscsi->state = ERR_DISCONNECTED;
-+ if (vscsi->tport.enabled)
-+ vscsi->state = ERR_DISCONNECTED;
-+ else
-+ vscsi->state = WAIT_ENABLED;
- pr_debug("adapter_idle, disc: flags 0x%x, state 0x%hx\n",
- vscsi->flags, vscsi->state);
- break;
-@@ -1221,7 +1372,7 @@ static long ibmvscsis_copy_crq_packet(struct scsi_info *vscsi,
- * @iue: Information Unit containing the Adapter Info MAD request
- *
- * EXECUTION ENVIRONMENT:
-- * Interrupt adpater lock is held
-+ * Interrupt adapter lock is held
- */
- static long ibmvscsis_adapter_info(struct scsi_info *vscsi,
- struct iu_entry *iue)
-@@ -1621,8 +1772,8 @@ static void ibmvscsis_send_messages(struct scsi_info *vscsi)
- be64_to_cpu(msg_hi),
- be64_to_cpu(cmd->rsp.tag));
-
-- pr_debug("send_messages: tag 0x%llx, rc %ld\n",
-- be64_to_cpu(cmd->rsp.tag), rc);
-+ pr_debug("send_messages: cmd %p, tag 0x%llx, rc %ld\n",
-+ cmd, be64_to_cpu(cmd->rsp.tag), rc);
-
- /* if all ok free up the command element resources */
- if (rc == H_SUCCESS) {
-@@ -1692,7 +1843,7 @@ static void ibmvscsis_send_mad_resp(struct scsi_info *vscsi,
- * @crq: Pointer to the CRQ entry containing the MAD request
- *
- * EXECUTION ENVIRONMENT:
-- * Interrupt called with adapter lock held
-+ * Interrupt, called with adapter lock held
- */
- static long ibmvscsis_mad(struct scsi_info *vscsi, struct viosrp_crq *crq)
- {
-@@ -1746,14 +1897,7 @@ static long ibmvscsis_mad(struct scsi_info *vscsi, struct viosrp_crq *crq)
-
- pr_debug("mad: type %d\n", be32_to_cpu(mad->type));
-
-- if (be16_to_cpu(mad->length) < 0) {
-- dev_err(&vscsi->dev, "mad: length is < 0\n");
-- ibmvscsis_post_disconnect(vscsi,
-- ERR_DISCONNECT_RECONNECT, 0);
-- rc = SRP_VIOLATION;
-- } else {
-- rc = ibmvscsis_process_mad(vscsi, iue);
-- }
-+ rc = ibmvscsis_process_mad(vscsi, iue);
-
- pr_debug("mad: status %hd, rc %ld\n", be16_to_cpu(mad->status),
- rc);
-@@ -1865,7 +2009,7 @@ static long ibmvscsis_srp_login_rej(struct scsi_info *vscsi,
- break;
- case H_PERMISSION:
- if (connection_broken(vscsi))
-- flag_bits = RESPONSE_Q_DOWN | CLIENT_FAILED;
-+ flag_bits = RESPONSE_Q_DOWN | CLIENT_FAILED;
- dev_err(&vscsi->dev, "login_rej: error copying to client, rc %ld\n",
- rc);
- ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT_RECONNECT,
-@@ -2090,248 +2234,98 @@ static void ibmvscsis_srp_cmd(struct scsi_info *vscsi, struct viosrp_crq *crq)
- break;
-
- case SRP_TSK_MGMT:
-- tsk = &vio_iu(iue)->srp.tsk_mgmt;
-- pr_debug("tsk_mgmt tag: %llu (0x%llx)\n", tsk->tag,
-- tsk->tag);
-- cmd->rsp.tag = tsk->tag;
-- vscsi->debit += 1;
-- cmd->type = TASK_MANAGEMENT;
-- list_add_tail(&cmd->list, &vscsi->schedule_q);
-- queue_work(vscsi->work_q, &cmd->work);
-- break;
--
-- case SRP_CMD:
-- pr_debug("srp_cmd tag: %llu (0x%llx)\n", srp->tag,
-- srp->tag);
-- cmd->rsp.tag = srp->tag;
-- vscsi->debit += 1;
-- cmd->type = SCSI_CDB;
-- /*
-- * We want to keep track of work waiting for
-- * the workqueue.
-- */
-- list_add_tail(&cmd->list, &vscsi->schedule_q);
-- queue_work(vscsi->work_q, &cmd->work);
-- break;
--
-- case SRP_I_LOGOUT:
-- rc = ibmvscsis_srp_i_logout(vscsi, cmd, crq);
-- break;
--
-- case SRP_CRED_RSP:
-- case SRP_AER_RSP:
-- default:
-- ibmvscsis_free_cmd_resources(vscsi, cmd);
-- dev_err(&vscsi->dev, "invalid srp cmd, opcode %d\n",
-- (uint)srp->opcode);
-- ibmvscsis_post_disconnect(vscsi,
-- ERR_DISCONNECT_RECONNECT, 0);
-- break;
-- }
-- } else if (srp->opcode == SRP_LOGIN_REQ && vscsi->state == CONNECTED) {
-- rc = ibmvscsis_srp_login(vscsi, cmd, crq);
-- } else {
-- ibmvscsis_free_cmd_resources(vscsi, cmd);
-- dev_err(&vscsi->dev, "Invalid state %d to handle srp cmd\n",
-- vscsi->state);
-- ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT_RECONNECT, 0);
-- }
--}
--
--/**
-- * ibmvscsis_ping_response() - Respond to a ping request
-- * @vscsi: Pointer to our adapter structure
-- *
-- * Let the client know that the server is alive and waiting on
-- * its native I/O stack.
-- * If any type of error occurs from the call to queue a ping
-- * response then the client is either not accepting or receiving
-- * interrupts. Disconnect with an error.
-- *
-- * EXECUTION ENVIRONMENT:
-- * Interrupt, interrupt lock held
-- */
--static long ibmvscsis_ping_response(struct scsi_info *vscsi)
--{
-- struct viosrp_crq *crq;
-- u64 buffer[2] = { 0, 0 };
-- long rc;
--
-- crq = (struct viosrp_crq *)&buffer;
-- crq->valid = VALID_CMD_RESP_EL;
-- crq->format = (u8)MESSAGE_IN_CRQ;
-- crq->status = PING_RESPONSE;
--
-- rc = h_send_crq(vscsi->dds.unit_id, cpu_to_be64(buffer[MSG_HI]),
-- cpu_to_be64(buffer[MSG_LOW]));
--
-- switch (rc) {
-- case H_SUCCESS:
-- break;
-- case H_CLOSED:
-- vscsi->flags |= CLIENT_FAILED;
-- case H_DROPPED:
-- vscsi->flags |= RESPONSE_Q_DOWN;
-- case H_REMOTE_PARM:
-- dev_err(&vscsi->dev, "ping_response: h_send_crq failed, rc %ld\n",
-- rc);
-- ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT_RECONNECT, 0);
-- break;
-- default:
-- dev_err(&vscsi->dev, "ping_response: h_send_crq returned unknown rc %ld\n",
-- rc);
-- ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT, 0);
-- break;
-- }
--
-- return rc;
--}
--
--/**
-- * ibmvscsis_handle_init_compl_msg() - Respond to an Init Complete Message
-- * @vscsi: Pointer to our adapter structure
-- *
-- * Must be called with interrupt lock held.
-- */
--static long ibmvscsis_handle_init_compl_msg(struct scsi_info *vscsi)
--{
-- long rc = ADAPT_SUCCESS;
--
-- switch (vscsi->state) {
-- case NO_QUEUE:
-- case ERR_DISCONNECT:
-- case ERR_DISCONNECT_RECONNECT:
-- case ERR_DISCONNECTED:
-- case UNCONFIGURING:
-- case UNDEFINED:
-- rc = ERROR;
-- break;
--
-- case WAIT_CONNECTION:
-- vscsi->state = CONNECTED;
-- break;
--
-- case WAIT_IDLE:
-- case SRP_PROCESSING:
-- case CONNECTED:
-- case WAIT_ENABLED:
-- case PART_UP_WAIT_ENAB:
-- default:
-- rc = ERROR;
-- dev_err(&vscsi->dev, "init_msg: invalid state %d to get init compl msg\n",
-- vscsi->state);
-- ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT_RECONNECT, 0);
-- break;
-- }
--
-- return rc;
--}
--
--/**
-- * ibmvscsis_handle_init_msg() - Respond to an Init Message
-- * @vscsi: Pointer to our adapter structure
-- *
-- * Must be called with interrupt lock held.
-- */
--static long ibmvscsis_handle_init_msg(struct scsi_info *vscsi)
--{
-- long rc = ADAPT_SUCCESS;
--
-- switch (vscsi->state) {
-- case WAIT_ENABLED:
-- vscsi->state = PART_UP_WAIT_ENAB;
-- break;
-+ tsk = &vio_iu(iue)->srp.tsk_mgmt;
-+ pr_debug("tsk_mgmt tag: %llu (0x%llx)\n", tsk->tag,
-+ tsk->tag);
-+ cmd->rsp.tag = tsk->tag;
-+ vscsi->debit += 1;
-+ cmd->type = TASK_MANAGEMENT;
-+ list_add_tail(&cmd->list, &vscsi->schedule_q);
-+ queue_work(vscsi->work_q, &cmd->work);
-+ break;
-
-- case WAIT_CONNECTION:
-- rc = ibmvscsis_send_init_message(vscsi, INIT_COMPLETE_MSG);
-- switch (rc) {
-- case H_SUCCESS:
-- vscsi->state = CONNECTED;
-+ case SRP_CMD:
-+ pr_debug("srp_cmd tag: %llu (0x%llx)\n", srp->tag,
-+ srp->tag);
-+ cmd->rsp.tag = srp->tag;
-+ vscsi->debit += 1;
-+ cmd->type = SCSI_CDB;
-+ /*
-+ * We want to keep track of work waiting for
-+ * the workqueue.
-+ */
-+ list_add_tail(&cmd->list, &vscsi->schedule_q);
-+ queue_work(vscsi->work_q, &cmd->work);
- break;
-
-- case H_PARAMETER:
-- dev_err(&vscsi->dev, "init_msg: failed to send, rc %ld\n",
-- rc);
-- ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT, 0);
-+ case SRP_I_LOGOUT:
-+ rc = ibmvscsis_srp_i_logout(vscsi, cmd, crq);
- break;
-
-- case H_DROPPED:
-- dev_err(&vscsi->dev, "init_msg: failed to send, rc %ld\n",
-- rc);
-- rc = ERROR;
-+ case SRP_CRED_RSP:
-+ case SRP_AER_RSP:
-+ default:
-+ ibmvscsis_free_cmd_resources(vscsi, cmd);
-+ dev_err(&vscsi->dev, "invalid srp cmd, opcode %d\n",
-+ (uint)srp->opcode);
- ibmvscsis_post_disconnect(vscsi,
- ERR_DISCONNECT_RECONNECT, 0);
- break;
--
-- case H_CLOSED:
-- pr_warn("init_msg: failed to send, rc %ld\n", rc);
-- rc = 0;
-- break;
- }
-- break;
--
-- case UNDEFINED:
-- rc = ERROR;
-- break;
--
-- case UNCONFIGURING:
-- break;
--
-- case PART_UP_WAIT_ENAB:
-- case CONNECTED:
-- case SRP_PROCESSING:
-- case WAIT_IDLE:
-- case NO_QUEUE:
-- case ERR_DISCONNECT:
-- case ERR_DISCONNECT_RECONNECT:
-- case ERR_DISCONNECTED:
-- default:
-- rc = ERROR;
-- dev_err(&vscsi->dev, "init_msg: invalid state %d to get init msg\n",
-+ } else if (srp->opcode == SRP_LOGIN_REQ && vscsi->state == CONNECTED) {
-+ rc = ibmvscsis_srp_login(vscsi, cmd, crq);
-+ } else {
-+ ibmvscsis_free_cmd_resources(vscsi, cmd);
-+ dev_err(&vscsi->dev, "Invalid state %d to handle srp cmd\n",
- vscsi->state);
- ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT_RECONNECT, 0);
-- break;
- }
--
-- return rc;
- }
-
- /**
-- * ibmvscsis_init_msg() - Respond to an init message
-+ * ibmvscsis_ping_response() - Respond to a ping request
- * @vscsi: Pointer to our adapter structure
-- * @crq: Pointer to CRQ element containing the Init Message
-+ *
-+ * Let the client know that the server is alive and waiting on
-+ * its native I/O stack.
-+ * If any type of error occurs from the call to queue a ping
-+ * response then the client is either not accepting or receiving
-+ * interrupts. Disconnect with an error.
- *
- * EXECUTION ENVIRONMENT:
- * Interrupt, interrupt lock held
- */
--static long ibmvscsis_init_msg(struct scsi_info *vscsi, struct viosrp_crq *crq)
-+static long ibmvscsis_ping_response(struct scsi_info *vscsi)
- {
-- long rc = ADAPT_SUCCESS;
-+ struct viosrp_crq *crq;
-+ u64 buffer[2] = { 0, 0 };
-+ long rc;
-
-- pr_debug("init_msg: state 0x%hx\n", vscsi->state);
-+ crq = (struct viosrp_crq *)&buffer;
-+ crq->valid = VALID_CMD_RESP_EL;
-+ crq->format = (u8)MESSAGE_IN_CRQ;
-+ crq->status = PING_RESPONSE;
-
-- rc = h_vioctl(vscsi->dds.unit_id, H_GET_PARTNER_INFO,
-- (u64)vscsi->map_ioba | ((u64)PAGE_SIZE << 32), 0, 0, 0,
-- 0);
-- if (rc == H_SUCCESS) {
-- vscsi->client_data.partition_number =
-- be64_to_cpu(*(u64 *)vscsi->map_buf);
-- pr_debug("init_msg, part num %d\n",
-- vscsi->client_data.partition_number);
-- } else {
-- pr_debug("init_msg h_vioctl rc %ld\n", rc);
-- rc = ADAPT_SUCCESS;
-- }
-+ rc = h_send_crq(vscsi->dds.unit_id, cpu_to_be64(buffer[MSG_HI]),
-+ cpu_to_be64(buffer[MSG_LOW]));
-
-- if (crq->format == INIT_MSG) {
-- rc = ibmvscsis_handle_init_msg(vscsi);
-- } else if (crq->format == INIT_COMPLETE_MSG) {
-- rc = ibmvscsis_handle_init_compl_msg(vscsi);
-- } else {
-- rc = ERROR;
-- dev_err(&vscsi->dev, "init_msg: invalid format %d\n",
-- (uint)crq->format);
-+ switch (rc) {
-+ case H_SUCCESS:
-+ break;
-+ case H_CLOSED:
-+ vscsi->flags |= CLIENT_FAILED;
-+ case H_DROPPED:
-+ vscsi->flags |= RESPONSE_Q_DOWN;
-+ case H_REMOTE_PARM:
-+ dev_err(&vscsi->dev, "ping_response: h_send_crq failed, rc %ld\n",
-+ rc);
- ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT_RECONNECT, 0);
-+ break;
-+ default:
-+ dev_err(&vscsi->dev, "ping_response: h_send_crq returned unknown rc %ld\n",
-+ rc);
-+ ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT, 0);
-+ break;
- }
-
- return rc;
-@@ -2392,7 +2386,7 @@ static long ibmvscsis_parse_command(struct scsi_info *vscsi,
- break;
-
- case VALID_TRANS_EVENT:
-- rc = ibmvscsis_trans_event(vscsi, crq);
-+ rc = ibmvscsis_trans_event(vscsi, crq);
- break;
-
- case VALID_INIT_MSG:
-@@ -2523,7 +2517,6 @@ static void ibmvscsis_parse_cmd(struct scsi_info *vscsi,
- dev_err(&vscsi->dev, "0x%llx: parsing SRP descriptor table failed.\n",
- srp->tag);
- goto fail;
-- return;
- }
-
- cmd->rsp.sol_not = srp->sol_not;
-@@ -2560,6 +2553,10 @@ static void ibmvscsis_parse_cmd(struct scsi_info *vscsi,
- data_len, attr, dir, 0);
- if (rc) {
- dev_err(&vscsi->dev, "target_submit_cmd failed, rc %d\n", rc);
-+ spin_lock_bh(&vscsi->intr_lock);
-+ list_del(&cmd->list);
-+ ibmvscsis_free_cmd_resources(vscsi, cmd);
-+ spin_unlock_bh(&vscsi->intr_lock);
- goto fail;
- }
- return;
-@@ -2639,6 +2636,9 @@ static void ibmvscsis_parse_task(struct scsi_info *vscsi,
- if (rc) {
- dev_err(&vscsi->dev, "target_submit_tmr failed, rc %d\n",
- rc);
-+ spin_lock_bh(&vscsi->intr_lock);
-+ list_del(&cmd->list);
-+ spin_unlock_bh(&vscsi->intr_lock);
- cmd->se_cmd.se_tmr_req->response =
- TMR_FUNCTION_REJECTED;
- }
-@@ -2787,36 +2787,6 @@ static irqreturn_t ibmvscsis_interrupt(int dummy, void *data)
- }
-
- /**
-- * ibmvscsis_check_q() - Helper function to Check Init Message Valid
-- * @vscsi: Pointer to our adapter structure
-- *
-- * Checks if a initialize message was queued by the initiatior
-- * while the timing window was open. This function is called from
-- * probe after the CRQ is created and interrupts are enabled.
-- * It would only be used by adapters who wait for some event before
-- * completing the init handshake with the client. For ibmvscsi, this
-- * event is waiting for the port to be enabled.
-- *
-- * EXECUTION ENVIRONMENT:
-- * Process level only, interrupt lock held
-- */
--static long ibmvscsis_check_q(struct scsi_info *vscsi)
--{
-- uint format;
-- long rc;
--
-- rc = ibmvscsis_check_init_msg(vscsi, &format);
-- if (rc)
-- ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT_RECONNECT, 0);
-- else if (format == UNUSED_FORMAT)
-- vscsi->state = WAIT_ENABLED;
-- else
-- vscsi->state = PART_UP_WAIT_ENAB;
--
-- return rc;
--}
--
--/**
- * ibmvscsis_enable_change_state() - Set new state based on enabled status
- * @vscsi: Pointer to our adapter structure
- *
-@@ -2827,77 +2797,19 @@ static long ibmvscsis_check_q(struct scsi_info *vscsi)
- */
- static long ibmvscsis_enable_change_state(struct scsi_info *vscsi)
- {
-+ int bytes;
- long rc = ADAPT_SUCCESS;
-
--handle_state_change:
-- switch (vscsi->state) {
-- case WAIT_ENABLED:
-- rc = ibmvscsis_send_init_message(vscsi, INIT_MSG);
-- switch (rc) {
-- case H_SUCCESS:
-- case H_DROPPED:
-- case H_CLOSED:
-- vscsi->state = WAIT_CONNECTION;
-- rc = ADAPT_SUCCESS;
-- break;
--
-- case H_PARAMETER:
-- break;
--
-- case H_HARDWARE:
-- break;
--
-- default:
-- vscsi->state = UNDEFINED;
-- rc = H_HARDWARE;
-- break;
-- }
-- break;
-- case PART_UP_WAIT_ENAB:
-- rc = ibmvscsis_send_init_message(vscsi, INIT_COMPLETE_MSG);
-- switch (rc) {
-- case H_SUCCESS:
-- vscsi->state = CONNECTED;
-- rc = ADAPT_SUCCESS;
-- break;
--
-- case H_DROPPED:
-- case H_CLOSED:
-- vscsi->state = WAIT_ENABLED;
-- goto handle_state_change;
--
-- case H_PARAMETER:
-- break;
--
-- case H_HARDWARE:
-- break;
--
-- default:
-- rc = H_HARDWARE;
-- break;
-- }
-- break;
--
-- case WAIT_CONNECTION:
-- case WAIT_IDLE:
-- case SRP_PROCESSING:
-- case CONNECTED:
-- rc = ADAPT_SUCCESS;
-- break;
-- /* should not be able to get here */
-- case UNCONFIGURING:
-- rc = ERROR;
-- vscsi->state = UNDEFINED;
-- break;
-+ bytes = vscsi->cmd_q.size * PAGE_SIZE;
-+ rc = h_reg_crq(vscsi->dds.unit_id, vscsi->cmd_q.crq_token, bytes);
-+ if (rc == H_CLOSED || rc == H_SUCCESS) {
-+ vscsi->state = WAIT_CONNECTION;
-+ rc = ibmvscsis_establish_new_q(vscsi);
-+ }
-
-- /* driver should never allow this to happen */
-- case ERR_DISCONNECT:
-- case ERR_DISCONNECT_RECONNECT:
-- default:
-- dev_err(&vscsi->dev, "in invalid state %d during enable_change_state\n",
-- vscsi->state);
-- rc = ADAPT_SUCCESS;
-- break;
-+ if (rc != ADAPT_SUCCESS) {
-+ vscsi->state = ERR_DISCONNECTED;
-+ vscsi->flags |= RESPONSE_Q_DOWN;
- }
-
- return rc;
-@@ -2917,7 +2829,6 @@ static long ibmvscsis_enable_change_state(struct scsi_info *vscsi)
- */
- static long ibmvscsis_create_command_q(struct scsi_info *vscsi, int num_cmds)
- {
-- long rc = 0;
- int pages;
- struct vio_dev *vdev = vscsi->dma_dev;
-
-@@ -2941,22 +2852,7 @@ static long ibmvscsis_create_command_q(struct scsi_info *vscsi, int num_cmds)
- return -ENOMEM;
- }
-
-- rc = h_reg_crq(vscsi->dds.unit_id, vscsi->cmd_q.crq_token, PAGE_SIZE);
-- if (rc) {
-- if (rc == H_CLOSED) {
-- vscsi->state = WAIT_ENABLED;
-- rc = 0;
-- } else {
-- dma_unmap_single(&vdev->dev, vscsi->cmd_q.crq_token,
-- PAGE_SIZE, DMA_BIDIRECTIONAL);
-- free_page((unsigned long)vscsi->cmd_q.base_addr);
-- rc = -ENODEV;
-- }
-- } else {
-- vscsi->state = WAIT_ENABLED;
-- }
--
-- return rc;
-+ return 0;
- }
-
- /**
-@@ -3271,7 +3167,7 @@ static void ibmvscsis_handle_crq(unsigned long data)
- /*
- * if we are in a path where we are waiting for all pending commands
- * to complete because we received a transport event and anything in
-- * the command queue is for a new connection, do nothing
-+ * the command queue is for a new connection, do nothing
- */
- if (TARGET_STOP(vscsi)) {
- vio_enable_interrupts(vscsi->dma_dev);
-@@ -3315,7 +3211,7 @@ static void ibmvscsis_handle_crq(unsigned long data)
- * everything but transport events on the queue
- *
- * need to decrement the queue index so we can
-- * look at the elment again
-+ * look at the element again
- */
- if (vscsi->cmd_q.index)
- vscsi->cmd_q.index -= 1;
-@@ -3379,7 +3275,8 @@ static int ibmvscsis_probe(struct vio_dev *vdev,
- INIT_LIST_HEAD(&vscsi->waiting_rsp);
- INIT_LIST_HEAD(&vscsi->active_q);
-
-- snprintf(vscsi->tport.tport_name, 256, "%s", dev_name(&vdev->dev));
-+ snprintf(vscsi->tport.tport_name, IBMVSCSIS_NAMELEN, "%s",
-+ dev_name(&vdev->dev));
-
- pr_debug("probe tport_name: %s\n", vscsi->tport.tport_name);
-
-@@ -3394,6 +3291,9 @@ static int ibmvscsis_probe(struct vio_dev *vdev,
- strncat(vscsi->eye, vdev->name, MAX_EYE);
-
- vscsi->dds.unit_id = vdev->unit_address;
-+ strncpy(vscsi->dds.partition_name, partition_name,
-+ sizeof(vscsi->dds.partition_name));
-+ vscsi->dds.partition_num = partition_number;
-
- spin_lock_bh(&ibmvscsis_dev_lock);
- list_add_tail(&vscsi->list, &ibmvscsis_dev_list);
-@@ -3470,6 +3370,7 @@ static int ibmvscsis_probe(struct vio_dev *vdev,
- (unsigned long)vscsi);
-
- init_completion(&vscsi->wait_idle);
-+ init_completion(&vscsi->unconfig);
-
- snprintf(wq_name, 24, "ibmvscsis%s", dev_name(&vdev->dev));
- vscsi->work_q = create_workqueue(wq_name);
-@@ -3486,31 +3387,12 @@ static int ibmvscsis_probe(struct vio_dev *vdev,
- goto destroy_WQ;
- }
-
-- spin_lock_bh(&vscsi->intr_lock);
-- vio_enable_interrupts(vdev);
-- if (rc) {
-- dev_err(&vscsi->dev, "enabling interrupts failed, rc %d\n", rc);
-- rc = -ENODEV;
-- spin_unlock_bh(&vscsi->intr_lock);
-- goto free_irq;
-- }
--
-- if (ibmvscsis_check_q(vscsi)) {
-- rc = ERROR;
-- dev_err(&vscsi->dev, "probe: check_q failed, rc %d\n", rc);
-- spin_unlock_bh(&vscsi->intr_lock);
-- goto disable_interrupt;
-- }
-- spin_unlock_bh(&vscsi->intr_lock);
-+ vscsi->state = WAIT_ENABLED;
-
- dev_set_drvdata(&vdev->dev, vscsi);
-
- return 0;
-
--disable_interrupt:
-- vio_disable_interrupts(vdev);
--free_irq:
-- free_irq(vdev->irq, vscsi);
- destroy_WQ:
- destroy_workqueue(vscsi->work_q);
- unmap_buf:
-@@ -3544,10 +3426,11 @@ static int ibmvscsis_remove(struct vio_dev *vdev)
-
- pr_debug("remove (%s)\n", dev_name(&vscsi->dma_dev->dev));
-
-- /*
-- * TBD: Need to handle if there are commands on the waiting_rsp q
-- * Actually, can there still be cmds outstanding to tcm?
-- */
-+ spin_lock_bh(&vscsi->intr_lock);
-+ ibmvscsis_post_disconnect(vscsi, UNCONFIGURING, 0);
-+ vscsi->flags |= CFG_SLEEPING;
-+ spin_unlock_bh(&vscsi->intr_lock);
-+ wait_for_completion(&vscsi->unconfig);
-
- vio_disable_interrupts(vdev);
- free_irq(vdev->irq, vscsi);
-@@ -3556,7 +3439,6 @@ static int ibmvscsis_remove(struct vio_dev *vdev)
- DMA_BIDIRECTIONAL);
- kfree(vscsi->map_buf);
- tasklet_kill(&vscsi->work_task);
-- ibmvscsis_unregister_command_q(vscsi);
- ibmvscsis_destroy_command_q(vscsi);
- ibmvscsis_freetimer(vscsi);
- ibmvscsis_free_cmds(vscsi);
-@@ -3610,7 +3492,7 @@ static int ibmvscsis_get_system_info(void)
-
- num = of_get_property(rootdn, "ibm,partition-no", NULL);
- if (num)
-- partition_number = *num;
-+ partition_number = of_read_number(num, 1);
-
- of_node_put(rootdn);
-
-@@ -3904,18 +3786,22 @@ static ssize_t ibmvscsis_tpg_enable_store(struct config_item *item,
- }
-
- if (tmp) {
-- tport->enabled = true;
- spin_lock_bh(&vscsi->intr_lock);
-+ tport->enabled = true;
- lrc = ibmvscsis_enable_change_state(vscsi);
- if (lrc)
- pr_err("enable_change_state failed, rc %ld state %d\n",
- lrc, vscsi->state);
- spin_unlock_bh(&vscsi->intr_lock);
- } else {
-+ spin_lock_bh(&vscsi->intr_lock);
- tport->enabled = false;
-+ /* This simulates the server going down */
-+ ibmvscsis_post_disconnect(vscsi, ERR_DISCONNECT, 0);
-+ spin_unlock_bh(&vscsi->intr_lock);
- }
-
-- pr_debug("tpg_enable_store, state %d\n", vscsi->state);
-+ pr_debug("tpg_enable_store, tmp %ld, state %d\n", tmp, vscsi->state);
-
- return count;
- }
-@@ -3985,10 +3871,10 @@ static struct attribute *ibmvscsis_dev_attrs[] = {
- ATTRIBUTE_GROUPS(ibmvscsis_dev);
-
- static struct class ibmvscsis_class = {
-- .name = "ibmvscsis",
-- .dev_release = ibmvscsis_dev_release,
-- .class_attrs = ibmvscsis_class_attrs,
-- .dev_groups = ibmvscsis_dev_groups,
-+ .name = "ibmvscsis",
-+ .dev_release = ibmvscsis_dev_release,
-+ .class_attrs = ibmvscsis_class_attrs,
-+ .dev_groups = ibmvscsis_dev_groups,
- };
-
- static struct vio_device_id ibmvscsis_device_table[] = {
-diff --git a/drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.h b/drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.h
-index 981a0c9..98b0ca7 100644
---- a/drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.h
-+++ b/drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.h
-@@ -204,8 +204,6 @@ struct scsi_info {
- struct list_head waiting_rsp;
- #define NO_QUEUE 0x00
- #define WAIT_ENABLED 0X01
-- /* driver has received an initialize command */
--#define PART_UP_WAIT_ENAB 0x02
- #define WAIT_CONNECTION 0x04
- /* have established a connection */
- #define CONNECTED 0x08
-@@ -259,6 +257,8 @@ struct scsi_info {
- #define SCHEDULE_DISCONNECT 0x00400
- /* disconnect handler is scheduled */
- #define DISCONNECT_SCHEDULED 0x00800
-+ /* remove function is sleeping */
-+#define CFG_SLEEPING 0x01000
- u32 flags;
- /* adapter lock */
- spinlock_t intr_lock;
-@@ -287,6 +287,7 @@ struct scsi_info {
-
- struct workqueue_struct *work_q;
- struct completion wait_idle;
-+ struct completion unconfig;
- struct device dev;
- struct vio_dev *dma_dev;
- struct srp_target target;
-diff --git a/drivers/tty/serial/8250/8250_pci.c b/drivers/tty/serial/8250/8250_pci.c
-index 4d09bd4..6e3e636 100644
---- a/drivers/tty/serial/8250/8250_pci.c
-+++ b/drivers/tty/serial/8250/8250_pci.c
-@@ -52,6 +52,7 @@ struct serial_private {
- struct pci_dev *dev;
- unsigned int nr;
- struct pci_serial_quirk *quirk;
-+ const struct pciserial_board *board;
- int line[0];
- };
-
-@@ -3871,6 +3872,7 @@ pciserial_init_ports(struct pci_dev *dev, const struct pciserial_board *board)
- }
- }
- priv->nr = i;
-+ priv->board = board;
- return priv;
-
- err_deinit:
-@@ -3881,7 +3883,7 @@ pciserial_init_ports(struct pci_dev *dev, const struct pciserial_board *board)
- }
- EXPORT_SYMBOL_GPL(pciserial_init_ports);
-
--void pciserial_remove_ports(struct serial_private *priv)
-+void pciserial_detach_ports(struct serial_private *priv)
- {
- struct pci_serial_quirk *quirk;
- int i;
-@@ -3895,7 +3897,11 @@ void pciserial_remove_ports(struct serial_private *priv)
- quirk = find_quirk(priv->dev);
- if (quirk->exit)
- quirk->exit(priv->dev);
-+}
-
-+void pciserial_remove_ports(struct serial_private *priv)
-+{
-+ pciserial_detach_ports(priv);
- kfree(priv);
- }
- EXPORT_SYMBOL_GPL(pciserial_remove_ports);
-@@ -5590,7 +5596,7 @@ static pci_ers_result_t serial8250_io_error_detected(struct pci_dev *dev,
- return PCI_ERS_RESULT_DISCONNECT;
-
- if (priv)
-- pciserial_suspend_ports(priv);
-+ pciserial_detach_ports(priv);
-
- pci_disable_device(dev);
-
-@@ -5615,9 +5621,18 @@ static pci_ers_result_t serial8250_io_slot_reset(struct pci_dev *dev)
- static void serial8250_io_resume(struct pci_dev *dev)
- {
- struct serial_private *priv = pci_get_drvdata(dev);
-+ const struct pciserial_board *board;
-
-- if (priv)
-- pciserial_resume_ports(priv);
-+ if (!priv)
-+ return;
-+
-+ board = priv->board;
-+ kfree(priv);
-+ priv = pciserial_init_ports(dev, board);
-+
-+ if (!IS_ERR(priv)) {
-+ pci_set_drvdata(dev, priv);
-+ }
- }
-
- static const struct pci_error_handlers serial8250_err_handler = {
-diff --git a/drivers/usb/gadget/udc/atmel_usba_udc.c b/drivers/usb/gadget/udc/atmel_usba_udc.c
-index 45bc997..a95b3e7 100644
---- a/drivers/usb/gadget/udc/atmel_usba_udc.c
-+++ b/drivers/usb/gadget/udc/atmel_usba_udc.c
-@@ -1978,7 +1978,8 @@ static struct usba_ep * atmel_udc_of_init(struct platform_device *pdev,
- dev_err(&pdev->dev, "of_probe: name error(%d)\n", ret);
- goto err;
- }
-- ep->ep.name = kasprintf(GFP_KERNEL, "ep%d", ep->index);
-+ sprintf(ep->name, "ep%d", ep->index);
-+ ep->ep.name = ep->name;
-
- ep->ep_regs = udc->regs + USBA_EPT_BASE(i);
- ep->dma_regs = udc->regs + USBA_DMA_BASE(i);
-diff --git a/drivers/usb/gadget/udc/atmel_usba_udc.h b/drivers/usb/gadget/udc/atmel_usba_udc.h
-index 3e1c9d5..b03b2eb 100644
---- a/drivers/usb/gadget/udc/atmel_usba_udc.h
-+++ b/drivers/usb/gadget/udc/atmel_usba_udc.h
-@@ -280,6 +280,7 @@ struct usba_ep {
- void __iomem *ep_regs;
- void __iomem *dma_regs;
- void __iomem *fifo;
-+ char name[8];
- struct usb_ep ep;
- struct usba_udc *udc;
-
-diff --git a/drivers/vfio/vfio_iommu_spapr_tce.c b/drivers/vfio/vfio_iommu_spapr_tce.c
-index 80378dd..c882357 100644
---- a/drivers/vfio/vfio_iommu_spapr_tce.c
-+++ b/drivers/vfio/vfio_iommu_spapr_tce.c
-@@ -31,49 +31,49 @@
- static void tce_iommu_detach_group(void *iommu_data,
- struct iommu_group *iommu_group);
-
--static long try_increment_locked_vm(long npages)
-+static long try_increment_locked_vm(struct mm_struct *mm, long npages)
- {
- long ret = 0, locked, lock_limit;
-
-- if (!current || !current->mm)
-- return -ESRCH; /* process exited */
-+ if (WARN_ON_ONCE(!mm))
-+ return -EPERM;
-
- if (!npages)
- return 0;
-
-- down_write(&current->mm->mmap_sem);
-- locked = current->mm->locked_vm + npages;
-+ down_write(&mm->mmap_sem);
-+ locked = mm->locked_vm + npages;
- lock_limit = rlimit(RLIMIT_MEMLOCK) >> PAGE_SHIFT;
- if (locked > lock_limit && !capable(CAP_IPC_LOCK))
- ret = -ENOMEM;
- else
-- current->mm->locked_vm += npages;
-+ mm->locked_vm += npages;
-
- pr_debug("[%d] RLIMIT_MEMLOCK +%ld %ld/%ld%s\n", current->pid,
- npages << PAGE_SHIFT,
-- current->mm->locked_vm << PAGE_SHIFT,
-+ mm->locked_vm << PAGE_SHIFT,
- rlimit(RLIMIT_MEMLOCK),
- ret ? " - exceeded" : "");
-
-- up_write(&current->mm->mmap_sem);
-+ up_write(&mm->mmap_sem);
-
- return ret;
- }
-
--static void decrement_locked_vm(long npages)
-+static void decrement_locked_vm(struct mm_struct *mm, long npages)
- {
-- if (!current || !current->mm || !npages)
-- return; /* process exited */
-+ if (!mm || !npages)
-+ return;
-
-- down_write(&current->mm->mmap_sem);
-- if (WARN_ON_ONCE(npages > current->mm->locked_vm))
-- npages = current->mm->locked_vm;
-- current->mm->locked_vm -= npages;
-+ down_write(&mm->mmap_sem);
-+ if (WARN_ON_ONCE(npages > mm->locked_vm))
-+ npages = mm->locked_vm;
-+ mm->locked_vm -= npages;
- pr_debug("[%d] RLIMIT_MEMLOCK -%ld %ld/%ld\n", current->pid,
- npages << PAGE_SHIFT,
-- current->mm->locked_vm << PAGE_SHIFT,
-+ mm->locked_vm << PAGE_SHIFT,
- rlimit(RLIMIT_MEMLOCK));
-- up_write(&current->mm->mmap_sem);
-+ up_write(&mm->mmap_sem);
- }
-
- /*
-@@ -89,6 +89,15 @@ struct tce_iommu_group {
- };
-
- /*
-+ * A container needs to remember which preregistered region it has
-+ * referenced to do proper cleanup at the userspace process exit.
-+ */
-+struct tce_iommu_prereg {
-+ struct list_head next;
-+ struct mm_iommu_table_group_mem_t *mem;
-+};
-+
-+/*
- * The container descriptor supports only a single group per container.
- * Required by the API as the container is not supplied with the IOMMU group
- * at the moment of initialization.
-@@ -97,24 +106,68 @@ struct tce_container {
- struct mutex lock;
- bool enabled;
- bool v2;
-+ bool def_window_pending;
- unsigned long locked_pages;
-+ struct mm_struct *mm;
- struct iommu_table *tables[IOMMU_TABLE_GROUP_MAX_TABLES];
- struct list_head group_list;
-+ struct list_head prereg_list;
- };
-
-+static long tce_iommu_mm_set(struct tce_container *container)
-+{
-+ if (container->mm) {
-+ if (container->mm == current->mm)
-+ return 0;
-+ return -EPERM;
-+ }
-+ BUG_ON(!current->mm);
-+ container->mm = current->mm;
-+ atomic_inc(&container->mm->mm_count);
-+
-+ return 0;
-+}
-+
-+static long tce_iommu_prereg_free(struct tce_container *container,
-+ struct tce_iommu_prereg *tcemem)
-+{
-+ long ret;
-+
-+ ret = mm_iommu_put(container->mm, tcemem->mem);
-+ if (ret)
-+ return ret;
-+
-+ list_del(&tcemem->next);
-+ kfree(tcemem);
-+
-+ return 0;
-+}
-+
- static long tce_iommu_unregister_pages(struct tce_container *container,
- __u64 vaddr, __u64 size)
- {
- struct mm_iommu_table_group_mem_t *mem;
-+ struct tce_iommu_prereg *tcemem;
-+ bool found = false;
-
- if ((vaddr & ~PAGE_MASK) || (size & ~PAGE_MASK))
- return -EINVAL;
-
-- mem = mm_iommu_find(vaddr, size >> PAGE_SHIFT);
-+ mem = mm_iommu_find(container->mm, vaddr, size >> PAGE_SHIFT);
- if (!mem)
- return -ENOENT;
-
-- return mm_iommu_put(mem);
-+ list_for_each_entry(tcemem, &container->prereg_list, next) {
-+ if (tcemem->mem == mem) {
-+ found = true;
-+ break;
-+ }
-+ }
-+
-+ if (!found)
-+ return -ENOENT;
-+
-+ return tce_iommu_prereg_free(container, tcemem);
- }
-
- static long tce_iommu_register_pages(struct tce_container *container,
-@@ -122,22 +175,36 @@ static long tce_iommu_register_pages(struct tce_container *container,
- {
- long ret = 0;
- struct mm_iommu_table_group_mem_t *mem = NULL;
-+ struct tce_iommu_prereg *tcemem;
- unsigned long entries = size >> PAGE_SHIFT;
-
- if ((vaddr & ~PAGE_MASK) || (size & ~PAGE_MASK) ||
- ((vaddr + size) < vaddr))
- return -EINVAL;
-
-- ret = mm_iommu_get(vaddr, entries, &mem);
-+ mem = mm_iommu_find(container->mm, vaddr, entries);
-+ if (mem) {
-+ list_for_each_entry(tcemem, &container->prereg_list, next) {
-+ if (tcemem->mem == mem)
-+ return -EBUSY;
-+ }
-+ }
-+
-+ ret = mm_iommu_get(container->mm, vaddr, entries, &mem);
- if (ret)
- return ret;
-
-+ tcemem = kzalloc(sizeof(*tcemem), GFP_KERNEL);
-+ tcemem->mem = mem;
-+ list_add(&tcemem->next, &container->prereg_list);
-+
- container->enabled = true;
-
- return 0;
- }
-
--static long tce_iommu_userspace_view_alloc(struct iommu_table *tbl)
-+static long tce_iommu_userspace_view_alloc(struct iommu_table *tbl,
-+ struct mm_struct *mm)
- {
- unsigned long cb = _ALIGN_UP(sizeof(tbl->it_userspace[0]) *
- tbl->it_size, PAGE_SIZE);
-@@ -146,13 +213,13 @@ static long tce_iommu_userspace_view_alloc(struct iommu_table *tbl)
-
- BUG_ON(tbl->it_userspace);
-
-- ret = try_increment_locked_vm(cb >> PAGE_SHIFT);
-+ ret = try_increment_locked_vm(mm, cb >> PAGE_SHIFT);
- if (ret)
- return ret;
-
- uas = vzalloc(cb);
- if (!uas) {
-- decrement_locked_vm(cb >> PAGE_SHIFT);
-+ decrement_locked_vm(mm, cb >> PAGE_SHIFT);
- return -ENOMEM;
- }
- tbl->it_userspace = uas;
-@@ -160,7 +227,8 @@ static long tce_iommu_userspace_view_alloc(struct iommu_table *tbl)
- return 0;
- }
-
--static void tce_iommu_userspace_view_free(struct iommu_table *tbl)
-+static void tce_iommu_userspace_view_free(struct iommu_table *tbl,
-+ struct mm_struct *mm)
- {
- unsigned long cb = _ALIGN_UP(sizeof(tbl->it_userspace[0]) *
- tbl->it_size, PAGE_SIZE);
-@@ -170,7 +238,7 @@ static void tce_iommu_userspace_view_free(struct iommu_table *tbl)
-
- vfree(tbl->it_userspace);
- tbl->it_userspace = NULL;
-- decrement_locked_vm(cb >> PAGE_SHIFT);
-+ decrement_locked_vm(mm, cb >> PAGE_SHIFT);
- }
-
- static bool tce_page_is_contained(struct page *page, unsigned page_shift)
-@@ -230,9 +298,6 @@ static int tce_iommu_enable(struct tce_container *container)
- struct iommu_table_group *table_group;
- struct tce_iommu_group *tcegrp;
-
-- if (!current->mm)
-- return -ESRCH; /* process exited */
--
- if (container->enabled)
- return -EBUSY;
-
-@@ -277,8 +342,12 @@ static int tce_iommu_enable(struct tce_container *container)
- if (!table_group->tce32_size)
- return -EPERM;
-
-+ ret = tce_iommu_mm_set(container);
-+ if (ret)
-+ return ret;
-+
- locked = table_group->tce32_size >> PAGE_SHIFT;
-- ret = try_increment_locked_vm(locked);
-+ ret = try_increment_locked_vm(container->mm, locked);
- if (ret)
- return ret;
-
-@@ -296,10 +365,8 @@ static void tce_iommu_disable(struct tce_container *container)
-
- container->enabled = false;
-
-- if (!current->mm)
-- return;
--
-- decrement_locked_vm(container->locked_pages);
-+ BUG_ON(!container->mm);
-+ decrement_locked_vm(container->mm, container->locked_pages);
- }
-
- static void *tce_iommu_open(unsigned long arg)
-@@ -317,6 +384,7 @@ static void *tce_iommu_open(unsigned long arg)
-
- mutex_init(&container->lock);
- INIT_LIST_HEAD_RCU(&container->group_list);
-+ INIT_LIST_HEAD_RCU(&container->prereg_list);
-
- container->v2 = arg == VFIO_SPAPR_TCE_v2_IOMMU;
-
-@@ -326,7 +394,8 @@ static void *tce_iommu_open(unsigned long arg)
- static int tce_iommu_clear(struct tce_container *container,
- struct iommu_table *tbl,
- unsigned long entry, unsigned long pages);
--static void tce_iommu_free_table(struct iommu_table *tbl);
-+static void tce_iommu_free_table(struct tce_container *container,
-+ struct iommu_table *tbl);
-
- static void tce_iommu_release(void *iommu_data)
- {
-@@ -351,10 +420,20 @@ static void tce_iommu_release(void *iommu_data)
- continue;
-
- tce_iommu_clear(container, tbl, tbl->it_offset, tbl->it_size);
-- tce_iommu_free_table(tbl);
-+ tce_iommu_free_table(container, tbl);
-+ }
-+
-+ while (!list_empty(&container->prereg_list)) {
-+ struct tce_iommu_prereg *tcemem;
-+
-+ tcemem = list_first_entry(&container->prereg_list,
-+ struct tce_iommu_prereg, next);
-+ WARN_ON_ONCE(tce_iommu_prereg_free(container, tcemem));
- }
-
- tce_iommu_disable(container);
-+ if (container->mm)
-+ mmdrop(container->mm);
- mutex_destroy(&container->lock);
-
- kfree(container);
-@@ -369,13 +448,14 @@ static void tce_iommu_unuse_page(struct tce_container *container,
- put_page(page);
- }
-
--static int tce_iommu_prereg_ua_to_hpa(unsigned long tce, unsigned long size,
-+static int tce_iommu_prereg_ua_to_hpa(struct tce_container *container,
-+ unsigned long tce, unsigned long size,
- unsigned long *phpa, struct mm_iommu_table_group_mem_t **pmem)
- {
- long ret = 0;
- struct mm_iommu_table_group_mem_t *mem;
-
-- mem = mm_iommu_lookup(tce, size);
-+ mem = mm_iommu_lookup(container->mm, tce, size);
- if (!mem)
- return -EINVAL;
-
-@@ -388,18 +468,18 @@ static int tce_iommu_prereg_ua_to_hpa(unsigned long tce, unsigned long size,
- return 0;
- }
-
--static void tce_iommu_unuse_page_v2(struct iommu_table *tbl,
-- unsigned long entry)
-+static void tce_iommu_unuse_page_v2(struct tce_container *container,
-+ struct iommu_table *tbl, unsigned long entry)
- {
- struct mm_iommu_table_group_mem_t *mem = NULL;
- int ret;
- unsigned long hpa = 0;
- unsigned long *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
-
-- if (!pua || !current || !current->mm)
-+ if (!pua)
- return;
-
-- ret = tce_iommu_prereg_ua_to_hpa(*pua, IOMMU_PAGE_SIZE(tbl),
-+ ret = tce_iommu_prereg_ua_to_hpa(container, *pua, IOMMU_PAGE_SIZE(tbl),
- &hpa, &mem);
- if (ret)
- pr_debug("%s: tce %lx at #%lx was not cached, ret=%d\n",
-@@ -429,7 +509,7 @@ static int tce_iommu_clear(struct tce_container *container,
- continue;
-
- if (container->v2) {
-- tce_iommu_unuse_page_v2(tbl, entry);
-+ tce_iommu_unuse_page_v2(container, tbl, entry);
- continue;
- }
-
-@@ -509,13 +589,19 @@ static long tce_iommu_build_v2(struct tce_container *container,
- unsigned long hpa;
- enum dma_data_direction dirtmp;
-
-+ if (!tbl->it_userspace) {
-+ ret = tce_iommu_userspace_view_alloc(tbl, container->mm);
-+ if (ret)
-+ return ret;
-+ }
-+
- for (i = 0; i < pages; ++i) {
- struct mm_iommu_table_group_mem_t *mem = NULL;
- unsigned long *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl,
- entry + i);
-
-- ret = tce_iommu_prereg_ua_to_hpa(tce, IOMMU_PAGE_SIZE(tbl),
-- &hpa, &mem);
-+ ret = tce_iommu_prereg_ua_to_hpa(container,
-+ tce, IOMMU_PAGE_SIZE(tbl), &hpa, &mem);
- if (ret)
- break;
-
-@@ -536,7 +622,7 @@ static long tce_iommu_build_v2(struct tce_container *container,
- ret = iommu_tce_xchg(tbl, entry + i, &hpa, &dirtmp);
- if (ret) {
- /* dirtmp cannot be DMA_NONE here */
-- tce_iommu_unuse_page_v2(tbl, entry + i);
-+ tce_iommu_unuse_page_v2(container, tbl, entry + i);
- pr_err("iommu_tce: %s failed ioba=%lx, tce=%lx, ret=%ld\n",
- __func__, entry << tbl->it_page_shift,
- tce, ret);
-@@ -544,7 +630,7 @@ static long tce_iommu_build_v2(struct tce_container *container,
- }
-
- if (dirtmp != DMA_NONE)
-- tce_iommu_unuse_page_v2(tbl, entry + i);
-+ tce_iommu_unuse_page_v2(container, tbl, entry + i);
-
- *pua = tce;
-
-@@ -572,7 +658,7 @@ static long tce_iommu_create_table(struct tce_container *container,
- if (!table_size)
- return -EINVAL;
-
-- ret = try_increment_locked_vm(table_size >> PAGE_SHIFT);
-+ ret = try_increment_locked_vm(container->mm, table_size >> PAGE_SHIFT);
- if (ret)
- return ret;
-
-@@ -582,25 +668,17 @@ static long tce_iommu_create_table(struct tce_container *container,
- WARN_ON(!ret && !(*ptbl)->it_ops->free);
- WARN_ON(!ret && ((*ptbl)->it_allocated_size != table_size));
-
-- if (!ret && container->v2) {
-- ret = tce_iommu_userspace_view_alloc(*ptbl);
-- if (ret)
-- (*ptbl)->it_ops->free(*ptbl);
-- }
--
-- if (ret)
-- decrement_locked_vm(table_size >> PAGE_SHIFT);
--
- return ret;
- }
-
--static void tce_iommu_free_table(struct iommu_table *tbl)
-+static void tce_iommu_free_table(struct tce_container *container,
-+ struct iommu_table *tbl)
- {
- unsigned long pages = tbl->it_allocated_size >> PAGE_SHIFT;
-
-- tce_iommu_userspace_view_free(tbl);
-+ tce_iommu_userspace_view_free(tbl, container->mm);
- tbl->it_ops->free(tbl);
-- decrement_locked_vm(pages);
-+ decrement_locked_vm(container->mm, pages);
- }
-
- static long tce_iommu_create_window(struct tce_container *container,
-@@ -663,7 +741,7 @@ static long tce_iommu_create_window(struct tce_container *container,
- table_group = iommu_group_get_iommudata(tcegrp->grp);
- table_group->ops->unset_window(table_group, num);
- }
-- tce_iommu_free_table(tbl);
-+ tce_iommu_free_table(container, tbl);
-
- return ret;
- }
-@@ -701,12 +779,41 @@ static long tce_iommu_remove_window(struct tce_container *container,
-
- /* Free table */
- tce_iommu_clear(container, tbl, tbl->it_offset, tbl->it_size);
-- tce_iommu_free_table(tbl);
-+ tce_iommu_free_table(container, tbl);
- container->tables[num] = NULL;
-
- return 0;
- }
-
-+static long tce_iommu_create_default_window(struct tce_container *container)
-+{
-+ long ret;
-+ __u64 start_addr = 0;
-+ struct tce_iommu_group *tcegrp;
-+ struct iommu_table_group *table_group;
-+
-+ if (!container->def_window_pending)
-+ return 0;
-+
-+ if (!tce_groups_attached(container))
-+ return -ENODEV;
-+
-+ tcegrp = list_first_entry(&container->group_list,
-+ struct tce_iommu_group, next);
-+ table_group = iommu_group_get_iommudata(tcegrp->grp);
-+ if (!table_group)
-+ return -ENODEV;
-+
-+ ret = tce_iommu_create_window(container, IOMMU_PAGE_SHIFT_4K,
-+ table_group->tce32_size, 1, &start_addr);
-+ WARN_ON_ONCE(!ret && start_addr);
-+
-+ if (!ret)
-+ container->def_window_pending = false;
-+
-+ return ret;
-+}
-+
- static long tce_iommu_ioctl(void *iommu_data,
- unsigned int cmd, unsigned long arg)
- {
-@@ -727,7 +834,17 @@ static long tce_iommu_ioctl(void *iommu_data,
- }
-
- return (ret < 0) ? 0 : ret;
-+ }
-+
-+ /*
-+ * Sanity check to prevent one userspace from manipulating
-+ * another userspace mm.
-+ */
-+ BUG_ON(!container);
-+ if (container->mm && container->mm != current->mm)
-+ return -EPERM;
-
-+ switch (cmd) {
- case VFIO_IOMMU_SPAPR_TCE_GET_INFO: {
- struct vfio_iommu_spapr_tce_info info;
- struct tce_iommu_group *tcegrp;
-@@ -797,6 +914,10 @@ static long tce_iommu_ioctl(void *iommu_data,
- VFIO_DMA_MAP_FLAG_WRITE))
- return -EINVAL;
-
-+ ret = tce_iommu_create_default_window(container);
-+ if (ret)
-+ return ret;
-+
- num = tce_iommu_find_table(container, param.iova, &tbl);
- if (num < 0)
- return -ENXIO;
-@@ -860,6 +981,10 @@ static long tce_iommu_ioctl(void *iommu_data,
- if (param.flags)
- return -EINVAL;
-
-+ ret = tce_iommu_create_default_window(container);
-+ if (ret)
-+ return ret;
-+
- num = tce_iommu_find_table(container, param.iova, &tbl);
- if (num < 0)
- return -ENXIO;
-@@ -888,6 +1013,10 @@ static long tce_iommu_ioctl(void *iommu_data,
- minsz = offsetofend(struct vfio_iommu_spapr_register_memory,
- size);
-
-+ ret = tce_iommu_mm_set(container);
-+ if (ret)
-+ return ret;
-+
- if (copy_from_user(&param, (void __user *)arg, minsz))
- return -EFAULT;
-
-@@ -911,6 +1040,9 @@ static long tce_iommu_ioctl(void *iommu_data,
- if (!container->v2)
- break;
-
-+ if (!container->mm)
-+ return -EPERM;
-+
- minsz = offsetofend(struct vfio_iommu_spapr_register_memory,
- size);
-
-@@ -969,6 +1101,10 @@ static long tce_iommu_ioctl(void *iommu_data,
- if (!container->v2)
- break;
-
-+ ret = tce_iommu_mm_set(container);
-+ if (ret)
-+ return ret;
-+
- if (!tce_groups_attached(container))
- return -ENXIO;
-
-@@ -986,6 +1122,10 @@ static long tce_iommu_ioctl(void *iommu_data,
-
- mutex_lock(&container->lock);
-
-+ ret = tce_iommu_create_default_window(container);
-+ if (ret)
-+ return ret;
-+
- ret = tce_iommu_create_window(container, create.page_shift,
- create.window_size, create.levels,
- &create.start_addr);
-@@ -1003,6 +1143,10 @@ static long tce_iommu_ioctl(void *iommu_data,
- if (!container->v2)
- break;
-
-+ ret = tce_iommu_mm_set(container);
-+ if (ret)
-+ return ret;
-+
- if (!tce_groups_attached(container))
- return -ENXIO;
-
-@@ -1018,6 +1162,11 @@ static long tce_iommu_ioctl(void *iommu_data,
- if (remove.flags)
- return -EINVAL;
-
-+ if (container->def_window_pending && !remove.start_addr) {
-+ container->def_window_pending = false;
-+ return 0;
-+ }
-+
- mutex_lock(&container->lock);
-
- ret = tce_iommu_remove_window(container, remove.start_addr);
-@@ -1043,7 +1192,7 @@ static void tce_iommu_release_ownership(struct tce_container *container,
- continue;
-
- tce_iommu_clear(container, tbl, tbl->it_offset, tbl->it_size);
-- tce_iommu_userspace_view_free(tbl);
-+ tce_iommu_userspace_view_free(tbl, container->mm);
- if (tbl->it_map)
- iommu_release_ownership(tbl);
-
-@@ -1062,10 +1211,7 @@ static int tce_iommu_take_ownership(struct tce_container *container,
- if (!tbl || !tbl->it_map)
- continue;
-
-- rc = tce_iommu_userspace_view_alloc(tbl);
-- if (!rc)
-- rc = iommu_take_ownership(tbl);
--
-+ rc = iommu_take_ownership(tbl);
- if (rc) {
- for (j = 0; j < i; ++j)
- iommu_release_ownership(
-@@ -1100,9 +1246,6 @@ static void tce_iommu_release_ownership_ddw(struct tce_container *container,
- static long tce_iommu_take_ownership_ddw(struct tce_container *container,
- struct iommu_table_group *table_group)
- {
-- long i, ret = 0;
-- struct iommu_table *tbl = NULL;
--
- if (!table_group->ops->create_table || !table_group->ops->set_window ||
- !table_group->ops->release_ownership) {
- WARN_ON_ONCE(1);
-@@ -1111,47 +1254,7 @@ static long tce_iommu_take_ownership_ddw(struct tce_container *container,
-
- table_group->ops->take_ownership(table_group);
-
-- /*
-- * If it the first group attached, check if there is
-- * a default DMA window and create one if none as
-- * the userspace expects it to exist.
-- */
-- if (!tce_groups_attached(container) && !container->tables[0]) {
-- ret = tce_iommu_create_table(container,
-- table_group,
-- 0, /* window number */
-- IOMMU_PAGE_SHIFT_4K,
-- table_group->tce32_size,
-- 1, /* default levels */
-- &tbl);
-- if (ret)
-- goto release_exit;
-- else
-- container->tables[0] = tbl;
-- }
--
-- /* Set all windows to the new group */
-- for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) {
-- tbl = container->tables[i];
--
-- if (!tbl)
-- continue;
--
-- /* Set the default window to a new group */
-- ret = table_group->ops->set_window(table_group, i, tbl);
-- if (ret)
-- goto release_exit;
-- }
--
- return 0;
--
--release_exit:
-- for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i)
-- table_group->ops->unset_window(table_group, i);
--
-- table_group->ops->release_ownership(table_group);
--
-- return ret;
- }
-
- static int tce_iommu_attach_group(void *iommu_data,
-@@ -1203,10 +1306,13 @@ static int tce_iommu_attach_group(void *iommu_data,
- }
-
- if (!table_group->ops || !table_group->ops->take_ownership ||
-- !table_group->ops->release_ownership)
-+ !table_group->ops->release_ownership) {
- ret = tce_iommu_take_ownership(container, table_group);
-- else
-+ } else {
- ret = tce_iommu_take_ownership_ddw(container, table_group);
-+ if (!tce_groups_attached(container) && !container->tables[0])
-+ container->def_window_pending = true;
-+ }
-
- if (!ret) {
- tcegrp->grp = iommu_group;
-diff --git a/include/linux/bpf_verifier.h b/include/linux/bpf_verifier.h
-index 6aaf425..a13b031 100644
---- a/include/linux/bpf_verifier.h
-+++ b/include/linux/bpf_verifier.h
-@@ -18,19 +18,12 @@
-
- struct bpf_reg_state {
- enum bpf_reg_type type;
-- /*
-- * Used to determine if any memory access using this register will
-- * result in a bad access.
-- */
-- s64 min_value;
-- u64 max_value;
- union {
- /* valid when type == CONST_IMM | PTR_TO_STACK | UNKNOWN_VALUE */
- s64 imm;
-
- /* valid when type == PTR_TO_PACKET* */
- struct {
-- u32 id;
- u16 off;
- u16 range;
- };
-@@ -40,6 +33,13 @@ struct bpf_reg_state {
- */
- struct bpf_map *map_ptr;
- };
-+ u32 id;
-+ /* Used to determine if any memory access using this register will
-+ * result in a bad access. These two fields must be last.
-+ * See states_equal()
-+ */
-+ s64 min_value;
-+ u64 max_value;
- };
-
- enum bpf_stack_slot_type {
-diff --git a/include/linux/dccp.h b/include/linux/dccp.h
-index 61d042b..6844929 100644
---- a/include/linux/dccp.h
-+++ b/include/linux/dccp.h
-@@ -163,6 +163,7 @@ struct dccp_request_sock {
- __u64 dreq_isr;
- __u64 dreq_gsr;
- __be32 dreq_service;
-+ spinlock_t dreq_lock;
- struct list_head dreq_featneg;
- __u32 dreq_timestamp_echo;
- __u32 dreq_timestamp_time;
-diff --git a/include/linux/hyperv.h b/include/linux/hyperv.h
-index 192eef2f..d596a07 100644
---- a/include/linux/hyperv.h
-+++ b/include/linux/hyperv.h
-@@ -1548,31 +1548,23 @@ static inline struct vmpacket_descriptor *
- get_next_pkt_raw(struct vmbus_channel *channel)
- {
- struct hv_ring_buffer_info *ring_info = &channel->inbound;
-- u32 read_loc = ring_info->priv_read_index;
-+ u32 priv_read_loc = ring_info->priv_read_index;
- void *ring_buffer = hv_get_ring_buffer(ring_info);
-- struct vmpacket_descriptor *cur_desc;
-- u32 packetlen;
- u32 dsize = ring_info->ring_datasize;
-- u32 delta = read_loc - ring_info->ring_buffer->read_index;
-+ /*
-+ * delta is the difference between what is available to read and
-+ * what was already consumed in place. We commit read index after
-+ * the whole batch is processed.
-+ */
-+ u32 delta = priv_read_loc >= ring_info->ring_buffer->read_index ?
-+ priv_read_loc - ring_info->ring_buffer->read_index :
-+ (dsize - ring_info->ring_buffer->read_index) + priv_read_loc;
- u32 bytes_avail_toread = (hv_get_bytes_to_read(ring_info) - delta);
-
- if (bytes_avail_toread < sizeof(struct vmpacket_descriptor))
- return NULL;
-
-- if ((read_loc + sizeof(*cur_desc)) > dsize)
-- return NULL;
--
-- cur_desc = ring_buffer + read_loc;
-- packetlen = cur_desc->len8 << 3;
--
-- /*
-- * If the packet under consideration is wrapping around,
-- * return failure.
-- */
-- if ((read_loc + packetlen + VMBUS_PKT_TRAILER) > (dsize - 1))
-- return NULL;
--
-- return cur_desc;
-+ return ring_buffer + priv_read_loc;
- }
-
- /*
-@@ -1584,16 +1576,14 @@ static inline void put_pkt_raw(struct vmbus_channel *channel,
- struct vmpacket_descriptor *desc)
- {
- struct hv_ring_buffer_info *ring_info = &channel->inbound;
-- u32 read_loc = ring_info->priv_read_index;
- u32 packetlen = desc->len8 << 3;
- u32 dsize = ring_info->ring_datasize;
-
-- if ((read_loc + packetlen + VMBUS_PKT_TRAILER) > dsize)
-- BUG();
- /*
- * Include the packet trailer.
- */
- ring_info->priv_read_index += packetlen + VMBUS_PKT_TRAILER;
-+ ring_info->priv_read_index %= dsize;
- }
-
- /*
-diff --git a/include/uapi/linux/packet_diag.h b/include/uapi/linux/packet_diag.h
-index d08c63f..0c5d5dd 100644
---- a/include/uapi/linux/packet_diag.h
-+++ b/include/uapi/linux/packet_diag.h
-@@ -64,7 +64,7 @@ struct packet_diag_mclist {
- __u32 pdmc_count;
- __u16 pdmc_type;
- __u16 pdmc_alen;
-- __u8 pdmc_addr[MAX_ADDR_LEN];
-+ __u8 pdmc_addr[32]; /* MAX_ADDR_LEN */
- };
-
- struct packet_diag_ring {
-diff --git a/kernel/bpf/verifier.c b/kernel/bpf/verifier.c
-index 8199821..85d1c94 100644
---- a/kernel/bpf/verifier.c
-+++ b/kernel/bpf/verifier.c
-@@ -212,9 +212,10 @@ static void print_verifier_state(struct bpf_verifier_state *state)
- else if (t == CONST_PTR_TO_MAP || t == PTR_TO_MAP_VALUE ||
- t == PTR_TO_MAP_VALUE_OR_NULL ||
- t == PTR_TO_MAP_VALUE_ADJ)
-- verbose("(ks=%d,vs=%d)",
-+ verbose("(ks=%d,vs=%d,id=%u)",
- reg->map_ptr->key_size,
-- reg->map_ptr->value_size);
-+ reg->map_ptr->value_size,
-+ reg->id);
- if (reg->min_value != BPF_REGISTER_MIN_RANGE)
- verbose(",min_value=%lld",
- (long long)reg->min_value);
-@@ -443,13 +444,19 @@ static void init_reg_state(struct bpf_reg_state *regs)
- regs[BPF_REG_1].type = PTR_TO_CTX;
- }
-
--static void mark_reg_unknown_value(struct bpf_reg_state *regs, u32 regno)
-+static void __mark_reg_unknown_value(struct bpf_reg_state *regs, u32 regno)
- {
-- BUG_ON(regno >= MAX_BPF_REG);
- regs[regno].type = UNKNOWN_VALUE;
-+ regs[regno].id = 0;
- regs[regno].imm = 0;
- }
-
-+static void mark_reg_unknown_value(struct bpf_reg_state *regs, u32 regno)
-+{
-+ BUG_ON(regno >= MAX_BPF_REG);
-+ __mark_reg_unknown_value(regs, regno);
-+}
-+
- static void reset_reg_range_values(struct bpf_reg_state *regs, u32 regno)
- {
- regs[regno].min_value = BPF_REGISTER_MIN_RANGE;
-@@ -1252,6 +1259,7 @@ static int check_call(struct bpf_verifier_env *env, int func_id)
- return -EINVAL;
- }
- regs[BPF_REG_0].map_ptr = meta.map_ptr;
-+ regs[BPF_REG_0].id = ++env->id_gen;
- } else {
- verbose("unknown return type %d of func %d\n",
- fn->ret_type, func_id);
-@@ -1668,8 +1676,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn)
- insn->src_reg);
- return -EACCES;
- }
-- regs[insn->dst_reg].type = UNKNOWN_VALUE;
-- regs[insn->dst_reg].map_ptr = NULL;
-+ mark_reg_unknown_value(regs, insn->dst_reg);
- }
- } else {
- /* case: R = imm
-@@ -1931,6 +1938,43 @@ static void reg_set_min_max_inv(struct bpf_reg_state *true_reg,
- check_reg_overflow(true_reg);
- }
-
-+static void mark_map_reg(struct bpf_reg_state *regs, u32 regno, u32 id,
-+ enum bpf_reg_type type)
-+{
-+ struct bpf_reg_state *reg = &regs[regno];
-+
-+ if (reg->type == PTR_TO_MAP_VALUE_OR_NULL && reg->id == id) {
-+ reg->type = type;
-+ /* We don't need id from this point onwards anymore, thus we
-+ * should better reset it, so that state pruning has chances
-+ * to take effect.
-+ */
-+ reg->id = 0;
-+ if (type == UNKNOWN_VALUE)
-+ __mark_reg_unknown_value(regs, regno);
-+ }
-+}
-+
-+/* The logic is similar to find_good_pkt_pointers(), both could eventually
-+ * be folded together at some point.
-+ */
-+static void mark_map_regs(struct bpf_verifier_state *state, u32 regno,
-+ enum bpf_reg_type type)
-+{
-+ struct bpf_reg_state *regs = state->regs;
-+ u32 id = regs[regno].id;
-+ int i;
-+
-+ for (i = 0; i < MAX_BPF_REG; i++)
-+ mark_map_reg(regs, i, id, type);
-+
-+ for (i = 0; i < MAX_BPF_STACK; i += BPF_REG_SIZE) {
-+ if (state->stack_slot_type[i] != STACK_SPILL)
-+ continue;
-+ mark_map_reg(state->spilled_regs, i / BPF_REG_SIZE, id, type);
-+ }
-+}
-+
- static int check_cond_jmp_op(struct bpf_verifier_env *env,
- struct bpf_insn *insn, int *insn_idx)
- {
-@@ -2018,18 +2062,13 @@ static int check_cond_jmp_op(struct bpf_verifier_env *env,
- if (BPF_SRC(insn->code) == BPF_K &&
- insn->imm == 0 && (opcode == BPF_JEQ || opcode == BPF_JNE) &&
- dst_reg->type == PTR_TO_MAP_VALUE_OR_NULL) {
-- if (opcode == BPF_JEQ) {
-- /* next fallthrough insn can access memory via
-- * this register
-- */
-- regs[insn->dst_reg].type = PTR_TO_MAP_VALUE;
-- /* branch targer cannot access it, since reg == 0 */
-- mark_reg_unknown_value(other_branch->regs,
-- insn->dst_reg);
-- } else {
-- other_branch->regs[insn->dst_reg].type = PTR_TO_MAP_VALUE;
-- mark_reg_unknown_value(regs, insn->dst_reg);
-- }
-+ /* Mark all identical map registers in each branch as either
-+ * safe or unknown depending R == 0 or R != 0 conditional.
-+ */
-+ mark_map_regs(this_branch, insn->dst_reg,
-+ opcode == BPF_JEQ ? PTR_TO_MAP_VALUE : UNKNOWN_VALUE);
-+ mark_map_regs(other_branch, insn->dst_reg,
-+ opcode == BPF_JEQ ? UNKNOWN_VALUE : PTR_TO_MAP_VALUE);
- } else if (BPF_SRC(insn->code) == BPF_X && opcode == BPF_JGT &&
- dst_reg->type == PTR_TO_PACKET &&
- regs[insn->src_reg].type == PTR_TO_PACKET_END) {
-@@ -2469,7 +2508,7 @@ static bool states_equal(struct bpf_verifier_env *env,
- * we didn't do a variable access into a map then we are a-ok.
- */
- if (!varlen_map_access &&
-- rold->type == rcur->type && rold->imm == rcur->imm)
-+ memcmp(rold, rcur, offsetofend(struct bpf_reg_state, id)) == 0)
- continue;
-
- /* If we didn't map access then again we don't care about the
-diff --git a/kernel/futex.c b/kernel/futex.c
-index 38b68c2..4c6b6e6 100644
---- a/kernel/futex.c
-+++ b/kernel/futex.c
-@@ -2813,7 +2813,6 @@ static int futex_wait_requeue_pi(u32 __user *uaddr, unsigned int flags,
- {
- struct hrtimer_sleeper timeout, *to = NULL;
- struct rt_mutex_waiter rt_waiter;
-- struct rt_mutex *pi_mutex = NULL;
- struct futex_hash_bucket *hb;
- union futex_key key2 = FUTEX_KEY_INIT;
- struct futex_q q = futex_q_init;
-@@ -2897,6 +2896,8 @@ static int futex_wait_requeue_pi(u32 __user *uaddr, unsigned int flags,
- if (q.pi_state && (q.pi_state->owner != current)) {
- spin_lock(q.lock_ptr);
- ret = fixup_pi_state_owner(uaddr2, &q, current);
-+ if (ret && rt_mutex_owner(&q.pi_state->pi_mutex) == current)
-+ rt_mutex_unlock(&q.pi_state->pi_mutex);
- /*
- * Drop the reference to the pi state which
- * the requeue_pi() code acquired for us.
-@@ -2905,6 +2906,8 @@ static int futex_wait_requeue_pi(u32 __user *uaddr, unsigned int flags,
- spin_unlock(q.lock_ptr);
- }
- } else {
-+ struct rt_mutex *pi_mutex;
-+
- /*
- * We have been woken up by futex_unlock_pi(), a timeout, or a
- * signal. futex_unlock_pi() will not destroy the lock_ptr nor
-@@ -2928,18 +2931,19 @@ static int futex_wait_requeue_pi(u32 __user *uaddr, unsigned int flags,
- if (res)
- ret = (res < 0) ? res : 0;
-
-+ /*
-+ * If fixup_pi_state_owner() faulted and was unable to handle
-+ * the fault, unlock the rt_mutex and return the fault to
-+ * userspace.
-+ */
-+ if (ret && rt_mutex_owner(pi_mutex) == current)
-+ rt_mutex_unlock(pi_mutex);
-+
- /* Unqueue and drop the lock. */
- unqueue_me_pi(&q);
- }
-
-- /*
-- * If fixup_pi_state_owner() faulted and was unable to handle the
-- * fault, unlock the rt_mutex and return the fault to userspace.
-- */
-- if (ret == -EFAULT) {
-- if (pi_mutex && rt_mutex_owner(pi_mutex) == current)
-- rt_mutex_unlock(pi_mutex);
-- } else if (ret == -EINTR) {
-+ if (ret == -EINTR) {
- /*
- * We've already been requeued, but cannot restart by calling
- * futex_lock_pi() directly. We could restart this syscall, but
-diff --git a/kernel/locking/rwsem-spinlock.c b/kernel/locking/rwsem-spinlock.c
-index 1591f6b..2bef4ab 100644
---- a/kernel/locking/rwsem-spinlock.c
-+++ b/kernel/locking/rwsem-spinlock.c
-@@ -216,10 +216,8 @@ int __sched __down_write_common(struct rw_semaphore *sem, int state)
- */
- if (sem->count == 0)
- break;
-- if (signal_pending_state(state, current)) {
-- ret = -EINTR;
-- goto out;
-- }
-+ if (signal_pending_state(state, current))
-+ goto out_nolock;
- set_task_state(tsk, state);
- raw_spin_unlock_irqrestore(&sem->wait_lock, flags);
- schedule();
-@@ -227,12 +225,19 @@ int __sched __down_write_common(struct rw_semaphore *sem, int state)
- }
- /* got the lock */
- sem->count = -1;
--out:
- list_del(&waiter.list);
-
- raw_spin_unlock_irqrestore(&sem->wait_lock, flags);
-
- return ret;
-+
-+out_nolock:
-+ list_del(&waiter.list);
-+ if (!list_empty(&sem->wait_list))
-+ __rwsem_do_wake(sem, 1);
-+ raw_spin_unlock_irqrestore(&sem->wait_lock, flags);
-+
-+ return -EINTR;
- }
-
- void __sched __down_write(struct rw_semaphore *sem)
-diff --git a/mm/slab.c b/mm/slab.c
-index bd878f0..1f82d16 100644
---- a/mm/slab.c
-+++ b/mm/slab.c
-@@ -2332,7 +2332,7 @@ static int drain_freelist(struct kmem_cache *cache,
- return nr_freed;
- }
-
--int __kmem_cache_shrink(struct kmem_cache *cachep, bool deactivate)
-+int __kmem_cache_shrink(struct kmem_cache *cachep)
- {
- int ret = 0;
- int node;
-@@ -2352,7 +2352,7 @@ int __kmem_cache_shrink(struct kmem_cache *cachep, bool deactivate)
-
- int __kmem_cache_shutdown(struct kmem_cache *cachep)
- {
-- return __kmem_cache_shrink(cachep, false);
-+ return __kmem_cache_shrink(cachep);
- }
-
- void __kmem_cache_release(struct kmem_cache *cachep)
-diff --git a/mm/slab.h b/mm/slab.h
-index bc05fdc..ceb7d70 100644
---- a/mm/slab.h
-+++ b/mm/slab.h
-@@ -146,7 +146,7 @@ static inline unsigned long kmem_cache_flags(unsigned long object_size,
-
- int __kmem_cache_shutdown(struct kmem_cache *);
- void __kmem_cache_release(struct kmem_cache *);
--int __kmem_cache_shrink(struct kmem_cache *, bool);
-+int __kmem_cache_shrink(struct kmem_cache *);
- void slab_kmem_cache_release(struct kmem_cache *);
-
- struct seq_file;
-diff --git a/mm/slab_common.c b/mm/slab_common.c
-index 329b038..5d2f24f 100644
---- a/mm/slab_common.c
-+++ b/mm/slab_common.c
-@@ -573,6 +573,29 @@ void memcg_deactivate_kmem_caches(struct mem_cgroup *memcg)
- get_online_cpus();
- get_online_mems();
-
-+#ifdef CONFIG_SLUB
-+ /*
-+ * In case of SLUB, we need to disable empty slab caching to
-+ * avoid pinning the offline memory cgroup by freeable kmem
-+ * pages charged to it. SLAB doesn't need this, as it
-+ * periodically purges unused slabs.
-+ */
-+ mutex_lock(&slab_mutex);
-+ list_for_each_entry(s, &slab_caches, list) {
-+ c = is_root_cache(s) ? cache_from_memcg_idx(s, idx) : NULL;
-+ if (c) {
-+ c->cpu_partial = 0;
-+ c->min_partial = 0;
-+ }
-+ }
-+ mutex_unlock(&slab_mutex);
-+ /*
-+ * kmem_cache->cpu_partial is checked locklessly (see
-+ * put_cpu_partial()). Make sure the change is visible.
-+ */
-+ synchronize_sched();
-+#endif
-+
- mutex_lock(&slab_mutex);
- list_for_each_entry(s, &slab_caches, list) {
- if (!is_root_cache(s))
-@@ -584,7 +607,7 @@ void memcg_deactivate_kmem_caches(struct mem_cgroup *memcg)
- if (!c)
- continue;
-
-- __kmem_cache_shrink(c, true);
-+ __kmem_cache_shrink(c);
- arr->entries[idx] = NULL;
- }
- mutex_unlock(&slab_mutex);
-@@ -755,7 +778,7 @@ int kmem_cache_shrink(struct kmem_cache *cachep)
- get_online_cpus();
- get_online_mems();
- kasan_cache_shrink(cachep);
-- ret = __kmem_cache_shrink(cachep, false);
-+ ret = __kmem_cache_shrink(cachep);
- put_online_mems();
- put_online_cpus();
- return ret;
-diff --git a/mm/slob.c b/mm/slob.c
-index 5ec1580..eac04d43 100644
---- a/mm/slob.c
-+++ b/mm/slob.c
-@@ -634,7 +634,7 @@ void __kmem_cache_release(struct kmem_cache *c)
- {
- }
-
--int __kmem_cache_shrink(struct kmem_cache *d, bool deactivate)
-+int __kmem_cache_shrink(struct kmem_cache *d)
- {
- return 0;
- }
-diff --git a/mm/slub.c b/mm/slub.c
-index 7aa0e97..58c7526 100644
---- a/mm/slub.c
-+++ b/mm/slub.c
-@@ -3887,7 +3887,7 @@ EXPORT_SYMBOL(kfree);
- * being allocated from last increasing the chance that the last objects
- * are freed in them.
- */
--int __kmem_cache_shrink(struct kmem_cache *s, bool deactivate)
-+int __kmem_cache_shrink(struct kmem_cache *s)
- {
- int node;
- int i;
-@@ -3899,21 +3899,6 @@ int __kmem_cache_shrink(struct kmem_cache *s, bool deactivate)
- unsigned long flags;
- int ret = 0;
-
-- if (deactivate) {
-- /*
-- * Disable empty slabs caching. Used to avoid pinning offline
-- * memory cgroups by kmem pages that can be freed.
-- */
-- s->cpu_partial = 0;
-- s->min_partial = 0;
--
-- /*
-- * s->cpu_partial is checked locklessly (see put_cpu_partial),
-- * so we have to make sure the change is visible.
-- */
-- synchronize_sched();
-- }
--
- flush_all(s);
- for_each_kmem_cache_node(s, node, n) {
- INIT_LIST_HEAD(&discard);
-@@ -3970,7 +3955,7 @@ static int slab_mem_going_offline_callback(void *arg)
-
- mutex_lock(&slab_mutex);
- list_for_each_entry(s, &slab_caches, list)
-- __kmem_cache_shrink(s, false);
-+ __kmem_cache_shrink(s);
- mutex_unlock(&slab_mutex);
-
- return 0;
-diff --git a/net/bridge/br_forward.c b/net/bridge/br_forward.c
-index 7cb41ae..8498e35 100644
---- a/net/bridge/br_forward.c
-+++ b/net/bridge/br_forward.c
-@@ -186,8 +186,9 @@ void br_flood(struct net_bridge *br, struct sk_buff *skb,
- /* Do not flood unicast traffic to ports that turn it off */
- if (pkt_type == BR_PKT_UNICAST && !(p->flags & BR_FLOOD))
- continue;
-+ /* Do not flood if mc off, except for traffic we originate */
- if (pkt_type == BR_PKT_MULTICAST &&
-- !(p->flags & BR_MCAST_FLOOD))
-+ !(p->flags & BR_MCAST_FLOOD) && skb->dev != br->dev)
- continue;
-
- /* Do not flood to ports that enable proxy ARP */
-diff --git a/net/bridge/br_input.c b/net/bridge/br_input.c
-index 855b72f..267b46a 100644
---- a/net/bridge/br_input.c
-+++ b/net/bridge/br_input.c
-@@ -29,6 +29,7 @@ EXPORT_SYMBOL(br_should_route_hook);
- static int
- br_netif_receive_skb(struct net *net, struct sock *sk, struct sk_buff *skb)
- {
-+ br_drop_fake_rtable(skb);
- return netif_receive_skb(skb);
- }
-
-diff --git a/net/bridge/br_netfilter_hooks.c b/net/bridge/br_netfilter_hooks.c
-index 7fbdbae..aa1df1a 100644
---- a/net/bridge/br_netfilter_hooks.c
-+++ b/net/bridge/br_netfilter_hooks.c
-@@ -521,21 +521,6 @@ static unsigned int br_nf_pre_routing(void *priv,
- }
-
-
--/* PF_BRIDGE/LOCAL_IN ************************************************/
--/* The packet is locally destined, which requires a real
-- * dst_entry, so detach the fake one. On the way up, the
-- * packet would pass through PRE_ROUTING again (which already
-- * took place when the packet entered the bridge), but we
-- * register an IPv4 PRE_ROUTING 'sabotage' hook that will
-- * prevent this from happening. */
--static unsigned int br_nf_local_in(void *priv,
-- struct sk_buff *skb,
-- const struct nf_hook_state *state)
--{
-- br_drop_fake_rtable(skb);
-- return NF_ACCEPT;
--}
--
- /* PF_BRIDGE/FORWARD *************************************************/
- static int br_nf_forward_finish(struct net *net, struct sock *sk, struct sk_buff *skb)
- {
-@@ -906,12 +891,6 @@ static struct nf_hook_ops br_nf_ops[] __read_mostly = {
- .priority = NF_BR_PRI_BRNF,
- },
- {
-- .hook = br_nf_local_in,
-- .pf = NFPROTO_BRIDGE,
-- .hooknum = NF_BR_LOCAL_IN,
-- .priority = NF_BR_PRI_BRNF,
-- },
-- {
- .hook = br_nf_forward_ip,
- .pf = NFPROTO_BRIDGE,
- .hooknum = NF_BR_FORWARD,
-diff --git a/net/core/dev.c b/net/core/dev.c
-index 60b0a604..2e04fd1 100644
---- a/net/core/dev.c
-+++ b/net/core/dev.c
-@@ -1697,27 +1697,54 @@ EXPORT_SYMBOL_GPL(net_dec_egress_queue);
- static struct static_key netstamp_needed __read_mostly;
- #ifdef HAVE_JUMP_LABEL
- static atomic_t netstamp_needed_deferred;
-+static atomic_t netstamp_wanted;
- static void netstamp_clear(struct work_struct *work)
- {
- int deferred = atomic_xchg(&netstamp_needed_deferred, 0);
-+ int wanted;
-
-- while (deferred--)
-- static_key_slow_dec(&netstamp_needed);
-+ wanted = atomic_add_return(deferred, &netstamp_wanted);
-+ if (wanted > 0)
-+ static_key_enable(&netstamp_needed);
-+ else
-+ static_key_disable(&netstamp_needed);
- }
- static DECLARE_WORK(netstamp_work, netstamp_clear);
- #endif
-
- void net_enable_timestamp(void)
- {
-+#ifdef HAVE_JUMP_LABEL
-+ int wanted;
-+
-+ while (1) {
-+ wanted = atomic_read(&netstamp_wanted);
-+ if (wanted <= 0)
-+ break;
-+ if (atomic_cmpxchg(&netstamp_wanted, wanted, wanted + 1) == wanted)
-+ return;
-+ }
-+ atomic_inc(&netstamp_needed_deferred);
-+ schedule_work(&netstamp_work);
-+#else
- static_key_slow_inc(&netstamp_needed);
-+#endif
- }
- EXPORT_SYMBOL(net_enable_timestamp);
-
- void net_disable_timestamp(void)
- {
- #ifdef HAVE_JUMP_LABEL
-- /* net_disable_timestamp() can be called from non process context */
-- atomic_inc(&netstamp_needed_deferred);
-+ int wanted;
-+
-+ while (1) {
-+ wanted = atomic_read(&netstamp_wanted);
-+ if (wanted <= 1)
-+ break;
-+ if (atomic_cmpxchg(&netstamp_wanted, wanted, wanted - 1) == wanted)
-+ return;
-+ }
-+ atomic_dec(&netstamp_needed_deferred);
- schedule_work(&netstamp_work);
- #else
- static_key_slow_dec(&netstamp_needed);
-diff --git a/net/core/skbuff.c b/net/core/skbuff.c
-index 1e3e008..f0f462c 100644
---- a/net/core/skbuff.c
-+++ b/net/core/skbuff.c
-@@ -3814,13 +3814,14 @@ void skb_complete_tx_timestamp(struct sk_buff *skb,
- if (!skb_may_tx_timestamp(sk, false))
- return;
-
-- /* take a reference to prevent skb_orphan() from freeing the socket */
-- sock_hold(sk);
--
-- *skb_hwtstamps(skb) = *hwtstamps;
-- __skb_complete_tx_timestamp(skb, sk, SCM_TSTAMP_SND);
--
-- sock_put(sk);
-+ /* Take a reference to prevent skb_orphan() from freeing the socket,
-+ * but only if the socket refcount is not zero.
-+ */
-+ if (likely(atomic_inc_not_zero(&sk->sk_refcnt))) {
-+ *skb_hwtstamps(skb) = *hwtstamps;
-+ __skb_complete_tx_timestamp(skb, sk, SCM_TSTAMP_SND);
-+ sock_put(sk);
-+ }
- }
- EXPORT_SYMBOL_GPL(skb_complete_tx_timestamp);
-
-@@ -3871,7 +3872,7 @@ void skb_complete_wifi_ack(struct sk_buff *skb, bool acked)
- {
- struct sock *sk = skb->sk;
- struct sock_exterr_skb *serr;
-- int err;
-+ int err = 1;
-
- skb->wifi_acked_valid = 1;
- skb->wifi_acked = acked;
-@@ -3881,14 +3882,15 @@ void skb_complete_wifi_ack(struct sk_buff *skb, bool acked)
- serr->ee.ee_errno = ENOMSG;
- serr->ee.ee_origin = SO_EE_ORIGIN_TXSTATUS;
-
-- /* take a reference to prevent skb_orphan() from freeing the socket */
-- sock_hold(sk);
--
-- err = sock_queue_err_skb(sk, skb);
-+ /* Take a reference to prevent skb_orphan() from freeing the socket,
-+ * but only if the socket refcount is not zero.
-+ */
-+ if (likely(atomic_inc_not_zero(&sk->sk_refcnt))) {
-+ err = sock_queue_err_skb(sk, skb);
-+ sock_put(sk);
-+ }
- if (err)
- kfree_skb(skb);
--
-- sock_put(sk);
- }
- EXPORT_SYMBOL_GPL(skb_complete_wifi_ack);
-
-diff --git a/net/dccp/ccids/ccid2.c b/net/dccp/ccids/ccid2.c
-index f053198..5e3a730 100644
---- a/net/dccp/ccids/ccid2.c
-+++ b/net/dccp/ccids/ccid2.c
-@@ -749,6 +749,7 @@ static void ccid2_hc_tx_exit(struct sock *sk)
- for (i = 0; i < hc->tx_seqbufc; i++)
- kfree(hc->tx_seqbuf[i]);
- hc->tx_seqbufc = 0;
-+ dccp_ackvec_parsed_cleanup(&hc->tx_av_chunks);
- }
-
- static void ccid2_hc_rx_packet_recv(struct sock *sk, struct sk_buff *skb)
-diff --git a/net/dccp/input.c b/net/dccp/input.c
-index 8fedc2d..4a05d78 100644
---- a/net/dccp/input.c
-+++ b/net/dccp/input.c
-@@ -577,6 +577,7 @@ int dccp_rcv_state_process(struct sock *sk, struct sk_buff *skb,
- struct dccp_sock *dp = dccp_sk(sk);
- struct dccp_skb_cb *dcb = DCCP_SKB_CB(skb);
- const int old_state = sk->sk_state;
-+ bool acceptable;
- int queued = 0;
-
- /*
-@@ -603,8 +604,13 @@ int dccp_rcv_state_process(struct sock *sk, struct sk_buff *skb,
- */
- if (sk->sk_state == DCCP_LISTEN) {
- if (dh->dccph_type == DCCP_PKT_REQUEST) {
-- if (inet_csk(sk)->icsk_af_ops->conn_request(sk,
-- skb) < 0)
-+ /* It is possible that we process SYN packets from backlog,
-+ * so we need to make sure to disable BH right there.
-+ */
-+ local_bh_disable();
-+ acceptable = inet_csk(sk)->icsk_af_ops->conn_request(sk, skb) >= 0;
-+ local_bh_enable();
-+ if (!acceptable)
- return 1;
- consume_skb(skb);
- return 0;
-diff --git a/net/dccp/ipv4.c b/net/dccp/ipv4.c
-index edbe59d..86b0933 100644
---- a/net/dccp/ipv4.c
-+++ b/net/dccp/ipv4.c
-@@ -289,7 +289,8 @@ static void dccp_v4_err(struct sk_buff *skb, u32 info)
-
- switch (type) {
- case ICMP_REDIRECT:
-- dccp_do_redirect(skb, sk);
-+ if (!sock_owned_by_user(sk))
-+ dccp_do_redirect(skb, sk);
- goto out;
- case ICMP_SOURCE_QUENCH:
- /* Just silently ignore these. */
-diff --git a/net/dccp/ipv6.c b/net/dccp/ipv6.c
-index 7506c03..237d62c 100644
---- a/net/dccp/ipv6.c
-+++ b/net/dccp/ipv6.c
-@@ -122,10 +122,12 @@ static void dccp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt,
- np = inet6_sk(sk);
-
- if (type == NDISC_REDIRECT) {
-- struct dst_entry *dst = __sk_dst_check(sk, np->dst_cookie);
-+ if (!sock_owned_by_user(sk)) {
-+ struct dst_entry *dst = __sk_dst_check(sk, np->dst_cookie);
-
-- if (dst)
-- dst->ops->redirect(dst, sk, skb);
-+ if (dst)
-+ dst->ops->redirect(dst, sk, skb);
-+ }
- goto out;
- }
-
-diff --git a/net/dccp/minisocks.c b/net/dccp/minisocks.c
-index 53eddf9..39e7e2b 100644
---- a/net/dccp/minisocks.c
-+++ b/net/dccp/minisocks.c
-@@ -122,6 +122,7 @@ struct sock *dccp_create_openreq_child(const struct sock *sk,
- /* It is still raw copy of parent, so invalidate
- * destructor and make plain sk_free() */
- newsk->sk_destruct = NULL;
-+ bh_unlock_sock(newsk);
- sk_free(newsk);
- return NULL;
- }
-@@ -145,6 +146,13 @@ struct sock *dccp_check_req(struct sock *sk, struct sk_buff *skb,
- struct dccp_request_sock *dreq = dccp_rsk(req);
- bool own_req;
-
-+ /* TCP/DCCP listeners became lockless.
-+ * DCCP stores complex state in its request_sock, so we need
-+ * a protection for them, now this code runs without being protected
-+ * by the parent (listener) lock.
-+ */
-+ spin_lock_bh(&dreq->dreq_lock);
-+
- /* Check for retransmitted REQUEST */
- if (dccp_hdr(skb)->dccph_type == DCCP_PKT_REQUEST) {
-
-@@ -159,7 +167,7 @@ struct sock *dccp_check_req(struct sock *sk, struct sk_buff *skb,
- inet_rtx_syn_ack(sk, req);
- }
- /* Network Duplicate, discard packet */
-- return NULL;
-+ goto out;
- }
-
- DCCP_SKB_CB(skb)->dccpd_reset_code = DCCP_RESET_CODE_PACKET_ERROR;
-@@ -185,20 +193,20 @@ struct sock *dccp_check_req(struct sock *sk, struct sk_buff *skb,
-
- child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL,
- req, &own_req);
-- if (!child)
-- goto listen_overflow;
--
-- return inet_csk_complete_hashdance(sk, child, req, own_req);
-+ if (child) {
-+ child = inet_csk_complete_hashdance(sk, child, req, own_req);
-+ goto out;
-+ }
-
--listen_overflow:
-- dccp_pr_debug("listen_overflow!\n");
- DCCP_SKB_CB(skb)->dccpd_reset_code = DCCP_RESET_CODE_TOO_BUSY;
- drop:
- if (dccp_hdr(skb)->dccph_type != DCCP_PKT_RESET)
- req->rsk_ops->send_reset(sk, skb);
-
- inet_csk_reqsk_queue_drop(sk, req);
-- return NULL;
-+out:
-+ spin_unlock_bh(&dreq->dreq_lock);
-+ return child;
- }
-
- EXPORT_SYMBOL_GPL(dccp_check_req);
-@@ -249,6 +257,7 @@ int dccp_reqsk_init(struct request_sock *req,
- {
- struct dccp_request_sock *dreq = dccp_rsk(req);
-
-+ spin_lock_init(&dreq->dreq_lock);
- inet_rsk(req)->ir_rmt_port = dccp_hdr(skb)->dccph_sport;
- inet_rsk(req)->ir_num = ntohs(dccp_hdr(skb)->dccph_dport);
- inet_rsk(req)->acked = 0;
-diff --git a/net/ipv4/af_inet.c b/net/ipv4/af_inet.c
-index 21514324..971b947 100644
---- a/net/ipv4/af_inet.c
-+++ b/net/ipv4/af_inet.c
-@@ -1460,8 +1460,10 @@ int inet_gro_complete(struct sk_buff *skb, int nhoff)
- int proto = iph->protocol;
- int err = -ENOSYS;
-
-- if (skb->encapsulation)
-+ if (skb->encapsulation) {
-+ skb_set_inner_protocol(skb, cpu_to_be16(ETH_P_IP));
- skb_set_inner_network_header(skb, nhoff);
-+ }
-
- csum_replace2(&iph->check, iph->tot_len, newlen);
- iph->tot_len = newlen;
-diff --git a/net/ipv4/route.c b/net/ipv4/route.c
-index d851cae..17e6fbf 100644
---- a/net/ipv4/route.c
-+++ b/net/ipv4/route.c
-@@ -1968,6 +1968,7 @@ int ip_route_input_noref(struct sk_buff *skb, __be32 daddr, __be32 saddr,
- {
- int res;
-
-+ tos &= IPTOS_RT_MASK;
- rcu_read_lock();
-
- /* Multicast recognition logic is moved from route cache to here.
-diff --git a/net/ipv4/tcp_input.c b/net/ipv4/tcp_input.c
-index c71d49c..ce42ded 100644
---- a/net/ipv4/tcp_input.c
-+++ b/net/ipv4/tcp_input.c
-@@ -5916,9 +5916,15 @@ int tcp_rcv_state_process(struct sock *sk, struct sk_buff *skb)
- if (th->syn) {
- if (th->fin)
- goto discard;
-- if (icsk->icsk_af_ops->conn_request(sk, skb) < 0)
-- return 1;
-+ /* It is possible that we process SYN packets from backlog,
-+ * so we need to make sure to disable BH right there.
-+ */
-+ local_bh_disable();
-+ acceptable = icsk->icsk_af_ops->conn_request(sk, skb) >= 0;
-+ local_bh_enable();
-
-+ if (!acceptable)
-+ return 1;
- consume_skb(skb);
- return 0;
- }
-diff --git a/net/ipv4/tcp_ipv4.c b/net/ipv4/tcp_ipv4.c
-index 2259114..6988566 100644
---- a/net/ipv4/tcp_ipv4.c
-+++ b/net/ipv4/tcp_ipv4.c
-@@ -269,10 +269,13 @@ EXPORT_SYMBOL(tcp_v4_connect);
- */
- void tcp_v4_mtu_reduced(struct sock *sk)
- {
-- struct dst_entry *dst;
- struct inet_sock *inet = inet_sk(sk);
-- u32 mtu = tcp_sk(sk)->mtu_info;
-+ struct dst_entry *dst;
-+ u32 mtu;
-
-+ if ((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_CLOSE))
-+ return;
-+ mtu = tcp_sk(sk)->mtu_info;
- dst = inet_csk_update_pmtu(sk, mtu);
- if (!dst)
- return;
-@@ -418,7 +421,8 @@ void tcp_v4_err(struct sk_buff *icmp_skb, u32 info)
-
- switch (type) {
- case ICMP_REDIRECT:
-- do_redirect(icmp_skb, sk);
-+ if (!sock_owned_by_user(sk))
-+ do_redirect(icmp_skb, sk);
- goto out;
- case ICMP_SOURCE_QUENCH:
- /* Just silently ignore these. */
-diff --git a/net/ipv4/tcp_timer.c b/net/ipv4/tcp_timer.c
-index 3ea1cf8..b1e65b3 100644
---- a/net/ipv4/tcp_timer.c
-+++ b/net/ipv4/tcp_timer.c
-@@ -249,7 +249,8 @@ void tcp_delack_timer_handler(struct sock *sk)
-
- sk_mem_reclaim_partial(sk);
-
-- if (sk->sk_state == TCP_CLOSE || !(icsk->icsk_ack.pending & ICSK_ACK_TIMER))
-+ if (((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN)) ||
-+ !(icsk->icsk_ack.pending & ICSK_ACK_TIMER))
- goto out;
-
- if (time_after(icsk->icsk_ack.timeout, jiffies)) {
-@@ -552,7 +553,8 @@ void tcp_write_timer_handler(struct sock *sk)
- struct inet_connection_sock *icsk = inet_csk(sk);
- int event;
-
-- if (sk->sk_state == TCP_CLOSE || !icsk->icsk_pending)
-+ if (((1 << sk->sk_state) & (TCPF_CLOSE | TCPF_LISTEN)) ||
-+ !icsk->icsk_pending)
- goto out;
-
- if (time_after(icsk->icsk_timeout, jiffies)) {
-diff --git a/net/ipv6/ip6_fib.c b/net/ipv6/ip6_fib.c
-index ef54852..8c88a37 100644
---- a/net/ipv6/ip6_fib.c
-+++ b/net/ipv6/ip6_fib.c
-@@ -908,6 +908,8 @@ static int fib6_add_rt2node(struct fib6_node *fn, struct rt6_info *rt,
- ins = &rt->dst.rt6_next;
- iter = *ins;
- while (iter) {
-+ if (iter->rt6i_metric > rt->rt6i_metric)
-+ break;
- if (rt6_qualify_for_ecmp(iter)) {
- *ins = iter->dst.rt6_next;
- fib6_purge_rt(iter, fn, info->nl_net);
-diff --git a/net/ipv6/ip6_offload.c b/net/ipv6/ip6_offload.c
-index fc7b401..33b04ec 100644
---- a/net/ipv6/ip6_offload.c
-+++ b/net/ipv6/ip6_offload.c
-@@ -294,8 +294,10 @@ static int ipv6_gro_complete(struct sk_buff *skb, int nhoff)
- struct ipv6hdr *iph = (struct ipv6hdr *)(skb->data + nhoff);
- int err = -ENOSYS;
-
-- if (skb->encapsulation)
-+ if (skb->encapsulation) {
-+ skb_set_inner_protocol(skb, cpu_to_be16(ETH_P_IPV6));
- skb_set_inner_network_header(skb, nhoff);
-+ }
-
- iph->payload_len = htons(skb->len - nhoff - sizeof(*iph));
-
-diff --git a/net/ipv6/ip6_output.c b/net/ipv6/ip6_output.c
-index 9a87bfb..e27b8fd 100644
---- a/net/ipv6/ip6_output.c
-+++ b/net/ipv6/ip6_output.c
-@@ -757,13 +757,14 @@ int ip6_fragment(struct net *net, struct sock *sk, struct sk_buff *skb,
- * Fragment the datagram.
- */
-
-- *prevhdr = NEXTHDR_FRAGMENT;
- troom = rt->dst.dev->needed_tailroom;
-
- /*
- * Keep copying data until we run out.
- */
- while (left > 0) {
-+ u8 *fragnexthdr_offset;
-+
- len = left;
- /* IF: it doesn't fit, use 'mtu' - the data space left */
- if (len > mtu)
-@@ -808,6 +809,10 @@ int ip6_fragment(struct net *net, struct sock *sk, struct sk_buff *skb,
- */
- skb_copy_from_linear_data(skb, skb_network_header(frag), hlen);
-
-+ fragnexthdr_offset = skb_network_header(frag);
-+ fragnexthdr_offset += prevhdr - skb_network_header(skb);
-+ *fragnexthdr_offset = NEXTHDR_FRAGMENT;
-+
- /*
- * Build fragment header.
- */
-diff --git a/net/ipv6/ip6_vti.c b/net/ipv6/ip6_vti.c
-index c299c1e..66c2b4b 100644
---- a/net/ipv6/ip6_vti.c
-+++ b/net/ipv6/ip6_vti.c
-@@ -691,6 +691,10 @@ vti6_parm_to_user(struct ip6_tnl_parm2 *u, const struct __ip6_tnl_parm *p)
- u->link = p->link;
- u->i_key = p->i_key;
- u->o_key = p->o_key;
-+ if (u->i_key)
-+ u->i_flags |= GRE_KEY;
-+ if (u->o_key)
-+ u->o_flags |= GRE_KEY;
- u->proto = p->proto;
-
- memcpy(u->name, p->name, sizeof(u->name));
-diff --git a/net/ipv6/netfilter/nf_conntrack_reasm.c b/net/ipv6/netfilter/nf_conntrack_reasm.c
-index 9948b5c..986d4ca 100644
---- a/net/ipv6/netfilter/nf_conntrack_reasm.c
-+++ b/net/ipv6/netfilter/nf_conntrack_reasm.c
-@@ -589,6 +589,7 @@ int nf_ct_frag6_gather(struct net *net, struct sk_buff *skb, u32 user)
- hdr = ipv6_hdr(skb);
- fhdr = (struct frag_hdr *)skb_transport_header(skb);
-
-+ skb_orphan(skb);
- fq = fq_find(net, fhdr->identification, user, &hdr->saddr, &hdr->daddr,
- skb->dev ? skb->dev->ifindex : 0, ip6_frag_ecn(hdr));
- if (fq == NULL) {
-diff --git a/net/ipv6/tcp_ipv6.c b/net/ipv6/tcp_ipv6.c
-index 6673965..b2e61a0 100644
---- a/net/ipv6/tcp_ipv6.c
-+++ b/net/ipv6/tcp_ipv6.c
-@@ -375,10 +375,12 @@ static void tcp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt,
- np = inet6_sk(sk);
-
- if (type == NDISC_REDIRECT) {
-- struct dst_entry *dst = __sk_dst_check(sk, np->dst_cookie);
-+ if (!sock_owned_by_user(sk)) {
-+ struct dst_entry *dst = __sk_dst_check(sk, np->dst_cookie);
-
-- if (dst)
-- dst->ops->redirect(dst, sk, skb);
-+ if (dst)
-+ dst->ops->redirect(dst, sk, skb);
-+ }
- goto out;
- }
-
-diff --git a/net/l2tp/l2tp_ip.c b/net/l2tp/l2tp_ip.c
-index c0f0750..ff750bb 100644
---- a/net/l2tp/l2tp_ip.c
-+++ b/net/l2tp/l2tp_ip.c
-@@ -388,7 +388,7 @@ static int l2tp_ip_backlog_recv(struct sock *sk, struct sk_buff *skb)
- drop:
- IP_INC_STATS(sock_net(sk), IPSTATS_MIB_INDISCARDS);
- kfree_skb(skb);
-- return -1;
-+ return 0;
- }
-
- /* Userspace will call sendmsg() on the tunnel socket to send L2TP
-diff --git a/net/mpls/af_mpls.c b/net/mpls/af_mpls.c
-index 5b77377..1309e2c 100644
---- a/net/mpls/af_mpls.c
-+++ b/net/mpls/af_mpls.c
-@@ -956,7 +956,8 @@ static void mpls_ifdown(struct net_device *dev, int event)
- /* fall through */
- case NETDEV_CHANGE:
- nh->nh_flags |= RTNH_F_LINKDOWN;
-- ACCESS_ONCE(rt->rt_nhn_alive) = rt->rt_nhn_alive - 1;
-+ if (event != NETDEV_UNREGISTER)
-+ ACCESS_ONCE(rt->rt_nhn_alive) = rt->rt_nhn_alive - 1;
- break;
- }
- if (event == NETDEV_UNREGISTER)
-@@ -1696,6 +1697,7 @@ static void mpls_net_exit(struct net *net)
- for (index = 0; index < platform_labels; index++) {
- struct mpls_route *rt = rtnl_dereference(platform_label[index]);
- RCU_INIT_POINTER(platform_label[index], NULL);
-+ mpls_notify_route(net, index, rt, NULL, NULL);
- mpls_rt_free(rt);
- }
- rtnl_unlock();
-diff --git a/net/openvswitch/conntrack.c b/net/openvswitch/conntrack.c
-index eab210b..48386bf 100644
---- a/net/openvswitch/conntrack.c
-+++ b/net/openvswitch/conntrack.c
-@@ -367,7 +367,6 @@ static int handle_fragments(struct net *net, struct sw_flow_key *key,
- } else if (key->eth.type == htons(ETH_P_IPV6)) {
- enum ip6_defrag_users user = IP6_DEFRAG_CONNTRACK_IN + zone;
-
-- skb_orphan(skb);
- memset(IP6CB(skb), 0, sizeof(struct inet6_skb_parm));
- err = nf_ct_frag6_gather(net, skb, user);
- if (err) {
-diff --git a/net/packet/af_packet.c b/net/packet/af_packet.c
-index 34de326..f2b04a7 100644
---- a/net/packet/af_packet.c
-+++ b/net/packet/af_packet.c
-@@ -3140,7 +3140,7 @@ static int packet_bind_spkt(struct socket *sock, struct sockaddr *uaddr,
- int addr_len)
- {
- struct sock *sk = sock->sk;
-- char name[15];
-+ char name[sizeof(uaddr->sa_data) + 1];
-
- /*
- * Check legality
-@@ -3148,7 +3148,11 @@ static int packet_bind_spkt(struct socket *sock, struct sockaddr *uaddr,
-
- if (addr_len != sizeof(struct sockaddr))
- return -EINVAL;
-- strlcpy(name, uaddr->sa_data, sizeof(name));
-+ /* uaddr->sa_data comes from the userspace, it's not guaranteed to be
-+ * zero-terminated.
-+ */
-+ memcpy(name, uaddr->sa_data, sizeof(uaddr->sa_data));
-+ name[sizeof(uaddr->sa_data)] = 0;
-
- return packet_do_bind(sk, name, 0, pkt_sk(sk)->num);
- }
-diff --git a/net/sched/act_api.c b/net/sched/act_api.c
-index c6c2a93..c651cfc 100644
---- a/net/sched/act_api.c
-+++ b/net/sched/act_api.c
-@@ -820,10 +820,8 @@ static int tca_action_flush(struct net *net, struct nlattr *nla,
- goto out_module_put;
-
- err = ops->walk(net, skb, &dcb, RTM_DELACTION, ops);
-- if (err < 0)
-+ if (err <= 0)
- goto out_module_put;
-- if (err == 0)
-- goto noflush_out;
-
- nla_nest_end(skb, nest);
-
-@@ -840,7 +838,6 @@ static int tca_action_flush(struct net *net, struct nlattr *nla,
- out_module_put:
- module_put(ops->owner);
- err_out:
--noflush_out:
- kfree_skb(skb);
- return err;
- }
-diff --git a/net/sched/act_connmark.c b/net/sched/act_connmark.c
-index eae07a2..1191179 100644
---- a/net/sched/act_connmark.c
-+++ b/net/sched/act_connmark.c
-@@ -113,6 +113,9 @@ static int tcf_connmark_init(struct net *net, struct nlattr *nla,
- if (ret < 0)
- return ret;
-
-+ if (!tb[TCA_CONNMARK_PARMS])
-+ return -EINVAL;
-+
- parm = nla_data(tb[TCA_CONNMARK_PARMS]);
-
- if (!tcf_hash_check(tn, parm->index, a, bind)) {
-diff --git a/net/sched/act_skbmod.c b/net/sched/act_skbmod.c
-index e7d9638..f85313d 100644
---- a/net/sched/act_skbmod.c
-+++ b/net/sched/act_skbmod.c
-@@ -228,7 +228,6 @@ static int tcf_skbmod_dump(struct sk_buff *skb, struct tc_action *a,
-
- return skb->len;
- nla_put_failure:
-- rcu_read_unlock();
- nlmsg_trim(skb, b);
- return -1;
- }
-diff --git a/net/strparser/strparser.c b/net/strparser/strparser.c
-index 41adf36..b5c279b 100644
---- a/net/strparser/strparser.c
-+++ b/net/strparser/strparser.c
-@@ -504,6 +504,7 @@ static int __init strp_mod_init(void)
-
- static void __exit strp_mod_exit(void)
- {
-+ destroy_workqueue(strp_wq);
- }
- module_init(strp_mod_init);
- module_exit(strp_mod_exit);
diff --git a/4.9.18/1017_linux-4.9.18.patch b/4.9.18/1017_linux-4.9.18.patch
deleted file mode 100644
index 3f957a2..0000000
--- a/4.9.18/1017_linux-4.9.18.patch
+++ /dev/null
@@ -1,876 +0,0 @@
-diff --git a/Makefile b/Makefile
-index 004f90a..c10d0e6 100644
---- a/Makefile
-+++ b/Makefile
-@@ -1,6 +1,6 @@
- VERSION = 4
- PATCHLEVEL = 9
--SUBLEVEL = 17
-+SUBLEVEL = 18
- EXTRAVERSION =
- NAME = Roaring Lionus
-
-diff --git a/arch/parisc/include/asm/cacheflush.h b/arch/parisc/include/asm/cacheflush.h
-index 7bd69bd..1d8c24d 100644
---- a/arch/parisc/include/asm/cacheflush.h
-+++ b/arch/parisc/include/asm/cacheflush.h
-@@ -45,28 +45,9 @@ static inline void flush_kernel_dcache_page(struct page *page)
-
- #define flush_kernel_dcache_range(start,size) \
- flush_kernel_dcache_range_asm((start), (start)+(size));
--/* vmap range flushes and invalidates. Architecturally, we don't need
-- * the invalidate, because the CPU should refuse to speculate once an
-- * area has been flushed, so invalidate is left empty */
--static inline void flush_kernel_vmap_range(void *vaddr, int size)
--{
-- unsigned long start = (unsigned long)vaddr;
--
-- flush_kernel_dcache_range_asm(start, start + size);
--}
--static inline void invalidate_kernel_vmap_range(void *vaddr, int size)
--{
-- unsigned long start = (unsigned long)vaddr;
-- void *cursor = vaddr;
-
-- for ( ; cursor < vaddr + size; cursor += PAGE_SIZE) {
-- struct page *page = vmalloc_to_page(cursor);
--
-- if (test_and_clear_bit(PG_dcache_dirty, &page->flags))
-- flush_kernel_dcache_page(page);
-- }
-- flush_kernel_dcache_range_asm(start, start + size);
--}
-+void flush_kernel_vmap_range(void *vaddr, int size);
-+void invalidate_kernel_vmap_range(void *vaddr, int size);
-
- #define flush_cache_vmap(start, end) flush_cache_all()
- #define flush_cache_vunmap(start, end) flush_cache_all()
-diff --git a/arch/parisc/kernel/cache.c b/arch/parisc/kernel/cache.c
-index 977f0a4f..53ec75f 100644
---- a/arch/parisc/kernel/cache.c
-+++ b/arch/parisc/kernel/cache.c
-@@ -633,3 +633,25 @@ flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, unsigned long
- __flush_cache_page(vma, vmaddr, PFN_PHYS(pfn));
- }
- }
-+
-+void flush_kernel_vmap_range(void *vaddr, int size)
-+{
-+ unsigned long start = (unsigned long)vaddr;
-+
-+ if ((unsigned long)size > parisc_cache_flush_threshold)
-+ flush_data_cache();
-+ else
-+ flush_kernel_dcache_range_asm(start, start + size);
-+}
-+EXPORT_SYMBOL(flush_kernel_vmap_range);
-+
-+void invalidate_kernel_vmap_range(void *vaddr, int size)
-+{
-+ unsigned long start = (unsigned long)vaddr;
-+
-+ if ((unsigned long)size > parisc_cache_flush_threshold)
-+ flush_data_cache();
-+ else
-+ flush_kernel_dcache_range_asm(start, start + size);
-+}
-+EXPORT_SYMBOL(invalidate_kernel_vmap_range);
-diff --git a/arch/parisc/kernel/process.c b/arch/parisc/kernel/process.c
-index 4063943..e81afc37 100644
---- a/arch/parisc/kernel/process.c
-+++ b/arch/parisc/kernel/process.c
-@@ -139,6 +139,8 @@ void machine_power_off(void)
-
- printk(KERN_EMERG "System shut down completed.\n"
- "Please power this system off now.");
-+
-+ for (;;);
- }
-
- void (*pm_power_off)(void) = machine_power_off;
-diff --git a/arch/powerpc/boot/zImage.lds.S b/arch/powerpc/boot/zImage.lds.S
-index 861e721..f080abf 100644
---- a/arch/powerpc/boot/zImage.lds.S
-+++ b/arch/powerpc/boot/zImage.lds.S
-@@ -68,6 +68,7 @@ SECTIONS
- }
-
- #ifdef CONFIG_PPC64_BOOT_WRAPPER
-+ . = ALIGN(256);
- .got :
- {
- __toc_start = .;
-diff --git a/drivers/cpufreq/cpufreq.c b/drivers/cpufreq/cpufreq.c
-index 6e6c1fb..272608f 100644
---- a/drivers/cpufreq/cpufreq.c
-+++ b/drivers/cpufreq/cpufreq.c
-@@ -680,9 +680,11 @@ static ssize_t show_cpuinfo_cur_freq(struct cpufreq_policy *policy,
- char *buf)
- {
- unsigned int cur_freq = __cpufreq_get(policy);
-- if (!cur_freq)
-- return sprintf(buf, "<unknown>");
-- return sprintf(buf, "%u\n", cur_freq);
-+
-+ if (cur_freq)
-+ return sprintf(buf, "%u\n", cur_freq);
-+
-+ return sprintf(buf, "<unknown>\n");
- }
-
- /**
-diff --git a/drivers/gpu/drm/amd/amdgpu/si_dpm.c b/drivers/gpu/drm/amd/amdgpu/si_dpm.c
-index b447a01..09e6a73 100644
---- a/drivers/gpu/drm/amd/amdgpu/si_dpm.c
-+++ b/drivers/gpu/drm/amd/amdgpu/si_dpm.c
-@@ -3506,6 +3506,12 @@ static void si_apply_state_adjust_rules(struct amdgpu_device *adev,
- max_sclk = 75000;
- max_mclk = 80000;
- }
-+ } else if (adev->asic_type == CHIP_OLAND) {
-+ if ((adev->pdev->device == 0x6604) &&
-+ (adev->pdev->subsystem_vendor == 0x1028) &&
-+ (adev->pdev->subsystem_device == 0x066F)) {
-+ max_sclk = 75000;
-+ }
- }
- /* Apply dpm quirks */
- while (p && p->chip_device != 0) {
-diff --git a/drivers/gpu/drm/vc4/vc4_drv.c b/drivers/gpu/drm/vc4/vc4_drv.c
-index 8703f56..246d1ae 100644
---- a/drivers/gpu/drm/vc4/vc4_drv.c
-+++ b/drivers/gpu/drm/vc4/vc4_drv.c
-@@ -61,21 +61,24 @@ static int vc4_get_param_ioctl(struct drm_device *dev, void *data,
- if (ret < 0)
- return ret;
- args->value = V3D_READ(V3D_IDENT0);
-- pm_runtime_put(&vc4->v3d->pdev->dev);
-+ pm_runtime_mark_last_busy(&vc4->v3d->pdev->dev);
-+ pm_runtime_put_autosuspend(&vc4->v3d->pdev->dev);
- break;
- case DRM_VC4_PARAM_V3D_IDENT1:
- ret = pm_runtime_get_sync(&vc4->v3d->pdev->dev);
- if (ret < 0)
- return ret;
- args->value = V3D_READ(V3D_IDENT1);
-- pm_runtime_put(&vc4->v3d->pdev->dev);
-+ pm_runtime_mark_last_busy(&vc4->v3d->pdev->dev);
-+ pm_runtime_put_autosuspend(&vc4->v3d->pdev->dev);
- break;
- case DRM_VC4_PARAM_V3D_IDENT2:
- ret = pm_runtime_get_sync(&vc4->v3d->pdev->dev);
- if (ret < 0)
- return ret;
- args->value = V3D_READ(V3D_IDENT2);
-- pm_runtime_put(&vc4->v3d->pdev->dev);
-+ pm_runtime_mark_last_busy(&vc4->v3d->pdev->dev);
-+ pm_runtime_put_autosuspend(&vc4->v3d->pdev->dev);
- break;
- case DRM_VC4_PARAM_SUPPORTS_BRANCHES:
- args->value = true;
-diff --git a/drivers/gpu/drm/vc4/vc4_gem.c b/drivers/gpu/drm/vc4/vc4_gem.c
-index 18e3717..ab30169 100644
---- a/drivers/gpu/drm/vc4/vc4_gem.c
-+++ b/drivers/gpu/drm/vc4/vc4_gem.c
-@@ -711,8 +711,10 @@ vc4_complete_exec(struct drm_device *dev, struct vc4_exec_info *exec)
- }
-
- mutex_lock(&vc4->power_lock);
-- if (--vc4->power_refcount == 0)
-- pm_runtime_put(&vc4->v3d->pdev->dev);
-+ if (--vc4->power_refcount == 0) {
-+ pm_runtime_mark_last_busy(&vc4->v3d->pdev->dev);
-+ pm_runtime_put_autosuspend(&vc4->v3d->pdev->dev);
-+ }
- mutex_unlock(&vc4->power_lock);
-
- kfree(exec);
-diff --git a/drivers/gpu/drm/vc4/vc4_v3d.c b/drivers/gpu/drm/vc4/vc4_v3d.c
-index e6d3c60..7cc346a 100644
---- a/drivers/gpu/drm/vc4/vc4_v3d.c
-+++ b/drivers/gpu/drm/vc4/vc4_v3d.c
-@@ -222,6 +222,8 @@ static int vc4_v3d_bind(struct device *dev, struct device *master, void *data)
- return ret;
- }
-
-+ pm_runtime_use_autosuspend(dev);
-+ pm_runtime_set_autosuspend_delay(dev, 40); /* a little over 2 frames. */
- pm_runtime_enable(dev);
-
- return 0;
-diff --git a/drivers/gpu/drm/vc4/vc4_validate_shaders.c b/drivers/gpu/drm/vc4/vc4_validate_shaders.c
-index 2543cf5..917321c 100644
---- a/drivers/gpu/drm/vc4/vc4_validate_shaders.c
-+++ b/drivers/gpu/drm/vc4/vc4_validate_shaders.c
-@@ -608,9 +608,7 @@ static bool
- vc4_validate_branches(struct vc4_shader_validation_state *validation_state)
- {
- uint32_t max_branch_target = 0;
-- bool found_shader_end = false;
- int ip;
-- int shader_end_ip = 0;
- int last_branch = -2;
-
- for (ip = 0; ip < validation_state->max_ip; ip++) {
-@@ -621,8 +619,13 @@ vc4_validate_branches(struct vc4_shader_validation_state *validation_state)
- uint32_t branch_target_ip;
-
- if (sig == QPU_SIG_PROG_END) {
-- shader_end_ip = ip;
-- found_shader_end = true;
-+ /* There are two delay slots after program end is
-+ * signaled that are still executed, then we're
-+ * finished. validation_state->max_ip is the
-+ * instruction after the last valid instruction in the
-+ * program.
-+ */
-+ validation_state->max_ip = ip + 3;
- continue;
- }
-
-@@ -676,15 +679,9 @@ vc4_validate_branches(struct vc4_shader_validation_state *validation_state)
- }
- set_bit(after_delay_ip, validation_state->branch_targets);
- max_branch_target = max(max_branch_target, after_delay_ip);
--
-- /* There are two delay slots after program end is signaled
-- * that are still executed, then we're finished.
-- */
-- if (found_shader_end && ip == shader_end_ip + 2)
-- break;
- }
-
-- if (max_branch_target > shader_end_ip) {
-+ if (max_branch_target > validation_state->max_ip - 3) {
- DRM_ERROR("Branch landed after QPU_SIG_PROG_END");
- return false;
- }
-diff --git a/drivers/isdn/gigaset/bas-gigaset.c b/drivers/isdn/gigaset/bas-gigaset.c
-index aecec6d..7f1c625 100644
---- a/drivers/isdn/gigaset/bas-gigaset.c
-+++ b/drivers/isdn/gigaset/bas-gigaset.c
-@@ -2317,6 +2317,9 @@ static int gigaset_probe(struct usb_interface *interface,
- return -ENODEV;
- }
-
-+ if (hostif->desc.bNumEndpoints < 1)
-+ return -ENODEV;
-+
- dev_info(&udev->dev,
- "%s: Device matched (Vendor: 0x%x, Product: 0x%x)\n",
- __func__, le16_to_cpu(udev->descriptor.idVendor),
-diff --git a/drivers/md/raid10.c b/drivers/md/raid10.c
-index 39fddda..55b5e0e 100644
---- a/drivers/md/raid10.c
-+++ b/drivers/md/raid10.c
-@@ -1470,7 +1470,25 @@ static void raid10_make_request(struct mddev *mddev, struct bio *bio)
- split = bio;
- }
-
-+ /*
-+ * If a bio is splitted, the first part of bio will pass
-+ * barrier but the bio is queued in current->bio_list (see
-+ * generic_make_request). If there is a raise_barrier() called
-+ * here, the second part of bio can't pass barrier. But since
-+ * the first part bio isn't dispatched to underlaying disks
-+ * yet, the barrier is never released, hence raise_barrier will
-+ * alays wait. We have a deadlock.
-+ * Note, this only happens in read path. For write path, the
-+ * first part of bio is dispatched in a schedule() call
-+ * (because of blk plug) or offloaded to raid10d.
-+ * Quitting from the function immediately can change the bio
-+ * order queued in bio_list and avoid the deadlock.
-+ */
- __make_request(mddev, split);
-+ if (split != bio && bio_data_dir(bio) == READ) {
-+ generic_make_request(bio);
-+ break;
-+ }
- } while (split != bio);
-
- /* In case raid10d snuck in to freeze_array */
-diff --git a/drivers/scsi/libiscsi.c b/drivers/scsi/libiscsi.c
-index f9b6fba..a530f08 100644
---- a/drivers/scsi/libiscsi.c
-+++ b/drivers/scsi/libiscsi.c
-@@ -560,8 +560,12 @@ static void iscsi_complete_task(struct iscsi_task *task, int state)
- WARN_ON_ONCE(task->state == ISCSI_TASK_FREE);
- task->state = state;
-
-- if (!list_empty(&task->running))
-+ spin_lock_bh(&conn->taskqueuelock);
-+ if (!list_empty(&task->running)) {
-+ pr_debug_once("%s while task on list", __func__);
- list_del_init(&task->running);
-+ }
-+ spin_unlock_bh(&conn->taskqueuelock);
-
- if (conn->task == task)
- conn->task = NULL;
-@@ -783,7 +787,9 @@ __iscsi_conn_send_pdu(struct iscsi_conn *conn, struct iscsi_hdr *hdr,
- if (session->tt->xmit_task(task))
- goto free_task;
- } else {
-+ spin_lock_bh(&conn->taskqueuelock);
- list_add_tail(&task->running, &conn->mgmtqueue);
-+ spin_unlock_bh(&conn->taskqueuelock);
- iscsi_conn_queue_work(conn);
- }
-
-@@ -1474,8 +1480,10 @@ void iscsi_requeue_task(struct iscsi_task *task)
- * this may be on the requeue list already if the xmit_task callout
- * is handling the r2ts while we are adding new ones
- */
-+ spin_lock_bh(&conn->taskqueuelock);
- if (list_empty(&task->running))
- list_add_tail(&task->running, &conn->requeue);
-+ spin_unlock_bh(&conn->taskqueuelock);
- iscsi_conn_queue_work(conn);
- }
- EXPORT_SYMBOL_GPL(iscsi_requeue_task);
-@@ -1512,22 +1520,26 @@ static int iscsi_data_xmit(struct iscsi_conn *conn)
- * only have one nop-out as a ping from us and targets should not
- * overflow us with nop-ins
- */
-+ spin_lock_bh(&conn->taskqueuelock);
- check_mgmt:
- while (!list_empty(&conn->mgmtqueue)) {
- conn->task = list_entry(conn->mgmtqueue.next,
- struct iscsi_task, running);
- list_del_init(&conn->task->running);
-+ spin_unlock_bh(&conn->taskqueuelock);
- if (iscsi_prep_mgmt_task(conn, conn->task)) {
- /* regular RX path uses back_lock */
- spin_lock_bh(&conn->session->back_lock);
- __iscsi_put_task(conn->task);
- spin_unlock_bh(&conn->session->back_lock);
- conn->task = NULL;
-+ spin_lock_bh(&conn->taskqueuelock);
- continue;
- }
- rc = iscsi_xmit_task(conn);
- if (rc)
- goto done;
-+ spin_lock_bh(&conn->taskqueuelock);
- }
-
- /* process pending command queue */
-@@ -1535,19 +1547,24 @@ static int iscsi_data_xmit(struct iscsi_conn *conn)
- conn->task = list_entry(conn->cmdqueue.next, struct iscsi_task,
- running);
- list_del_init(&conn->task->running);
-+ spin_unlock_bh(&conn->taskqueuelock);
- if (conn->session->state == ISCSI_STATE_LOGGING_OUT) {
- fail_scsi_task(conn->task, DID_IMM_RETRY);
-+ spin_lock_bh(&conn->taskqueuelock);
- continue;
- }
- rc = iscsi_prep_scsi_cmd_pdu(conn->task);
- if (rc) {
- if (rc == -ENOMEM || rc == -EACCES) {
-+ spin_lock_bh(&conn->taskqueuelock);
- list_add_tail(&conn->task->running,
- &conn->cmdqueue);
- conn->task = NULL;
-+ spin_unlock_bh(&conn->taskqueuelock);
- goto done;
- } else
- fail_scsi_task(conn->task, DID_ABORT);
-+ spin_lock_bh(&conn->taskqueuelock);
- continue;
- }
- rc = iscsi_xmit_task(conn);
-@@ -1558,6 +1575,7 @@ static int iscsi_data_xmit(struct iscsi_conn *conn)
- * we need to check the mgmt queue for nops that need to
- * be sent to aviod starvation
- */
-+ spin_lock_bh(&conn->taskqueuelock);
- if (!list_empty(&conn->mgmtqueue))
- goto check_mgmt;
- }
-@@ -1577,12 +1595,15 @@ static int iscsi_data_xmit(struct iscsi_conn *conn)
- conn->task = task;
- list_del_init(&conn->task->running);
- conn->task->state = ISCSI_TASK_RUNNING;
-+ spin_unlock_bh(&conn->taskqueuelock);
- rc = iscsi_xmit_task(conn);
- if (rc)
- goto done;
-+ spin_lock_bh(&conn->taskqueuelock);
- if (!list_empty(&conn->mgmtqueue))
- goto check_mgmt;
- }
-+ spin_unlock_bh(&conn->taskqueuelock);
- spin_unlock_bh(&conn->session->frwd_lock);
- return -ENODATA;
-
-@@ -1738,7 +1759,9 @@ int iscsi_queuecommand(struct Scsi_Host *host, struct scsi_cmnd *sc)
- goto prepd_reject;
- }
- } else {
-+ spin_lock_bh(&conn->taskqueuelock);
- list_add_tail(&task->running, &conn->cmdqueue);
-+ spin_unlock_bh(&conn->taskqueuelock);
- iscsi_conn_queue_work(conn);
- }
-
-@@ -2897,6 +2920,7 @@ iscsi_conn_setup(struct iscsi_cls_session *cls_session, int dd_size,
- INIT_LIST_HEAD(&conn->mgmtqueue);
- INIT_LIST_HEAD(&conn->cmdqueue);
- INIT_LIST_HEAD(&conn->requeue);
-+ spin_lock_init(&conn->taskqueuelock);
- INIT_WORK(&conn->xmitwork, iscsi_xmitworker);
-
- /* allocate login_task used for the login/text sequences */
-diff --git a/drivers/scsi/lpfc/lpfc_init.c b/drivers/scsi/lpfc/lpfc_init.c
-index 734a042..f7e3f27 100644
---- a/drivers/scsi/lpfc/lpfc_init.c
-+++ b/drivers/scsi/lpfc/lpfc_init.c
-@@ -11393,6 +11393,7 @@ static struct pci_driver lpfc_driver = {
- .id_table = lpfc_id_table,
- .probe = lpfc_pci_probe_one,
- .remove = lpfc_pci_remove_one,
-+ .shutdown = lpfc_pci_remove_one,
- .suspend = lpfc_pci_suspend_one,
- .resume = lpfc_pci_resume_one,
- .err_handler = &lpfc_err_handler,
-diff --git a/drivers/scsi/qla2xxx/qla_target.c b/drivers/scsi/qla2xxx/qla_target.c
-index bff9689..feab7ea 100644
---- a/drivers/scsi/qla2xxx/qla_target.c
-+++ b/drivers/scsi/qla2xxx/qla_target.c
-@@ -5375,16 +5375,22 @@ qlt_send_busy(struct scsi_qla_host *vha,
-
- static int
- qlt_chk_qfull_thresh_hold(struct scsi_qla_host *vha,
-- struct atio_from_isp *atio)
-+ struct atio_from_isp *atio, bool ha_locked)
- {
- struct qla_hw_data *ha = vha->hw;
- uint16_t status;
-+ unsigned long flags;
-
- if (ha->tgt.num_pend_cmds < Q_FULL_THRESH_HOLD(ha))
- return 0;
-
-+ if (!ha_locked)
-+ spin_lock_irqsave(&ha->hardware_lock, flags);
- status = temp_sam_status;
- qlt_send_busy(vha, atio, status);
-+ if (!ha_locked)
-+ spin_unlock_irqrestore(&ha->hardware_lock, flags);
-+
- return 1;
- }
-
-@@ -5429,7 +5435,7 @@ static void qlt_24xx_atio_pkt(struct scsi_qla_host *vha,
-
-
- if (likely(atio->u.isp24.fcp_cmnd.task_mgmt_flags == 0)) {
-- rc = qlt_chk_qfull_thresh_hold(vha, atio);
-+ rc = qlt_chk_qfull_thresh_hold(vha, atio, ha_locked);
- if (rc != 0) {
- tgt->atio_irq_cmd_count--;
- return;
-@@ -5552,7 +5558,7 @@ static void qlt_response_pkt(struct scsi_qla_host *vha, response_t *pkt)
- break;
- }
-
-- rc = qlt_chk_qfull_thresh_hold(vha, atio);
-+ rc = qlt_chk_qfull_thresh_hold(vha, atio, true);
- if (rc != 0) {
- tgt->irq_cmd_count--;
- return;
-@@ -6794,6 +6800,8 @@ qlt_handle_abts_recv_work(struct work_struct *work)
- spin_lock_irqsave(&ha->hardware_lock, flags);
- qlt_response_pkt_all_vps(vha, (response_t *)&op->atio);
- spin_unlock_irqrestore(&ha->hardware_lock, flags);
-+
-+ kfree(op);
- }
-
- void
-diff --git a/drivers/target/target_core_pscsi.c b/drivers/target/target_core_pscsi.c
-index 9125d93..ef1c8c1 100644
---- a/drivers/target/target_core_pscsi.c
-+++ b/drivers/target/target_core_pscsi.c
-@@ -154,7 +154,7 @@ static void pscsi_tape_read_blocksize(struct se_device *dev,
-
- buf = kzalloc(12, GFP_KERNEL);
- if (!buf)
-- return;
-+ goto out_free;
-
- memset(cdb, 0, MAX_COMMAND_SIZE);
- cdb[0] = MODE_SENSE;
-@@ -169,9 +169,10 @@ static void pscsi_tape_read_blocksize(struct se_device *dev,
- * If MODE_SENSE still returns zero, set the default value to 1024.
- */
- sdev->sector_size = (buf[9] << 16) | (buf[10] << 8) | (buf[11]);
-+out_free:
- if (!sdev->sector_size)
- sdev->sector_size = 1024;
--out_free:
-+
- kfree(buf);
- }
-
-@@ -314,9 +315,10 @@ static int pscsi_add_device_to_list(struct se_device *dev,
- sd->lun, sd->queue_depth);
- }
-
-- dev->dev_attrib.hw_block_size = sd->sector_size;
-+ dev->dev_attrib.hw_block_size =
-+ min_not_zero((int)sd->sector_size, 512);
- dev->dev_attrib.hw_max_sectors =
-- min_t(int, sd->host->max_sectors, queue_max_hw_sectors(q));
-+ min_not_zero(sd->host->max_sectors, queue_max_hw_sectors(q));
- dev->dev_attrib.hw_queue_depth = sd->queue_depth;
-
- /*
-@@ -339,8 +341,10 @@ static int pscsi_add_device_to_list(struct se_device *dev,
- /*
- * For TYPE_TAPE, attempt to determine blocksize with MODE_SENSE.
- */
-- if (sd->type == TYPE_TAPE)
-+ if (sd->type == TYPE_TAPE) {
- pscsi_tape_read_blocksize(dev, sd);
-+ dev->dev_attrib.hw_block_size = sd->sector_size;
-+ }
- return 0;
- }
-
-@@ -406,7 +410,7 @@ static int pscsi_create_type_disk(struct se_device *dev, struct scsi_device *sd)
- /*
- * Called with struct Scsi_Host->host_lock called.
- */
--static int pscsi_create_type_rom(struct se_device *dev, struct scsi_device *sd)
-+static int pscsi_create_type_nondisk(struct se_device *dev, struct scsi_device *sd)
- __releases(sh->host_lock)
- {
- struct pscsi_hba_virt *phv = dev->se_hba->hba_ptr;
-@@ -433,28 +437,6 @@ static int pscsi_create_type_rom(struct se_device *dev, struct scsi_device *sd)
- return 0;
- }
-
--/*
-- * Called with struct Scsi_Host->host_lock called.
-- */
--static int pscsi_create_type_other(struct se_device *dev,
-- struct scsi_device *sd)
-- __releases(sh->host_lock)
--{
-- struct pscsi_hba_virt *phv = dev->se_hba->hba_ptr;
-- struct Scsi_Host *sh = sd->host;
-- int ret;
--
-- spin_unlock_irq(sh->host_lock);
-- ret = pscsi_add_device_to_list(dev, sd);
-- if (ret)
-- return ret;
--
-- pr_debug("CORE_PSCSI[%d] - Added Type: %s for %d:%d:%d:%llu\n",
-- phv->phv_host_id, scsi_device_type(sd->type), sh->host_no,
-- sd->channel, sd->id, sd->lun);
-- return 0;
--}
--
- static int pscsi_configure_device(struct se_device *dev)
- {
- struct se_hba *hba = dev->se_hba;
-@@ -542,11 +524,8 @@ static int pscsi_configure_device(struct se_device *dev)
- case TYPE_DISK:
- ret = pscsi_create_type_disk(dev, sd);
- break;
-- case TYPE_ROM:
-- ret = pscsi_create_type_rom(dev, sd);
-- break;
- default:
-- ret = pscsi_create_type_other(dev, sd);
-+ ret = pscsi_create_type_nondisk(dev, sd);
- break;
- }
-
-@@ -611,8 +590,7 @@ static void pscsi_free_device(struct se_device *dev)
- else if (pdv->pdv_lld_host)
- scsi_host_put(pdv->pdv_lld_host);
-
-- if ((sd->type == TYPE_DISK) || (sd->type == TYPE_ROM))
-- scsi_device_put(sd);
-+ scsi_device_put(sd);
-
- pdv->pdv_sd = NULL;
- }
-@@ -1069,7 +1047,6 @@ static sector_t pscsi_get_blocks(struct se_device *dev)
- if (pdv->pdv_bd && pdv->pdv_bd->bd_part)
- return pdv->pdv_bd->bd_part->nr_sects;
-
-- dump_stack();
- return 0;
- }
-
-diff --git a/drivers/target/target_core_sbc.c b/drivers/target/target_core_sbc.c
-index aabd660..a53fb23 100644
---- a/drivers/target/target_core_sbc.c
-+++ b/drivers/target/target_core_sbc.c
-@@ -1104,9 +1104,15 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
- return ret;
- break;
- case VERIFY:
-+ case VERIFY_16:
- size = 0;
-- sectors = transport_get_sectors_10(cdb);
-- cmd->t_task_lba = transport_lba_32(cdb);
-+ if (cdb[0] == VERIFY) {
-+ sectors = transport_get_sectors_10(cdb);
-+ cmd->t_task_lba = transport_lba_32(cdb);
-+ } else {
-+ sectors = transport_get_sectors_16(cdb);
-+ cmd->t_task_lba = transport_lba_64(cdb);
-+ }
- cmd->execute_cmd = sbc_emulate_noop;
- goto check_lba;
- case REZERO_UNIT:
-diff --git a/fs/ext4/super.c b/fs/ext4/super.c
-index afe29ba..5fa9ba1 100644
---- a/fs/ext4/super.c
-+++ b/fs/ext4/super.c
-@@ -3830,7 +3830,7 @@ static int ext4_fill_super(struct super_block *sb, void *data, int silent)
- db_count = (sbi->s_groups_count + EXT4_DESC_PER_BLOCK(sb) - 1) /
- EXT4_DESC_PER_BLOCK(sb);
- if (ext4_has_feature_meta_bg(sb)) {
-- if (le32_to_cpu(es->s_first_meta_bg) >= db_count) {
-+ if (le32_to_cpu(es->s_first_meta_bg) > db_count) {
- ext4_msg(sb, KERN_WARNING,
- "first meta block group too large: %u "
- "(group descriptor block count %u)",
-diff --git a/fs/gfs2/incore.h b/fs/gfs2/incore.h
-index a6a3389..51519c2 100644
---- a/fs/gfs2/incore.h
-+++ b/fs/gfs2/incore.h
-@@ -207,7 +207,7 @@ struct lm_lockname {
- struct gfs2_sbd *ln_sbd;
- u64 ln_number;
- unsigned int ln_type;
--};
-+} __packed __aligned(sizeof(int));
-
- #define lm_name_equal(name1, name2) \
- (((name1)->ln_number == (name2)->ln_number) && \
-diff --git a/fs/nfs/nfs4proc.c b/fs/nfs/nfs4proc.c
-index 609840d..1536aeb 100644
---- a/fs/nfs/nfs4proc.c
-+++ b/fs/nfs/nfs4proc.c
-@@ -7426,11 +7426,11 @@ static void nfs4_exchange_id_release(void *data)
- struct nfs41_exchange_id_data *cdata =
- (struct nfs41_exchange_id_data *)data;
-
-- nfs_put_client(cdata->args.client);
- if (cdata->xprt) {
- xprt_put(cdata->xprt);
- rpc_clnt_xprt_switch_put(cdata->args.client->cl_rpcclient);
- }
-+ nfs_put_client(cdata->args.client);
- kfree(cdata->res.impl_id);
- kfree(cdata->res.server_scope);
- kfree(cdata->res.server_owner);
-@@ -7537,10 +7537,8 @@ static int _nfs4_proc_exchange_id(struct nfs_client *clp, struct rpc_cred *cred,
- task_setup_data.callback_data = calldata;
-
- task = rpc_run_task(&task_setup_data);
-- if (IS_ERR(task)) {
-- status = PTR_ERR(task);
-- goto out_impl_id;
-- }
-+ if (IS_ERR(task))
-+ return PTR_ERR(task);
-
- if (!xprt) {
- status = rpc_wait_for_completion_task(task);
-@@ -7568,6 +7566,7 @@ static int _nfs4_proc_exchange_id(struct nfs_client *clp, struct rpc_cred *cred,
- kfree(calldata->res.server_owner);
- out_calldata:
- kfree(calldata);
-+ nfs_put_client(clp);
- goto out;
- }
-
-diff --git a/include/linux/log2.h b/include/linux/log2.h
-index fd7ff3d..f38fae2 100644
---- a/include/linux/log2.h
-+++ b/include/linux/log2.h
-@@ -16,12 +16,6 @@
- #include <linux/bitops.h>
-
- /*
-- * deal with unrepresentable constant logarithms
-- */
--extern __attribute__((const, noreturn))
--int ____ilog2_NaN(void);
--
--/*
- * non-constant log of base 2 calculators
- * - the arch may override these in asm/bitops.h if they can be implemented
- * more efficiently than using fls() and fls64()
-@@ -85,7 +79,7 @@ unsigned long __rounddown_pow_of_two(unsigned long n)
- #define ilog2(n) \
- ( \
- __builtin_constant_p(n) ? ( \
-- (n) < 1 ? ____ilog2_NaN() : \
-+ (n) < 2 ? 0 : \
- (n) & (1ULL << 63) ? 63 : \
- (n) & (1ULL << 62) ? 62 : \
- (n) & (1ULL << 61) ? 61 : \
-@@ -148,10 +142,7 @@ unsigned long __rounddown_pow_of_two(unsigned long n)
- (n) & (1ULL << 4) ? 4 : \
- (n) & (1ULL << 3) ? 3 : \
- (n) & (1ULL << 2) ? 2 : \
-- (n) & (1ULL << 1) ? 1 : \
-- (n) & (1ULL << 0) ? 0 : \
-- ____ilog2_NaN() \
-- ) : \
-+ 1 ) : \
- (sizeof(n) <= 4) ? \
- __ilog2_u32(n) : \
- __ilog2_u64(n) \
-diff --git a/include/scsi/libiscsi.h b/include/scsi/libiscsi.h
-index 4d1c46a..c7b1dc7 100644
---- a/include/scsi/libiscsi.h
-+++ b/include/scsi/libiscsi.h
-@@ -196,6 +196,7 @@ struct iscsi_conn {
- struct iscsi_task *task; /* xmit task in progress */
-
- /* xmit */
-+ spinlock_t taskqueuelock; /* protects the next three lists */
- struct list_head mgmtqueue; /* mgmt (control) xmit queue */
- struct list_head cmdqueue; /* data-path cmd queue */
- struct list_head requeue; /* tasks needing another run */
-diff --git a/kernel/cgroup_pids.c b/kernel/cgroup_pids.c
-index 2bd6737..a57242e 100644
---- a/kernel/cgroup_pids.c
-+++ b/kernel/cgroup_pids.c
-@@ -229,7 +229,7 @@ static int pids_can_fork(struct task_struct *task)
- /* Only log the first time events_limit is incremented. */
- if (atomic64_inc_return(&pids->events_limit) == 1) {
- pr_info("cgroup: fork rejected by pids controller in ");
-- pr_cont_cgroup_path(task_cgroup(current, pids_cgrp_id));
-+ pr_cont_cgroup_path(css->cgroup);
- pr_cont("\n");
- }
- cgroup_file_notify(&pids->events_file);
-diff --git a/kernel/events/core.c b/kernel/events/core.c
-index 4b33231..07c0dc8 100644
---- a/kernel/events/core.c
-+++ b/kernel/events/core.c
-@@ -10333,6 +10333,17 @@ void perf_event_free_task(struct task_struct *task)
- continue;
-
- mutex_lock(&ctx->mutex);
-+ raw_spin_lock_irq(&ctx->lock);
-+ /*
-+ * Destroy the task <-> ctx relation and mark the context dead.
-+ *
-+ * This is important because even though the task hasn't been
-+ * exposed yet the context has been (through child_list).
-+ */
-+ RCU_INIT_POINTER(task->perf_event_ctxp[ctxn], NULL);
-+ WRITE_ONCE(ctx->task, TASK_TOMBSTONE);
-+ put_task_struct(task); /* cannot be last */
-+ raw_spin_unlock_irq(&ctx->lock);
- again:
- list_for_each_entry_safe(event, tmp, &ctx->pinned_groups,
- group_entry)
-@@ -10586,7 +10597,7 @@ static int perf_event_init_context(struct task_struct *child, int ctxn)
- ret = inherit_task_group(event, parent, parent_ctx,
- child, ctxn, &inherited_all);
- if (ret)
-- break;
-+ goto out_unlock;
- }
-
- /*
-@@ -10602,7 +10613,7 @@ static int perf_event_init_context(struct task_struct *child, int ctxn)
- ret = inherit_task_group(event, parent, parent_ctx,
- child, ctxn, &inherited_all);
- if (ret)
-- break;
-+ goto out_unlock;
- }
-
- raw_spin_lock_irqsave(&parent_ctx->lock, flags);
-@@ -10630,6 +10641,7 @@ static int perf_event_init_context(struct task_struct *child, int ctxn)
- }
-
- raw_spin_unlock_irqrestore(&parent_ctx->lock, flags);
-+out_unlock:
- mutex_unlock(&parent_ctx->mutex);
-
- perf_unpin_context(parent_ctx);
-diff --git a/mm/percpu.c b/mm/percpu.c
-index 2557143..f014ceb 100644
---- a/mm/percpu.c
-+++ b/mm/percpu.c
-@@ -1010,8 +1010,11 @@ static void __percpu *pcpu_alloc(size_t size, size_t align, bool reserved,
- mutex_unlock(&pcpu_alloc_mutex);
- }
-
-- if (chunk != pcpu_reserved_chunk)
-+ if (chunk != pcpu_reserved_chunk) {
-+ spin_lock_irqsave(&pcpu_lock, flags);
- pcpu_nr_empty_pop_pages -= occ_pages;
-+ spin_unlock_irqrestore(&pcpu_lock, flags);
-+ }
-
- if (pcpu_nr_empty_pop_pages < PCPU_EMPTY_POP_PAGES_LOW)
- pcpu_schedule_balance_work();
-diff --git a/net/sunrpc/xprtrdma/verbs.c b/net/sunrpc/xprtrdma/verbs.c
-index e2c37061..69502fa 100644
---- a/net/sunrpc/xprtrdma/verbs.c
-+++ b/net/sunrpc/xprtrdma/verbs.c
-@@ -486,7 +486,8 @@ rpcrdma_ep_create(struct rpcrdma_ep *ep, struct rpcrdma_ia *ia,
- struct ib_cq *sendcq, *recvcq;
- int rc;
-
-- max_sge = min(ia->ri_device->attrs.max_sge, RPCRDMA_MAX_SEND_SGES);
-+ max_sge = min_t(unsigned int, ia->ri_device->attrs.max_sge,
-+ RPCRDMA_MAX_SEND_SGES);
- if (max_sge < RPCRDMA_MIN_SEND_SGES) {
- pr_warn("rpcrdma: HCA provides only %d send SGEs\n", max_sge);
- return -ENOMEM;
-diff --git a/tools/include/linux/log2.h b/tools/include/linux/log2.h
-index 4144666..d5677d3 100644
---- a/tools/include/linux/log2.h
-+++ b/tools/include/linux/log2.h
-@@ -13,12 +13,6 @@
- #define _TOOLS_LINUX_LOG2_H
-
- /*
-- * deal with unrepresentable constant logarithms
-- */
--extern __attribute__((const, noreturn))
--int ____ilog2_NaN(void);
--
--/*
- * non-constant log of base 2 calculators
- * - the arch may override these in asm/bitops.h if they can be implemented
- * more efficiently than using fls() and fls64()
-@@ -78,7 +72,7 @@ unsigned long __rounddown_pow_of_two(unsigned long n)
- #define ilog2(n) \
- ( \
- __builtin_constant_p(n) ? ( \
-- (n) < 1 ? ____ilog2_NaN() : \
-+ (n) < 2 ? 0 : \
- (n) & (1ULL << 63) ? 63 : \
- (n) & (1ULL << 62) ? 62 : \
- (n) & (1ULL << 61) ? 61 : \
-@@ -141,10 +135,7 @@ unsigned long __rounddown_pow_of_two(unsigned long n)
- (n) & (1ULL << 4) ? 4 : \
- (n) & (1ULL << 3) ? 3 : \
- (n) & (1ULL << 2) ? 2 : \
-- (n) & (1ULL << 1) ? 1 : \
-- (n) & (1ULL << 0) ? 0 : \
-- ____ilog2_NaN() \
-- ) : \
-+ 1 ) : \
- (sizeof(n) <= 4) ? \
- __ilog2_u32(n) : \
- __ilog2_u64(n) \
diff --git a/4.9.18/0000_README b/4.9.20/0000_README
index 8c12f63..a960856 100644
--- a/4.9.18/0000_README
+++ b/4.9.20/0000_README
@@ -2,15 +2,7 @@ README
-----------------------------------------------------------------------------
Individual Patch Descriptions:
-----------------------------------------------------------------------------
-Patch: 1016_linux-4.9.17.patch
-From: http://www.kernel.org
-Desc: Linux 4.9.17
-
-Patch: 1017_linux-4.9.18.patch
-From: http://www.kernel.org
-Desc: Linux 4.9.18
-
-Patch: 4420_grsecurity-3.1-4.9.18-201703261106.patch
+Patch: 4420_grsecurity-3.1-4.9.20-201703310823.patch
From: http://www.grsecurity.net
Desc: hardened-sources base patch from upstream grsecurity
diff --git a/4.9.18/4420_grsecurity-3.1-4.9.18-201703261106.patch b/4.9.20/4420_grsecurity-3.1-4.9.20-201703310823.patch
index 3659b97..f803149 100644
--- a/4.9.18/4420_grsecurity-3.1-4.9.18-201703261106.patch
+++ b/4.9.20/4420_grsecurity-3.1-4.9.20-201703310823.patch
@@ -419,7 +419,7 @@ index 3d0ae15..84e5412 100644
cmd_syscalls = $(CONFIG_SHELL) $< $(CC) $(c_flags) $(missing_syscalls_flags)
diff --git a/Makefile b/Makefile
-index c10d0e6..54799eb2 100644
+index 4496018..3f9a080 100644
--- a/Makefile
+++ b/Makefile
@@ -302,7 +302,9 @@ CONFIG_SHELL := $(shell if [ -x "$$BASH" ]; then echo $$BASH; \
@@ -7351,10 +7351,10 @@ index 1652f36..0e22377 100644
{
struct pt_regs *regs;
diff --git a/arch/mips/kernel/ptrace.c b/arch/mips/kernel/ptrace.c
-index a92994d..e389b11 100644
+index bf83dc1..775bed8 100644
--- a/arch/mips/kernel/ptrace.c
+++ b/arch/mips/kernel/ptrace.c
-@@ -882,6 +882,10 @@ long arch_ptrace(struct task_struct *child, long request,
+@@ -883,6 +883,10 @@ long arch_ptrace(struct task_struct *child, long request,
return ret;
}
@@ -7365,7 +7365,7 @@ index a92994d..e389b11 100644
/*
* Notification of system call entry/exit
* - triggered by current->work.syscall_trace
-@@ -899,6 +903,11 @@ asmlinkage long syscall_trace_enter(struct pt_regs *regs, long syscall)
+@@ -900,6 +904,11 @@ asmlinkage long syscall_trace_enter(struct pt_regs *regs, long syscall)
if (secure_computing(NULL) == -1)
return -1;
@@ -11324,7 +11324,7 @@ index 79cc0d1..46d6233 100644
.getproplen = prom_getproplen,
.getproperty = prom_getproperty,
diff --git a/arch/sparc/kernel/ptrace_64.c b/arch/sparc/kernel/ptrace_64.c
-index ac082dd..7170942 100644
+index 7037ca3..070b51b 100644
--- a/arch/sparc/kernel/ptrace_64.c
+++ b/arch/sparc/kernel/ptrace_64.c
@@ -1068,6 +1068,10 @@ long arch_ptrace(struct task_struct *child, long request,
@@ -36277,7 +36277,7 @@ index 69b8f8a..0cf39f5 100644
vcpu->arch.regs_avail = ~((1 << VCPU_REGS_RIP) | (1 << VCPU_REGS_RSP)
diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
-index 731044e..399463d 100644
+index e5bc139..5a1766b 100644
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -2005,8 +2005,8 @@ static int xen_hvm_config(struct kvm_vcpu *vcpu, u64 data)
@@ -49817,7 +49817,7 @@ index 4d3ec92..cf501fc 100644
ret = cpufreq_register_driver(&dt_cpufreq_driver);
if (ret)
diff --git a/drivers/cpufreq/cpufreq.c b/drivers/cpufreq/cpufreq.c
-index 272608f..5c4a47a 100644
+index cac4a92..93c0aed 100644
--- a/drivers/cpufreq/cpufreq.c
+++ b/drivers/cpufreq/cpufreq.c
@@ -528,12 +528,12 @@ EXPORT_SYMBOL_GPL(cpufreq_driver_resolve_freq);
@@ -49835,7 +49835,7 @@ index 272608f..5c4a47a 100644
const char *buf, size_t count)
{
int ret, enable;
-@@ -2116,7 +2116,7 @@ void cpufreq_unregister_governor(struct cpufreq_governor *governor)
+@@ -2119,7 +2119,7 @@ void cpufreq_unregister_governor(struct cpufreq_governor *governor)
read_unlock_irqrestore(&cpufreq_driver_lock, flags);
mutex_lock(&cpufreq_governor_mutex);
@@ -49844,7 +49844,7 @@ index 272608f..5c4a47a 100644
mutex_unlock(&cpufreq_governor_mutex);
return;
}
-@@ -2336,13 +2336,17 @@ int cpufreq_boost_trigger_state(int state)
+@@ -2339,13 +2339,17 @@ int cpufreq_boost_trigger_state(int state)
return 0;
write_lock_irqsave(&cpufreq_driver_lock, flags);
@@ -49864,7 +49864,7 @@ index 272608f..5c4a47a 100644
write_unlock_irqrestore(&cpufreq_driver_lock, flags);
pr_err("%s: Cannot %s BOOST\n",
-@@ -2383,7 +2387,9 @@ int cpufreq_enable_boost_support(void)
+@@ -2386,7 +2390,9 @@ int cpufreq_enable_boost_support(void)
if (cpufreq_boost_supported())
return 0;
@@ -49875,7 +49875,7 @@ index 272608f..5c4a47a 100644
/* This will get removed on driver unregister */
return create_boost_sysfs_file();
-@@ -2441,8 +2447,11 @@ int cpufreq_register_driver(struct cpufreq_driver *driver_data)
+@@ -2444,8 +2450,11 @@ int cpufreq_register_driver(struct cpufreq_driver *driver_data)
cpufreq_driver = driver_data;
write_unlock_irqrestore(&cpufreq_driver_lock, flags);
@@ -52248,7 +52248,7 @@ index 1fd6eac..e4206c9 100644
return 0;
}
diff --git a/drivers/gpu/drm/drm_fops.c b/drivers/gpu/drm/drm_fops.c
-index e84faec..03aaa9f 100644
+index f5815e1..106f6e1 100644
--- a/drivers/gpu/drm/drm_fops.c
+++ b/drivers/gpu/drm/drm_fops.c
@@ -132,7 +132,7 @@ int drm_open(struct inode *inode, struct file *filp)
@@ -52935,7 +52935,7 @@ index 97f3a56..32c712e 100644
ret = drm_ioctl(filp, cmd, arg);
diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c
-index 3fc286c..4c19f25 100644
+index 3fc286cd..4c19f25 100644
--- a/drivers/gpu/drm/i915/i915_irq.c
+++ b/drivers/gpu/drm/i915/i915_irq.c
@@ -4511,15 +4511,16 @@ void intel_irq_init(struct drm_i915_private *dev_priv)
@@ -55328,7 +55328,7 @@ index c13fb5b..55a3802 100644
*off += size;
diff --git a/drivers/hv/channel.c b/drivers/hv/channel.c
-index be34547..df73ac5 100644
+index 1606e7f..b207d4b 100644
--- a/drivers/hv/channel.c
+++ b/drivers/hv/channel.c
@@ -404,7 +404,7 @@ int vmbus_establish_gpadl(struct vmbus_channel *channel, void *kbuffer,
@@ -55340,7 +55340,7 @@ index be34547..df73ac5 100644
ret = create_gpadl_header(kbuffer, size, &msginfo);
if (ret)
-@@ -734,9 +734,7 @@ int vmbus_sendpacket_pagebuffer_ctl(struct vmbus_channel *channel,
+@@ -737,9 +737,7 @@ int vmbus_sendpacket_pagebuffer_ctl(struct vmbus_channel *channel,
* Adjust the size down since vmbus_channel_packet_page_buffer is the
* largest size we support
*/
@@ -57462,10 +57462,10 @@ index 4a95b22..874c182 100644
#include <linux/gameport.h>
#include <linux/jiffies.h>
diff --git a/drivers/input/misc/ims-pcu.c b/drivers/input/misc/ims-pcu.c
-index 9c0ea36..1e1a411 100644
+index f4e8fbe..0efd9d6 100644
--- a/drivers/input/misc/ims-pcu.c
+++ b/drivers/input/misc/ims-pcu.c
-@@ -1855,7 +1855,7 @@ static int ims_pcu_identify_type(struct ims_pcu *pcu, u8 *device_id)
+@@ -1859,7 +1859,7 @@ static int ims_pcu_identify_type(struct ims_pcu *pcu, u8 *device_id)
static int ims_pcu_init_application_mode(struct ims_pcu *pcu)
{
@@ -57474,7 +57474,7 @@ index 9c0ea36..1e1a411 100644
const struct ims_pcu_device_info *info;
int error;
-@@ -1886,7 +1886,7 @@ static int ims_pcu_init_application_mode(struct ims_pcu *pcu)
+@@ -1890,7 +1890,7 @@ static int ims_pcu_init_application_mode(struct ims_pcu *pcu)
}
/* Device appears to be operable, complete initialization */
@@ -65843,10 +65843,10 @@ index 9b56b40..f183a4d 100644
struct lance_private *lp = netdev_priv(dev);
int entry, skblen, len;
diff --git a/drivers/net/ethernet/amd/xgbe/xgbe-common.h b/drivers/net/ethernet/amd/xgbe/xgbe-common.h
-index bbef959..999ab1d 100644
+index 1592e1c..26df6c5 100644
--- a/drivers/net/ethernet/amd/xgbe/xgbe-common.h
+++ b/drivers/net/ethernet/amd/xgbe/xgbe-common.h
-@@ -1283,14 +1283,14 @@ do { \
+@@ -1285,14 +1285,14 @@ do { \
* operations, everything works on mask values.
*/
#define XMDIO_READ(_pdata, _mmd, _reg) \
@@ -65935,10 +65935,10 @@ index b3bc87f..5bdfdd3 100644
+ .wrapper_rx_desc_init = xgbe_wrapper_rx_descriptor_init,
+};
diff --git a/drivers/net/ethernet/amd/xgbe/xgbe-dev.c b/drivers/net/ethernet/amd/xgbe/xgbe-dev.c
-index 1babcc1..aa7f8f4e 100644
+index ca106d4..36c4702 100644
--- a/drivers/net/ethernet/amd/xgbe/xgbe-dev.c
+++ b/drivers/net/ethernet/amd/xgbe/xgbe-dev.c
-@@ -2816,7 +2816,7 @@ static void xgbe_powerdown_rx(struct xgbe_prv_data *pdata)
+@@ -2818,7 +2818,7 @@ static void xgbe_powerdown_rx(struct xgbe_prv_data *pdata)
static int xgbe_init(struct xgbe_prv_data *pdata)
{
@@ -65947,7 +65947,7 @@ index 1babcc1..aa7f8f4e 100644
int ret;
DBGPR("-->xgbe_init\n");
-@@ -2882,107 +2882,102 @@ static int xgbe_init(struct xgbe_prv_data *pdata)
+@@ -2884,107 +2884,102 @@ static int xgbe_init(struct xgbe_prv_data *pdata)
return 0;
}
@@ -66133,7 +66133,7 @@ index 1babcc1..aa7f8f4e 100644
+ .set_rss_lookup_table = xgbe_set_rss_lookup_table,
+};
diff --git a/drivers/net/ethernet/amd/xgbe/xgbe-drv.c b/drivers/net/ethernet/amd/xgbe/xgbe-drv.c
-index 7f9216d..26872f6 100644
+index 0f0f3014..882be95 100644
--- a/drivers/net/ethernet/amd/xgbe/xgbe-drv.c
+++ b/drivers/net/ethernet/amd/xgbe/xgbe-drv.c
@@ -245,7 +245,7 @@ static int xgbe_maybe_stop_tx_queue(struct xgbe_channel *channel,
@@ -66380,7 +66380,7 @@ index 7f9216d..26872f6 100644
struct xgbe_ring *ring = channel->rx_ring;
struct xgbe_ring_data *rdata;
-@@ -1794,8 +1794,8 @@ static struct sk_buff *xgbe_create_skb(struct xgbe_prv_data *pdata,
+@@ -1812,8 +1812,8 @@ static unsigned int xgbe_rx_buf2_len(struct xgbe_ring_data *rdata,
static int xgbe_tx_poll(struct xgbe_channel *channel)
{
struct xgbe_prv_data *pdata = channel->pdata;
@@ -66391,7 +66391,7 @@ index 7f9216d..26872f6 100644
struct xgbe_ring *ring = channel->tx_ring;
struct xgbe_ring_data *rdata;
struct xgbe_ring_desc *rdesc;
-@@ -1865,7 +1865,7 @@ static int xgbe_tx_poll(struct xgbe_channel *channel)
+@@ -1883,7 +1883,7 @@ static int xgbe_tx_poll(struct xgbe_channel *channel)
static int xgbe_rx_poll(struct xgbe_channel *channel, int budget)
{
struct xgbe_prv_data *pdata = channel->pdata;
@@ -68516,7 +68516,7 @@ index 75d07fa..d766d8e 100644
struct mlx4_dev_persistent *persist = pci_get_drvdata(pdev);
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/main.c b/drivers/net/ethernet/mellanox/mlx5/core/main.c
-index 0c9ef87..c10ec50 100644
+index 7a196a0..da60bf8 100644
--- a/drivers/net/ethernet/mellanox/mlx5/core/main.c
+++ b/drivers/net/ethernet/mellanox/mlx5/core/main.c
@@ -1312,7 +1312,7 @@ static void remove_one(struct pci_dev *pdev)
@@ -70074,10 +70074,10 @@ index 51fc0c3..6cc1baa 100644
#define VIRTNET_DRIVER_VERSION "1.0.0"
diff --git a/drivers/net/vrf.c b/drivers/net/vrf.c
-index bc744ac..2abf77e 100644
+index a2afb8e..6d66a2e 100644
--- a/drivers/net/vrf.c
+++ b/drivers/net/vrf.c
-@@ -1297,7 +1297,7 @@ static const struct nla_policy vrf_nl_policy[IFLA_VRF_MAX + 1] = {
+@@ -1299,7 +1299,7 @@ static const struct nla_policy vrf_nl_policy[IFLA_VRF_MAX + 1] = {
[IFLA_VRF_TABLE] = { .type = NLA_U32 },
};
@@ -70086,7 +70086,7 @@ index bc744ac..2abf77e 100644
.kind = DRV_NAME,
.priv_size = sizeof(struct net_vrf),
-@@ -1334,7 +1334,7 @@ static int vrf_device_event(struct notifier_block *unused,
+@@ -1336,7 +1336,7 @@ static int vrf_device_event(struct notifier_block *unused,
return NOTIFY_DONE;
}
@@ -86198,7 +86198,7 @@ index 479e223..ba82b75 100644
wake_up(&usb_kill_urb_queue);
usb_put_urb(urb);
diff --git a/drivers/usb/core/hub.c b/drivers/usb/core/hub.c
-index aef81a1..cf6b268 100644
+index c28ccf1..0f884ac 100644
--- a/drivers/usb/core/hub.c
+++ b/drivers/usb/core/hub.c
@@ -26,6 +26,7 @@
@@ -86220,6 +86220,19 @@ index aef81a1..cf6b268 100644
if (hub_is_superspeed(hub->hdev))
unit_load = 150;
else
+diff --git a/drivers/usb/core/message.c b/drivers/usb/core/message.c
+index 3a47077..5cf8b9c 100644
+--- a/drivers/usb/core/message.c
++++ b/drivers/usb/core/message.c
+@@ -982,7 +982,7 @@ EXPORT_SYMBOL_GPL(usb_get_status);
+ * Return: Zero on success, or else the status code returned by the
+ * underlying usb_control_msg() call.
+ */
+-int usb_clear_halt(struct usb_device *dev, int pipe)
++int usb_clear_halt(struct usb_device *dev, unsigned int pipe)
+ {
+ int result;
+ int endp = usb_pipeendpoint(pipe);
diff --git a/drivers/usb/core/sysfs.c b/drivers/usb/core/sysfs.c
index c953a0f..54c64f4 100644
--- a/drivers/usb/core/sysfs.c
@@ -86894,6 +86907,37 @@ index 8fae28b..8b4bfec 100644
/* Dynamic bitflag definitions (us->dflags): used in set_bit() etc. */
+diff --git a/drivers/usb/usbip/stub_rx.c b/drivers/usb/usbip/stub_rx.c
+index 191b176..960b4ae 100644
+--- a/drivers/usb/usbip/stub_rx.c
++++ b/drivers/usb/usbip/stub_rx.c
+@@ -80,7 +80,7 @@ static int tweak_clear_halt_cmd(struct urb *urb)
+ struct usb_ctrlrequest *req;
+ int target_endp;
+ int target_dir;
+- int target_pipe;
++ unsigned int target_pipe;
+ int ret;
+
+ req = (struct usb_ctrlrequest *) urb->setup_packet;
+@@ -336,7 +336,7 @@ static struct stub_priv *stub_priv_alloc(struct stub_device *sdev,
+ return priv;
+ }
+
+-static int get_pipe(struct stub_device *sdev, int epnum, int dir)
++static unsigned int get_pipe(struct stub_device *sdev, int epnum, int dir)
+ {
+ struct usb_device *udev = sdev->udev;
+ struct usb_host_endpoint *ep;
+@@ -447,7 +447,7 @@ static void stub_recv_cmd_submit(struct stub_device *sdev,
+ struct stub_priv *priv;
+ struct usbip_device *ud = &sdev->ud;
+ struct usb_device *udev = sdev->udev;
+- int pipe = get_pipe(sdev, pdu->base.ep, pdu->base.direction);
++ unsigned int pipe = get_pipe(sdev, pdu->base.ep, pdu->base.direction);
+
+ priv = stub_priv_alloc(sdev, pdu);
+ if (!priv)
diff --git a/drivers/usb/usbip/vhci.h b/drivers/usb/usbip/vhci.h
index 88b71c4..31cc1ca6 100644
--- a/drivers/usb/usbip/vhci.h
@@ -87227,7 +87271,7 @@ index 9269d56..78d2a06 100644
};
EXPORT_SYMBOL_GPL(dummy_con);
diff --git a/drivers/video/console/fbcon.c b/drivers/video/console/fbcon.c
-index b87f5cf..6aad4f8 100644
+index 4db10d7..582743d 100644
--- a/drivers/video/console/fbcon.c
+++ b/drivers/video/console/fbcon.c
@@ -106,7 +106,7 @@ static int fbcon_softback_size = 32768;
@@ -107732,10 +107776,10 @@ index 42145be..1f1db90 100644
static ssize_t session_write_kbytes_show(struct ext4_attr *a,
struct ext4_sb_info *sbi, char *buf)
diff --git a/fs/ext4/xattr.c b/fs/ext4/xattr.c
-index 4448ed3..523c675 100644
+index 3eeed8f..d68ad95 100644
--- a/fs/ext4/xattr.c
+++ b/fs/ext4/xattr.c
-@@ -414,7 +414,7 @@ static int
+@@ -409,7 +409,7 @@ static int
ext4_xattr_list_entries(struct dentry *dentry, struct ext4_xattr_entry *entry,
char *buffer, size_t buffer_size)
{
@@ -107744,7 +107788,7 @@ index 4448ed3..523c675 100644
for (; !IS_LAST_ENTRY(entry); entry = EXT4_XATTR_NEXT(entry)) {
const struct xattr_handler *handler =
-@@ -435,9 +435,10 @@ ext4_xattr_list_entries(struct dentry *dentry, struct ext4_xattr_entry *entry,
+@@ -430,9 +430,10 @@ ext4_xattr_list_entries(struct dentry *dentry, struct ext4_xattr_entry *entry,
*buffer++ = 0;
}
rest -= size;
@@ -136720,7 +136764,7 @@ index ede6b97..1f5b11f 100644
int xts_crypt(struct blkcipher_desc *desc, struct scatterlist *dst,
struct scatterlist *src, unsigned int nbytes,
diff --git a/include/drm/drmP.h b/include/drm/drmP.h
-index 6726440..96d599d 100644
+index e9fb2e8..872cabe 100644
--- a/include/drm/drmP.h
+++ b/include/drm/drmP.h
@@ -61,6 +61,7 @@
@@ -136767,7 +136811,7 @@ index 6726440..96d599d 100644
/**
* Creates a driver or general drm_ioctl_desc array entry for the given
-@@ -713,7 +716,8 @@ struct drm_driver {
+@@ -714,7 +717,8 @@ struct drm_driver {
/* List of devices hanging off this driver with stealth attach. */
struct list_head legacy_dev_list;
@@ -136777,7 +136821,7 @@ index 6726440..96d599d 100644
enum drm_minor_type {
DRM_MINOR_PRIMARY,
-@@ -731,7 +735,8 @@ struct drm_info_list {
+@@ -732,7 +736,8 @@ struct drm_info_list {
int (*show)(struct seq_file*, void*); /** show callback */
u32 driver_features; /**< Required driver features for this entry */
void *data;
@@ -136787,7 +136831,7 @@ index 6726440..96d599d 100644
/**
* debugfs node structure. This structure represents a debugfs file.
-@@ -792,7 +797,7 @@ struct drm_device {
+@@ -793,7 +798,7 @@ struct drm_device {
/** \name Usage Counters */
/*@{ */
@@ -144884,7 +144928,7 @@ index 33383ca..44211d6 100644
static __always_inline void put_unaligned_le16(u16 val, void *p)
diff --git a/include/linux/usb.h b/include/linux/usb.h
-index eba1f10..94c966f 100644
+index eba1f10..eac1b52 100644
--- a/include/linux/usb.h
+++ b/include/linux/usb.h
@@ -370,7 +370,7 @@ struct usb_bus {
@@ -144905,6 +144949,15 @@ index eba1f10..94c966f 100644
unsigned long active_duration;
+@@ -1700,7 +1700,7 @@ extern int usb_string(struct usb_device *dev, int index,
+ char *buf, size_t size);
+
+ /* wrappers that also update important state inside usbcore */
+-extern int usb_clear_halt(struct usb_device *dev, int pipe);
++extern int usb_clear_halt(struct usb_device *dev, unsigned int pipe);
+ extern int usb_reset_configuration(struct usb_device *dev);
+ extern int usb_set_interface(struct usb_device *dev, int ifnum, int alternate);
+ extern void usb_reset_endpoint(struct usb_device *dev, unsigned int epaddr);
@@ -1793,10 +1793,10 @@ void usb_sg_wait(struct usb_sg_request *io);
/* NOTE: these are not the standard USB_ENDPOINT_XFER_* values!! */
@@ -144920,6 +144973,15 @@ index eba1f10..94c966f 100644
#define usb_pipein(pipe) ((pipe) & USB_DIR_IN)
#define usb_pipeout(pipe) (!usb_pipein(pipe))
+@@ -1845,7 +1845,7 @@ usb_pipe_endpoint(struct usb_device *dev, unsigned int pipe)
+ /*-------------------------------------------------------------------------*/
+
+ static inline __u16
+-usb_maxpacket(struct usb_device *udev, int pipe, int is_out)
++usb_maxpacket(struct usb_device *udev, unsigned int pipe, int is_out)
+ {
+ struct usb_host_endpoint *ep;
+ unsigned epnum = usb_pipeendpoint(pipe);
diff --git a/include/linux/usb/hcd.h b/include/linux/usb/hcd.h
index 66fc137..9602956 100644
--- a/include/linux/usb/hcd.h
@@ -152684,7 +152746,7 @@ index 154fd68..f95f804 100644
(void *)current->task_state_change,
(void *)current->task_state_change);
diff --git a/kernel/sched/deadline.c b/kernel/sched/deadline.c
-index 37e2449..61f57aa 100644
+index c95c512..16f39ee 100644
--- a/kernel/sched/deadline.c
+++ b/kernel/sched/deadline.c
@@ -219,8 +219,8 @@ static inline bool need_pull_dl_task(struct rq *rq, struct task_struct *prev)
@@ -152846,7 +152908,7 @@ index c242944..c6a1086 100644
struct rq *this_rq = this_rq();
enum cpu_idle_type idle = this_rq->idle_balance ?
diff --git a/kernel/sched/rt.c b/kernel/sched/rt.c
-index 2516b8d..251b6ab 100644
+index f139f22..c040b45 100644
--- a/kernel/sched/rt.c
+++ b/kernel/sched/rt.c
@@ -362,8 +362,8 @@ static inline int has_pushable_tasks(struct rq *rq)
@@ -158185,7 +158247,7 @@ index 66ce6b4..c5f0a41 100644
err = -EPERM;
goto out;
diff --git a/mm/mlock.c b/mm/mlock.c
-index 665ab75..41833e6 100644
+index 665ab75..70e0033 100644
--- a/mm/mlock.c
+++ b/mm/mlock.c
@@ -14,6 +14,7 @@
@@ -158231,6 +158293,20 @@ index 665ab75..41833e6 100644
newflags |= flags;
/* Here we know that vma->vm_start <= nstart < vma->vm_end. */
+@@ -629,11 +639,11 @@ static int apply_vma_lock_flags(unsigned long start, size_t len,
+ * is also counted.
+ * Return value: previously mlocked page counts
+ */
+-static int count_mm_mlocked_page_nr(struct mm_struct *mm,
++static unsigned long count_mm_mlocked_page_nr(struct mm_struct *mm,
+ unsigned long start, size_t len)
+ {
+ struct vm_area_struct *vma;
+- int count = 0;
++ unsigned long count = 0;
+
+ if (mm == NULL)
+ mm = current->mm;
@@ -695,6 +705,10 @@ static __must_check int do_mlock(unsigned long start, size_t len, vm_flags_t fla
}
@@ -164507,7 +164583,7 @@ index f0f462c..e5d59e8 100644
}
diff --git a/net/core/sock.c b/net/core/sock.c
-index bc6543f..d9e3e41 100644
+index 470a204..f5adedf 100644
--- a/net/core/sock.c
+++ b/net/core/sock.c
@@ -411,13 +411,13 @@ int __sock_queue_rcv_skb(struct sock *sk, struct sk_buff *skb)
@@ -164633,7 +164709,7 @@ index bc6543f..d9e3e41 100644
return -EFAULT;
lenout:
if (put_user(len, optlen))
-@@ -1517,7 +1520,7 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
+@@ -1522,7 +1525,7 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
newsk->sk_dst_cache = NULL;
newsk->sk_wmem_queued = 0;
newsk->sk_forward_alloc = 0;
@@ -164642,7 +164718,7 @@ index bc6543f..d9e3e41 100644
newsk->sk_send_head = NULL;
newsk->sk_userlocks = sk->sk_userlocks & ~SOCK_BINDPORT_LOCK;
-@@ -1547,7 +1550,7 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
+@@ -1558,7 +1561,7 @@ struct sock *sk_clone_lock(const struct sock *sk, const gfp_t priority)
newsk->sk_err_soft = 0;
newsk->sk_priority = 0;
newsk->sk_incoming_cpu = raw_smp_processor_id();
@@ -164651,7 +164727,7 @@ index bc6543f..d9e3e41 100644
mem_cgroup_sk_alloc(newsk);
cgroup_sk_alloc(&newsk->sk_cgrp_data);
-@@ -2477,7 +2480,7 @@ void sock_init_data(struct socket *sock, struct sock *sk)
+@@ -2488,7 +2491,7 @@ void sock_init_data(struct socket *sock, struct sock *sk)
*/
smp_wmb();
atomic_set(&sk->sk_refcnt, 1);
@@ -164660,7 +164736,7 @@ index bc6543f..d9e3e41 100644
}
EXPORT_SYMBOL(sock_init_data);
-@@ -2601,6 +2604,7 @@ void sock_enable_timestamp(struct sock *sk, int flag)
+@@ -2612,6 +2615,7 @@ void sock_enable_timestamp(struct sock *sk, int flag)
int sock_recv_errqueue(struct sock *sk, struct msghdr *msg, int len,
int level, int type)
{
@@ -164668,7 +164744,7 @@ index bc6543f..d9e3e41 100644
struct sock_exterr_skb *serr;
struct sk_buff *skb;
int copied, err;
-@@ -2622,7 +2626,8 @@ int sock_recv_errqueue(struct sock *sk, struct msghdr *msg, int len,
+@@ -2633,7 +2637,8 @@ int sock_recv_errqueue(struct sock *sk, struct msghdr *msg, int len,
sock_recv_timestamp(msg, sk, skb);
serr = SKB_EXT_ERR(skb);
@@ -164678,7 +164754,7 @@ index bc6543f..d9e3e41 100644
msg->msg_flags |= MSG_ERRQUEUE;
err = copied;
-@@ -2885,8 +2890,9 @@ static int req_prot_init(const struct proto *prot)
+@@ -2891,8 +2896,9 @@ static int req_prot_init(const struct proto *prot)
int proto_register(struct proto *prot, int alloc_slab)
{
if (alloc_slab) {
@@ -164689,7 +164765,7 @@ index bc6543f..d9e3e41 100644
NULL);
if (prot->slab == NULL) {
-@@ -3074,7 +3080,7 @@ static __net_exit void proto_exit_net(struct net *net)
+@@ -3080,7 +3086,7 @@ static __net_exit void proto_exit_net(struct net *net)
}
@@ -165264,10 +165340,10 @@ index 062a67c..cb05c97 100644
.exit = devinet_exit_net,
};
diff --git a/net/ipv4/fib_frontend.c b/net/ipv4/fib_frontend.c
-index 5b03d7f..6c62eaf 100644
+index 6789e48..d779c45 100644
--- a/net/ipv4/fib_frontend.c
+++ b/net/ipv4/fib_frontend.c
-@@ -1140,12 +1140,12 @@ static int fib_inetaddr_event(struct notifier_block *this, unsigned long event,
+@@ -1141,12 +1141,12 @@ static int fib_inetaddr_event(struct notifier_block *this, unsigned long event,
#ifdef CONFIG_IP_ROUTE_MULTIPATH
fib_sync_up(dev, RTNH_F_DEAD);
#endif
@@ -165282,7 +165358,7 @@ index 5b03d7f..6c62eaf 100644
if (!ifa->ifa_dev->ifa_list) {
/* Last address was deleted from this interface.
* Disable IP.
-@@ -1185,7 +1185,7 @@ static int fib_netdev_event(struct notifier_block *this, unsigned long event, vo
+@@ -1186,7 +1186,7 @@ static int fib_netdev_event(struct notifier_block *this, unsigned long event, vo
#ifdef CONFIG_IP_ROUTE_MULTIPATH
fib_sync_up(dev, RTNH_F_DEAD);
#endif
@@ -166119,7 +166195,7 @@ index 80bc36b..d70d622 100644
.exit = ipv4_sysctl_exit_net,
};
diff --git a/net/ipv4/tcp_input.c b/net/ipv4/tcp_input.c
-index ce42ded..9c93e33 100644
+index 7727ffe..9488999 100644
--- a/net/ipv4/tcp_input.c
+++ b/net/ipv4/tcp_input.c
@@ -288,11 +288,13 @@ static void tcp_ecn_rcv_synack(struct tcp_sock *tp, const struct tcphdr *th)
@@ -166270,7 +166346,7 @@ index bf1f3b2..83f355d 100644
.exit = tcp_net_metrics_exit,
};
diff --git a/net/ipv4/tcp_minisocks.c b/net/ipv4/tcp_minisocks.c
-index 6234eba..8007145 100644
+index 8615a6b..772fcdb 100644
--- a/net/ipv4/tcp_minisocks.c
+++ b/net/ipv4/tcp_minisocks.c
@@ -27,6 +27,10 @@
@@ -166284,7 +166360,7 @@ index 6234eba..8007145 100644
int sysctl_tcp_abort_on_overflow __read_mostly;
struct inet_timewait_death_row tcp_death_row = {
-@@ -786,7 +790,10 @@ struct sock *tcp_check_req(struct sock *sk, struct sk_buff *skb,
+@@ -787,7 +791,10 @@ struct sock *tcp_check_req(struct sock *sk, struct sk_buff *skb,
* avoid becoming vulnerable to outside attack aiming at
* resetting legit local connections.
*/
@@ -167308,7 +167384,7 @@ index b2e61a0..bf47484 100644
}
diff --git a/net/ipv6/udp.c b/net/ipv6/udp.c
-index e4a8000..ae30c92 100644
+index 40a289f..c8715aa 100644
--- a/net/ipv6/udp.c
+++ b/net/ipv6/udp.c
@@ -78,6 +78,10 @@ static u32 udp6_ehashfn(const struct net *net,
@@ -172274,7 +172350,7 @@ index 0917f04..f4e3d8c 100644
if (!proc_create("x25/route", S_IRUGO, init_net.proc_net,
diff --git a/net/xfrm/xfrm_policy.c b/net/xfrm/xfrm_policy.c
-index 5bf7e1bf..5ef3f83 100644
+index e0437a7..05fba66 100644
--- a/net/xfrm/xfrm_policy.c
+++ b/net/xfrm/xfrm_policy.c
@@ -338,7 +338,7 @@ static void xfrm_policy_kill(struct xfrm_policy *policy)
@@ -172501,10 +172577,10 @@ index 35a7e79..35847ab 100644
__xfrm_sysctl_init(net);
diff --git a/net/xfrm/xfrm_user.c b/net/xfrm/xfrm_user.c
-index 671a1d0..1b8c39e 100644
+index a7e27e1..0040091 100644
--- a/net/xfrm/xfrm_user.c
+++ b/net/xfrm/xfrm_user.c
-@@ -2471,7 +2471,7 @@ static int xfrm_user_rcv_msg(struct sk_buff *skb, struct nlmsghdr *nlh)
+@@ -2478,7 +2478,7 @@ static int xfrm_user_rcv_msg(struct sk_buff *skb, struct nlmsghdr *nlh)
return -EINVAL;
{
@@ -224530,7 +224606,7 @@ index cd0e0eb..89543da 100644
}
}
diff --git a/sound/core/seq/seq_clientmgr.c b/sound/core/seq/seq_clientmgr.c
-index 4c93520..e4032f9 100644
+index f3b1d7f..6645b81 100644
--- a/sound/core/seq/seq_clientmgr.c
+++ b/sound/core/seq/seq_clientmgr.c
@@ -403,7 +403,7 @@ static ssize_t snd_seq_read(struct file *file, char __user *buf, size_t count,
@@ -224568,7 +224644,7 @@ index 4c93520..e4032f9 100644
}
#endif
diff --git a/sound/core/seq/seq_fifo.c b/sound/core/seq/seq_fifo.c
-index 86240d0..08b468d 100644
+index 3f4efcb..da7bb45 100644
--- a/sound/core/seq/seq_fifo.c
+++ b/sound/core/seq/seq_fifo.c
@@ -50,7 +50,7 @@ struct snd_seq_fifo *snd_seq_fifo_new(int poolsize)
@@ -224580,7 +224656,7 @@ index 86240d0..08b468d 100644
f->head = NULL;
f->tail = NULL;
-@@ -96,7 +96,7 @@ void snd_seq_fifo_clear(struct snd_seq_fifo *f)
+@@ -99,7 +99,7 @@ void snd_seq_fifo_clear(struct snd_seq_fifo *f)
unsigned long flags;
/* clear overflow flag */
@@ -224589,7 +224665,7 @@ index 86240d0..08b468d 100644
snd_use_lock_sync(&f->use_lock);
spin_lock_irqsave(&f->lock, flags);
-@@ -123,7 +123,7 @@ int snd_seq_fifo_event_in(struct snd_seq_fifo *f,
+@@ -126,7 +126,7 @@ int snd_seq_fifo_event_in(struct snd_seq_fifo *f,
err = snd_seq_event_dup(f->pool, event, &cell, 1, NULL); /* always non-blocking */
if (err < 0) {
if ((err == -ENOMEM) || (err == -EAGAIN))
@@ -224612,7 +224688,7 @@ index 062c446..a4b6f4c 100644
};
diff --git a/sound/core/seq/seq_memory.c b/sound/core/seq/seq_memory.c
-index dfa5156..05c2b75 100644
+index 5847c44..cfec4ed 100644
--- a/sound/core/seq/seq_memory.c
+++ b/sound/core/seq/seq_memory.c
@@ -87,7 +87,7 @@ int snd_seq_dump_var_event(const struct snd_seq_event *event,
diff --git a/4.9.18/4425_grsec_remove_EI_PAX.patch b/4.9.20/4425_grsec_remove_EI_PAX.patch
index 594598a..594598a 100644
--- a/4.9.18/4425_grsec_remove_EI_PAX.patch
+++ b/4.9.20/4425_grsec_remove_EI_PAX.patch
diff --git a/4.9.18/4426_default_XATTR_PAX_FLAGS.patch b/4.9.20/4426_default_XATTR_PAX_FLAGS.patch
index f7e97b5..f7e97b5 100644
--- a/4.9.18/4426_default_XATTR_PAX_FLAGS.patch
+++ b/4.9.20/4426_default_XATTR_PAX_FLAGS.patch
diff --git a/4.9.18/4427_force_XATTR_PAX_tmpfs.patch b/4.9.20/4427_force_XATTR_PAX_tmpfs.patch
index 3871139..3871139 100644
--- a/4.9.18/4427_force_XATTR_PAX_tmpfs.patch
+++ b/4.9.20/4427_force_XATTR_PAX_tmpfs.patch
diff --git a/4.9.18/4430_grsec-remove-localversion-grsec.patch b/4.9.20/4430_grsec-remove-localversion-grsec.patch
index 31cf878..31cf878 100644
--- a/4.9.18/4430_grsec-remove-localversion-grsec.patch
+++ b/4.9.20/4430_grsec-remove-localversion-grsec.patch
diff --git a/4.9.18/4435_grsec-mute-warnings.patch b/4.9.20/4435_grsec-mute-warnings.patch
index 8929222..8929222 100644
--- a/4.9.18/4435_grsec-mute-warnings.patch
+++ b/4.9.20/4435_grsec-mute-warnings.patch
diff --git a/4.9.18/4440_grsec-remove-protected-paths.patch b/4.9.20/4440_grsec-remove-protected-paths.patch
index 741546d..741546d 100644
--- a/4.9.18/4440_grsec-remove-protected-paths.patch
+++ b/4.9.20/4440_grsec-remove-protected-paths.patch
diff --git a/4.9.18/4450_grsec-kconfig-default-gids.patch b/4.9.20/4450_grsec-kconfig-default-gids.patch
index cee6e27..cee6e27 100644
--- a/4.9.18/4450_grsec-kconfig-default-gids.patch
+++ b/4.9.20/4450_grsec-kconfig-default-gids.patch
diff --git a/4.9.18/4465_selinux-avc_audit-log-curr_ip.patch b/4.9.20/4465_selinux-avc_audit-log-curr_ip.patch
index 06a5294..06a5294 100644
--- a/4.9.18/4465_selinux-avc_audit-log-curr_ip.patch
+++ b/4.9.20/4465_selinux-avc_audit-log-curr_ip.patch
diff --git a/4.9.18/4470_disable-compat_vdso.patch b/4.9.20/4470_disable-compat_vdso.patch
index a1401d8..a1401d8 100644
--- a/4.9.18/4470_disable-compat_vdso.patch
+++ b/4.9.20/4470_disable-compat_vdso.patch
diff --git a/4.9.18/4475_emutramp_default_on.patch b/4.9.20/4475_emutramp_default_on.patch
index feb8c7b..feb8c7b 100644
--- a/4.9.18/4475_emutramp_default_on.patch
+++ b/4.9.20/4475_emutramp_default_on.patch