diff options
-rw-r--r-- | 0000_README | 4 | ||||
-rw-r--r-- | 1006_linux-5.13.7.patch | 781 |
2 files changed, 785 insertions, 0 deletions
diff --git a/0000_README b/0000_README index 87d634dd..79d1b681 100644 --- a/0000_README +++ b/0000_README @@ -67,6 +67,10 @@ Patch: 1005_linux-5.13.6.patch From: http://www.kernel.org Desc: Linux 5.13.6 +Patch: 1006_linux-5.13.7.patch +From: http://www.kernel.org +Desc: Linux 5.13.7 + Patch: 1500_XATTR_USER_PREFIX.patch From: https://bugs.gentoo.org/show_bug.cgi?id=470644 Desc: Support for namespace user.pax.* on tmpfs. diff --git a/1006_linux-5.13.7.patch b/1006_linux-5.13.7.patch new file mode 100644 index 00000000..840c6008 --- /dev/null +++ b/1006_linux-5.13.7.patch @@ -0,0 +1,781 @@ +diff --git a/Makefile b/Makefile +index 96967f8951933..614327400aea2 100644 +--- a/Makefile ++++ b/Makefile +@@ -1,7 +1,7 @@ + # SPDX-License-Identifier: GPL-2.0 + VERSION = 5 + PATCHLEVEL = 13 +-SUBLEVEL = 6 ++SUBLEVEL = 7 + EXTRAVERSION = + NAME = Opossums on Parade + +diff --git a/arch/arm/boot/dts/versatile-ab.dts b/arch/arm/boot/dts/versatile-ab.dts +index 37bd41ff8dffa..151c0220047dd 100644 +--- a/arch/arm/boot/dts/versatile-ab.dts ++++ b/arch/arm/boot/dts/versatile-ab.dts +@@ -195,16 +195,15 @@ + #size-cells = <1>; + ranges; + +- vic: intc@10140000 { ++ vic: interrupt-controller@10140000 { + compatible = "arm,versatile-vic"; + interrupt-controller; + #interrupt-cells = <1>; + reg = <0x10140000 0x1000>; +- clear-mask = <0xffffffff>; + valid-mask = <0xffffffff>; + }; + +- sic: intc@10003000 { ++ sic: interrupt-controller@10003000 { + compatible = "arm,versatile-sic"; + interrupt-controller; + #interrupt-cells = <1>; +diff --git a/arch/arm/boot/dts/versatile-pb.dts b/arch/arm/boot/dts/versatile-pb.dts +index 06a0fdf24026c..e7e751a858d81 100644 +--- a/arch/arm/boot/dts/versatile-pb.dts ++++ b/arch/arm/boot/dts/versatile-pb.dts +@@ -7,7 +7,7 @@ + + amba { + /* The Versatile PB is using more SIC IRQ lines than the AB */ +- sic: intc@10003000 { ++ sic: interrupt-controller@10003000 { + clear-mask = <0xffffffff>; + /* + * Valid interrupt lines mask according to +diff --git a/drivers/firmware/arm_scmi/driver.c b/drivers/firmware/arm_scmi/driver.c +index 74986bf96656f..c9fda6261c6b0 100644 +--- a/drivers/firmware/arm_scmi/driver.c ++++ b/drivers/firmware/arm_scmi/driver.c +@@ -47,7 +47,6 @@ enum scmi_error_codes { + SCMI_ERR_GENERIC = -8, /* Generic Error */ + SCMI_ERR_HARDWARE = -9, /* Hardware Error */ + SCMI_ERR_PROTOCOL = -10,/* Protocol Error */ +- SCMI_ERR_MAX + }; + + /* List of all SCMI devices active in system */ +@@ -166,8 +165,10 @@ static const int scmi_linux_errmap[] = { + + static inline int scmi_to_linux_errno(int errno) + { +- if (errno < SCMI_SUCCESS && errno > SCMI_ERR_MAX) +- return scmi_linux_errmap[-errno]; ++ int err_idx = -errno; ++ ++ if (err_idx >= SCMI_SUCCESS && err_idx < ARRAY_SIZE(scmi_linux_errmap)) ++ return scmi_linux_errmap[err_idx]; + return -EIO; + } + +@@ -1029,8 +1030,9 @@ static int __scmi_xfer_info_init(struct scmi_info *sinfo, + const struct scmi_desc *desc = sinfo->desc; + + /* Pre-allocated messages, no more than what hdr.seq can support */ +- if (WARN_ON(desc->max_msg >= MSG_TOKEN_MAX)) { +- dev_err(dev, "Maximum message of %d exceeds supported %ld\n", ++ if (WARN_ON(!desc->max_msg || desc->max_msg > MSG_TOKEN_MAX)) { ++ dev_err(dev, ++ "Invalid maximum messages %d, not in range [1 - %lu]\n", + desc->max_msg, MSG_TOKEN_MAX); + return -EINVAL; + } +diff --git a/drivers/gpu/drm/ttm/ttm_range_manager.c b/drivers/gpu/drm/ttm/ttm_range_manager.c +index 707e5c1528967..ed053fd15c90b 100644 +--- a/drivers/gpu/drm/ttm/ttm_range_manager.c ++++ b/drivers/gpu/drm/ttm/ttm_range_manager.c +@@ -146,6 +146,9 @@ int ttm_range_man_fini(struct ttm_device *bdev, + struct drm_mm *mm = &rman->mm; + int ret; + ++ if (!man) ++ return 0; ++ + ttm_resource_manager_set_used(man, false); + + ret = ttm_resource_manager_evict_all(bdev, man); +diff --git a/drivers/nvme/host/pci.c b/drivers/nvme/host/pci.c +index fb1c5ae0da39d..d963f25fc7aed 100644 +--- a/drivers/nvme/host/pci.c ++++ b/drivers/nvme/host/pci.c +@@ -1562,6 +1562,28 @@ static void nvme_init_queue(struct nvme_queue *nvmeq, u16 qid) + wmb(); /* ensure the first interrupt sees the initialization */ + } + ++/* ++ * Try getting shutdown_lock while setting up IO queues. ++ */ ++static int nvme_setup_io_queues_trylock(struct nvme_dev *dev) ++{ ++ /* ++ * Give up if the lock is being held by nvme_dev_disable. ++ */ ++ if (!mutex_trylock(&dev->shutdown_lock)) ++ return -ENODEV; ++ ++ /* ++ * Controller is in wrong state, fail early. ++ */ ++ if (dev->ctrl.state != NVME_CTRL_CONNECTING) { ++ mutex_unlock(&dev->shutdown_lock); ++ return -ENODEV; ++ } ++ ++ return 0; ++} ++ + static int nvme_create_queue(struct nvme_queue *nvmeq, int qid, bool polled) + { + struct nvme_dev *dev = nvmeq->dev; +@@ -1590,8 +1612,11 @@ static int nvme_create_queue(struct nvme_queue *nvmeq, int qid, bool polled) + goto release_cq; + + nvmeq->cq_vector = vector; +- nvme_init_queue(nvmeq, qid); + ++ result = nvme_setup_io_queues_trylock(dev); ++ if (result) ++ return result; ++ nvme_init_queue(nvmeq, qid); + if (!polled) { + result = queue_request_irq(nvmeq); + if (result < 0) +@@ -1599,10 +1624,12 @@ static int nvme_create_queue(struct nvme_queue *nvmeq, int qid, bool polled) + } + + set_bit(NVMEQ_ENABLED, &nvmeq->flags); ++ mutex_unlock(&dev->shutdown_lock); + return result; + + release_sq: + dev->online_queues--; ++ mutex_unlock(&dev->shutdown_lock); + adapter_delete_sq(dev, qid); + release_cq: + adapter_delete_cq(dev, qid); +@@ -2176,7 +2203,18 @@ static int nvme_setup_io_queues(struct nvme_dev *dev) + if (nr_io_queues == 0) + return 0; + +- clear_bit(NVMEQ_ENABLED, &adminq->flags); ++ /* ++ * Free IRQ resources as soon as NVMEQ_ENABLED bit transitions ++ * from set to unset. If there is a window to it is truely freed, ++ * pci_free_irq_vectors() jumping into this window will crash. ++ * And take lock to avoid racing with pci_free_irq_vectors() in ++ * nvme_dev_disable() path. ++ */ ++ result = nvme_setup_io_queues_trylock(dev); ++ if (result) ++ return result; ++ if (test_and_clear_bit(NVMEQ_ENABLED, &adminq->flags)) ++ pci_free_irq(pdev, 0, adminq); + + if (dev->cmb_use_sqes) { + result = nvme_cmb_qdepth(dev, nr_io_queues, +@@ -2192,14 +2230,17 @@ static int nvme_setup_io_queues(struct nvme_dev *dev) + result = nvme_remap_bar(dev, size); + if (!result) + break; +- if (!--nr_io_queues) +- return -ENOMEM; ++ if (!--nr_io_queues) { ++ result = -ENOMEM; ++ goto out_unlock; ++ } + } while (1); + adminq->q_db = dev->dbs; + + retry: + /* Deregister the admin queue's interrupt */ +- pci_free_irq(pdev, 0, adminq); ++ if (test_and_clear_bit(NVMEQ_ENABLED, &adminq->flags)) ++ pci_free_irq(pdev, 0, adminq); + + /* + * If we enable msix early due to not intx, disable it again before +@@ -2208,8 +2249,10 @@ static int nvme_setup_io_queues(struct nvme_dev *dev) + pci_free_irq_vectors(pdev); + + result = nvme_setup_irqs(dev, nr_io_queues); +- if (result <= 0) +- return -EIO; ++ if (result <= 0) { ++ result = -EIO; ++ goto out_unlock; ++ } + + dev->num_vecs = result; + result = max(result - 1, 1); +@@ -2223,8 +2266,9 @@ static int nvme_setup_io_queues(struct nvme_dev *dev) + */ + result = queue_request_irq(adminq); + if (result) +- return result; ++ goto out_unlock; + set_bit(NVMEQ_ENABLED, &adminq->flags); ++ mutex_unlock(&dev->shutdown_lock); + + result = nvme_create_io_queues(dev); + if (result || dev->online_queues < 2) +@@ -2233,6 +2277,9 @@ static int nvme_setup_io_queues(struct nvme_dev *dev) + if (dev->online_queues - 1 < dev->max_qid) { + nr_io_queues = dev->online_queues - 1; + nvme_disable_io_queues(dev); ++ result = nvme_setup_io_queues_trylock(dev); ++ if (result) ++ return result; + nvme_suspend_io_queues(dev); + goto retry; + } +@@ -2241,6 +2288,9 @@ static int nvme_setup_io_queues(struct nvme_dev *dev) + dev->io_queues[HCTX_TYPE_READ], + dev->io_queues[HCTX_TYPE_POLL]); + return 0; ++out_unlock: ++ mutex_unlock(&dev->shutdown_lock); ++ return result; + } + + static void nvme_del_queue_end(struct request *req, blk_status_t error) +diff --git a/fs/cifs/smb2ops.c b/fs/cifs/smb2ops.c +index 64cad843ce723..398c941e38974 100644 +--- a/fs/cifs/smb2ops.c ++++ b/fs/cifs/smb2ops.c +@@ -555,8 +555,8 @@ parse_server_interfaces(struct network_interface_info_ioctl_rsp *buf, + p = buf; + while (bytes_left >= sizeof(*p)) { + info->speed = le64_to_cpu(p->LinkSpeed); +- info->rdma_capable = le32_to_cpu(p->Capability & RDMA_CAPABLE); +- info->rss_capable = le32_to_cpu(p->Capability & RSS_CAPABLE); ++ info->rdma_capable = le32_to_cpu(p->Capability & RDMA_CAPABLE) ? 1 : 0; ++ info->rss_capable = le32_to_cpu(p->Capability & RSS_CAPABLE) ? 1 : 0; + + cifs_dbg(FYI, "%s: adding iface %zu\n", __func__, *iface_count); + cifs_dbg(FYI, "%s: speed %zu bps\n", __func__, info->speed); +diff --git a/fs/hfs/bfind.c b/fs/hfs/bfind.c +index 4af318fbda774..ef9498a6e88ac 100644 +--- a/fs/hfs/bfind.c ++++ b/fs/hfs/bfind.c +@@ -25,7 +25,19 @@ int hfs_find_init(struct hfs_btree *tree, struct hfs_find_data *fd) + fd->key = ptr + tree->max_key_len + 2; + hfs_dbg(BNODE_REFS, "find_init: %d (%p)\n", + tree->cnid, __builtin_return_address(0)); +- mutex_lock(&tree->tree_lock); ++ switch (tree->cnid) { ++ case HFS_CAT_CNID: ++ mutex_lock_nested(&tree->tree_lock, CATALOG_BTREE_MUTEX); ++ break; ++ case HFS_EXT_CNID: ++ mutex_lock_nested(&tree->tree_lock, EXTENTS_BTREE_MUTEX); ++ break; ++ case HFS_ATTR_CNID: ++ mutex_lock_nested(&tree->tree_lock, ATTR_BTREE_MUTEX); ++ break; ++ default: ++ return -EINVAL; ++ } + return 0; + } + +diff --git a/fs/hfs/bnode.c b/fs/hfs/bnode.c +index b63a4df7327b6..c0a73a6ffb28b 100644 +--- a/fs/hfs/bnode.c ++++ b/fs/hfs/bnode.c +@@ -15,16 +15,31 @@ + + #include "btree.h" + +-void hfs_bnode_read(struct hfs_bnode *node, void *buf, +- int off, int len) ++void hfs_bnode_read(struct hfs_bnode *node, void *buf, int off, int len) + { + struct page *page; ++ int pagenum; ++ int bytes_read; ++ int bytes_to_read; ++ void *vaddr; + + off += node->page_offset; +- page = node->page[0]; ++ pagenum = off >> PAGE_SHIFT; ++ off &= ~PAGE_MASK; /* compute page offset for the first page */ + +- memcpy(buf, kmap(page) + off, len); +- kunmap(page); ++ for (bytes_read = 0; bytes_read < len; bytes_read += bytes_to_read) { ++ if (pagenum >= node->tree->pages_per_bnode) ++ break; ++ page = node->page[pagenum]; ++ bytes_to_read = min_t(int, len - bytes_read, PAGE_SIZE - off); ++ ++ vaddr = kmap_atomic(page); ++ memcpy(buf + bytes_read, vaddr + off, bytes_to_read); ++ kunmap_atomic(vaddr); ++ ++ pagenum++; ++ off = 0; /* page offset only applies to the first page */ ++ } + } + + u16 hfs_bnode_read_u16(struct hfs_bnode *node, int off) +diff --git a/fs/hfs/btree.h b/fs/hfs/btree.h +index 4ba45caf59392..0e6baee932453 100644 +--- a/fs/hfs/btree.h ++++ b/fs/hfs/btree.h +@@ -13,6 +13,13 @@ typedef int (*btree_keycmp)(const btree_key *, const btree_key *); + + #define NODE_HASH_SIZE 256 + ++/* B-tree mutex nested subclasses */ ++enum hfs_btree_mutex_classes { ++ CATALOG_BTREE_MUTEX, ++ EXTENTS_BTREE_MUTEX, ++ ATTR_BTREE_MUTEX, ++}; ++ + /* A HFS BTree held in memory */ + struct hfs_btree { + struct super_block *sb; +diff --git a/fs/hfs/super.c b/fs/hfs/super.c +index 44d07c9e3a7f0..12d9bae393631 100644 +--- a/fs/hfs/super.c ++++ b/fs/hfs/super.c +@@ -420,14 +420,12 @@ static int hfs_fill_super(struct super_block *sb, void *data, int silent) + if (!res) { + if (fd.entrylength > sizeof(rec) || fd.entrylength < 0) { + res = -EIO; +- goto bail; ++ goto bail_hfs_find; + } + hfs_bnode_read(fd.bnode, &rec, fd.entryoffset, fd.entrylength); + } +- if (res) { +- hfs_find_exit(&fd); +- goto bail_no_root; +- } ++ if (res) ++ goto bail_hfs_find; + res = -EINVAL; + root_inode = hfs_iget(sb, &fd.search_key->cat, &rec); + hfs_find_exit(&fd); +@@ -443,6 +441,8 @@ static int hfs_fill_super(struct super_block *sb, void *data, int silent) + /* everything's okay */ + return 0; + ++bail_hfs_find: ++ hfs_find_exit(&fd); + bail_no_root: + pr_err("get root inode failed\n"); + bail: +diff --git a/fs/internal.h b/fs/internal.h +index 6aeae7ef33803..728f8d70d7f1d 100644 +--- a/fs/internal.h ++++ b/fs/internal.h +@@ -61,7 +61,6 @@ extern void __init chrdev_init(void); + */ + extern const struct fs_context_operations legacy_fs_context_ops; + extern int parse_monolithic_mount_data(struct fs_context *, void *); +-extern void fc_drop_locked(struct fs_context *); + extern void vfs_clean_context(struct fs_context *fc); + extern int finish_clean_context(struct fs_context *fc); + +diff --git a/fs/iomap/seek.c b/fs/iomap/seek.c +index dab1b02eba5b7..ce6fb810854fe 100644 +--- a/fs/iomap/seek.c ++++ b/fs/iomap/seek.c +@@ -35,23 +35,20 @@ loff_t + iomap_seek_hole(struct inode *inode, loff_t offset, const struct iomap_ops *ops) + { + loff_t size = i_size_read(inode); +- loff_t length = size - offset; + loff_t ret; + + /* Nothing to be found before or beyond the end of the file. */ + if (offset < 0 || offset >= size) + return -ENXIO; + +- while (length > 0) { +- ret = iomap_apply(inode, offset, length, IOMAP_REPORT, ops, +- &offset, iomap_seek_hole_actor); ++ while (offset < size) { ++ ret = iomap_apply(inode, offset, size - offset, IOMAP_REPORT, ++ ops, &offset, iomap_seek_hole_actor); + if (ret < 0) + return ret; + if (ret == 0) + break; +- + offset += ret; +- length -= ret; + } + + return offset; +@@ -83,27 +80,23 @@ loff_t + iomap_seek_data(struct inode *inode, loff_t offset, const struct iomap_ops *ops) + { + loff_t size = i_size_read(inode); +- loff_t length = size - offset; + loff_t ret; + + /* Nothing to be found before or beyond the end of the file. */ + if (offset < 0 || offset >= size) + return -ENXIO; + +- while (length > 0) { +- ret = iomap_apply(inode, offset, length, IOMAP_REPORT, ops, +- &offset, iomap_seek_data_actor); ++ while (offset < size) { ++ ret = iomap_apply(inode, offset, size - offset, IOMAP_REPORT, ++ ops, &offset, iomap_seek_data_actor); + if (ret < 0) + return ret; + if (ret == 0) +- break; +- ++ return offset; + offset += ret; +- length -= ret; + } + +- if (length <= 0) +- return -ENXIO; +- return offset; ++ /* We've reached the end of the file without finding data */ ++ return -ENXIO; + } + EXPORT_SYMBOL_GPL(iomap_seek_data); +diff --git a/include/linux/fs_context.h b/include/linux/fs_context.h +index 37e1e8f7f08da..5b44b0195a28a 100644 +--- a/include/linux/fs_context.h ++++ b/include/linux/fs_context.h +@@ -139,6 +139,7 @@ extern int vfs_parse_fs_string(struct fs_context *fc, const char *key, + extern int generic_parse_monolithic(struct fs_context *fc, void *data); + extern int vfs_get_tree(struct fs_context *fc); + extern void put_fs_context(struct fs_context *fc); ++extern void fc_drop_locked(struct fs_context *fc); + + /* + * sget() wrappers to be called from the ->get_tree() op. +diff --git a/include/net/busy_poll.h b/include/net/busy_poll.h +index 73af4a64a5999..40296ed976a97 100644 +--- a/include/net/busy_poll.h ++++ b/include/net/busy_poll.h +@@ -38,7 +38,7 @@ static inline bool net_busy_loop_on(void) + + static inline bool sk_can_busy_loop(const struct sock *sk) + { +- return sk->sk_ll_usec && !signal_pending(current); ++ return READ_ONCE(sk->sk_ll_usec) && !signal_pending(current); + } + + bool sk_busy_loop_end(void *p, unsigned long start_time); +diff --git a/include/net/sctp/constants.h b/include/net/sctp/constants.h +index 14a0d22c91133..bf23a2ed92da8 100644 +--- a/include/net/sctp/constants.h ++++ b/include/net/sctp/constants.h +@@ -342,8 +342,7 @@ enum { + #define SCTP_SCOPE_POLICY_MAX SCTP_SCOPE_POLICY_LINK + + /* Based on IPv4 scoping <draft-stewart-tsvwg-sctp-ipv4-00.txt>, +- * SCTP IPv4 unusable addresses: 0.0.0.0/8, 224.0.0.0/4, 198.18.0.0/24, +- * 192.88.99.0/24. ++ * SCTP IPv4 unusable addresses: 0.0.0.0/8, 224.0.0.0/4, 192.88.99.0/24. + * Also, RFC 8.4, non-unicast addresses are not considered valid SCTP + * addresses. + */ +@@ -351,7 +350,6 @@ enum { + ((htonl(INADDR_BROADCAST) == a) || \ + ipv4_is_multicast(a) || \ + ipv4_is_zeronet(a) || \ +- ipv4_is_test_198(a) || \ + ipv4_is_anycast_6to4(a)) + + /* Flags used for the bind address copy functions. */ +diff --git a/kernel/cgroup/cgroup-v1.c b/kernel/cgroup/cgroup-v1.c +index d189eff4c92ff..583790d2060ce 100644 +--- a/kernel/cgroup/cgroup-v1.c ++++ b/kernel/cgroup/cgroup-v1.c +@@ -1225,9 +1225,7 @@ int cgroup1_get_tree(struct fs_context *fc) + ret = cgroup_do_get_tree(fc); + + if (!ret && percpu_ref_is_dying(&ctx->root->cgrp.self.refcnt)) { +- struct super_block *sb = fc->root->d_sb; +- dput(fc->root); +- deactivate_locked_super(sb); ++ fc_drop_locked(fc); + ret = 1; + } + +diff --git a/kernel/rcu/tasks.h b/kernel/rcu/tasks.h +index 350ebf5051f97..fcef5f0c60b8b 100644 +--- a/kernel/rcu/tasks.h ++++ b/kernel/rcu/tasks.h +@@ -908,10 +908,9 @@ static bool trc_inspect_reader(struct task_struct *t, void *arg) + in_qs = likely(!t->trc_reader_nesting); + } + +- // Mark as checked. Because this is called from the grace-period +- // kthread, also remove the task from the holdout list. ++ // Mark as checked so that the grace-period kthread will ++ // remove it from the holdout list. + t->trc_reader_checked = true; +- trc_del_holdout(t); + + if (in_qs) + return true; // Already in quiescent state, done!!! +@@ -938,7 +937,6 @@ static void trc_wait_for_one_reader(struct task_struct *t, + // The current task had better be in a quiescent state. + if (t == current) { + t->trc_reader_checked = true; +- trc_del_holdout(t); + WARN_ON_ONCE(t->trc_reader_nesting); + return; + } +diff --git a/kernel/workqueue.c b/kernel/workqueue.c +index 50142fc08902d..f148eacda55a9 100644 +--- a/kernel/workqueue.c ++++ b/kernel/workqueue.c +@@ -3676,15 +3676,21 @@ static void pwq_unbound_release_workfn(struct work_struct *work) + unbound_release_work); + struct workqueue_struct *wq = pwq->wq; + struct worker_pool *pool = pwq->pool; +- bool is_last; ++ bool is_last = false; + +- if (WARN_ON_ONCE(!(wq->flags & WQ_UNBOUND))) +- return; ++ /* ++ * when @pwq is not linked, it doesn't hold any reference to the ++ * @wq, and @wq is invalid to access. ++ */ ++ if (!list_empty(&pwq->pwqs_node)) { ++ if (WARN_ON_ONCE(!(wq->flags & WQ_UNBOUND))) ++ return; + +- mutex_lock(&wq->mutex); +- list_del_rcu(&pwq->pwqs_node); +- is_last = list_empty(&wq->pwqs); +- mutex_unlock(&wq->mutex); ++ mutex_lock(&wq->mutex); ++ list_del_rcu(&pwq->pwqs_node); ++ is_last = list_empty(&wq->pwqs); ++ mutex_unlock(&wq->mutex); ++ } + + mutex_lock(&wq_pool_mutex); + put_unbound_pool(pool); +diff --git a/net/802/garp.c b/net/802/garp.c +index 400bd857e5f57..f6012f8e59f00 100644 +--- a/net/802/garp.c ++++ b/net/802/garp.c +@@ -203,6 +203,19 @@ static void garp_attr_destroy(struct garp_applicant *app, struct garp_attr *attr + kfree(attr); + } + ++static void garp_attr_destroy_all(struct garp_applicant *app) ++{ ++ struct rb_node *node, *next; ++ struct garp_attr *attr; ++ ++ for (node = rb_first(&app->gid); ++ next = node ? rb_next(node) : NULL, node != NULL; ++ node = next) { ++ attr = rb_entry(node, struct garp_attr, node); ++ garp_attr_destroy(app, attr); ++ } ++} ++ + static int garp_pdu_init(struct garp_applicant *app) + { + struct sk_buff *skb; +@@ -609,6 +622,7 @@ void garp_uninit_applicant(struct net_device *dev, struct garp_application *appl + + spin_lock_bh(&app->lock); + garp_gid_event(app, GARP_EVENT_TRANSMIT_PDU); ++ garp_attr_destroy_all(app); + garp_pdu_queue(app); + spin_unlock_bh(&app->lock); + +diff --git a/net/802/mrp.c b/net/802/mrp.c +index bea6e43d45a0d..35e04cc5390c4 100644 +--- a/net/802/mrp.c ++++ b/net/802/mrp.c +@@ -292,6 +292,19 @@ static void mrp_attr_destroy(struct mrp_applicant *app, struct mrp_attr *attr) + kfree(attr); + } + ++static void mrp_attr_destroy_all(struct mrp_applicant *app) ++{ ++ struct rb_node *node, *next; ++ struct mrp_attr *attr; ++ ++ for (node = rb_first(&app->mad); ++ next = node ? rb_next(node) : NULL, node != NULL; ++ node = next) { ++ attr = rb_entry(node, struct mrp_attr, node); ++ mrp_attr_destroy(app, attr); ++ } ++} ++ + static int mrp_pdu_init(struct mrp_applicant *app) + { + struct sk_buff *skb; +@@ -895,6 +908,7 @@ void mrp_uninit_applicant(struct net_device *dev, struct mrp_application *appl) + + spin_lock_bh(&app->lock); + mrp_mad_event(app, MRP_EVENT_TX); ++ mrp_attr_destroy_all(app); + mrp_pdu_queue(app); + spin_unlock_bh(&app->lock); + +diff --git a/net/core/sock.c b/net/core/sock.c +index 2003c5ebb4c2e..37d732fe3fcf9 100644 +--- a/net/core/sock.c ++++ b/net/core/sock.c +@@ -1172,7 +1172,7 @@ set_sndbuf: + if (val < 0) + ret = -EINVAL; + else +- sk->sk_ll_usec = val; ++ WRITE_ONCE(sk->sk_ll_usec, val); + } + break; + case SO_PREFER_BUSY_POLL: +diff --git a/net/ipv6/ip6_output.c b/net/ipv6/ip6_output.c +index b7ffb4f227a45..6062ad1d5b510 100644 +--- a/net/ipv6/ip6_output.c ++++ b/net/ipv6/ip6_output.c +@@ -60,10 +60,38 @@ static int ip6_finish_output2(struct net *net, struct sock *sk, struct sk_buff * + { + struct dst_entry *dst = skb_dst(skb); + struct net_device *dev = dst->dev; ++ unsigned int hh_len = LL_RESERVED_SPACE(dev); ++ int delta = hh_len - skb_headroom(skb); + const struct in6_addr *nexthop; + struct neighbour *neigh; + int ret; + ++ /* Be paranoid, rather than too clever. */ ++ if (unlikely(delta > 0) && dev->header_ops) { ++ /* pskb_expand_head() might crash, if skb is shared */ ++ if (skb_shared(skb)) { ++ struct sk_buff *nskb = skb_clone(skb, GFP_ATOMIC); ++ ++ if (likely(nskb)) { ++ if (skb->sk) ++ skb_set_owner_w(nskb, skb->sk); ++ consume_skb(skb); ++ } else { ++ kfree_skb(skb); ++ } ++ skb = nskb; ++ } ++ if (skb && ++ pskb_expand_head(skb, SKB_DATA_ALIGN(delta), 0, GFP_ATOMIC)) { ++ kfree_skb(skb); ++ skb = NULL; ++ } ++ if (!skb) { ++ IP6_INC_STATS(net, ip6_dst_idev(dst), IPSTATS_MIB_OUTDISCARDS); ++ return -ENOMEM; ++ } ++ } ++ + if (ipv6_addr_is_multicast(&ipv6_hdr(skb)->daddr)) { + struct inet6_dev *idev = ip6_dst_idev(skb_dst(skb)); + +diff --git a/net/sctp/protocol.c b/net/sctp/protocol.c +index 25192b378e2ec..9b444df5e53ee 100644 +--- a/net/sctp/protocol.c ++++ b/net/sctp/protocol.c +@@ -398,7 +398,8 @@ static enum sctp_scope sctp_v4_scope(union sctp_addr *addr) + retval = SCTP_SCOPE_LINK; + } else if (ipv4_is_private_10(addr->v4.sin_addr.s_addr) || + ipv4_is_private_172(addr->v4.sin_addr.s_addr) || +- ipv4_is_private_192(addr->v4.sin_addr.s_addr)) { ++ ipv4_is_private_192(addr->v4.sin_addr.s_addr) || ++ ipv4_is_test_198(addr->v4.sin_addr.s_addr)) { + retval = SCTP_SCOPE_PRIVATE; + } else { + retval = SCTP_SCOPE_GLOBAL; +diff --git a/net/unix/af_unix.c b/net/unix/af_unix.c +index 5d1192ceb1397..68a9591d0144c 100644 +--- a/net/unix/af_unix.c ++++ b/net/unix/af_unix.c +@@ -1522,6 +1522,53 @@ out: + return err; + } + ++static void unix_peek_fds(struct scm_cookie *scm, struct sk_buff *skb) ++{ ++ scm->fp = scm_fp_dup(UNIXCB(skb).fp); ++ ++ /* ++ * Garbage collection of unix sockets starts by selecting a set of ++ * candidate sockets which have reference only from being in flight ++ * (total_refs == inflight_refs). This condition is checked once during ++ * the candidate collection phase, and candidates are marked as such, so ++ * that non-candidates can later be ignored. While inflight_refs is ++ * protected by unix_gc_lock, total_refs (file count) is not, hence this ++ * is an instantaneous decision. ++ * ++ * Once a candidate, however, the socket must not be reinstalled into a ++ * file descriptor while the garbage collection is in progress. ++ * ++ * If the above conditions are met, then the directed graph of ++ * candidates (*) does not change while unix_gc_lock is held. ++ * ++ * Any operations that changes the file count through file descriptors ++ * (dup, close, sendmsg) does not change the graph since candidates are ++ * not installed in fds. ++ * ++ * Dequeing a candidate via recvmsg would install it into an fd, but ++ * that takes unix_gc_lock to decrement the inflight count, so it's ++ * serialized with garbage collection. ++ * ++ * MSG_PEEK is special in that it does not change the inflight count, ++ * yet does install the socket into an fd. The following lock/unlock ++ * pair is to ensure serialization with garbage collection. It must be ++ * done between incrementing the file count and installing the file into ++ * an fd. ++ * ++ * If garbage collection starts after the barrier provided by the ++ * lock/unlock, then it will see the elevated refcount and not mark this ++ * as a candidate. If a garbage collection is already in progress ++ * before the file count was incremented, then the lock/unlock pair will ++ * ensure that garbage collection is finished before progressing to ++ * installing the fd. ++ * ++ * (*) A -> B where B is on the queue of A or B is on the queue of C ++ * which is on the queue of listening socket A. ++ */ ++ spin_lock(&unix_gc_lock); ++ spin_unlock(&unix_gc_lock); ++} ++ + static int unix_scm_to_skb(struct scm_cookie *scm, struct sk_buff *skb, bool send_fds) + { + int err = 0; +@@ -2171,7 +2218,7 @@ static int unix_dgram_recvmsg(struct socket *sock, struct msghdr *msg, + sk_peek_offset_fwd(sk, size); + + if (UNIXCB(skb).fp) +- scm.fp = scm_fp_dup(UNIXCB(skb).fp); ++ unix_peek_fds(&scm, skb); + } + err = (flags & MSG_TRUNC) ? skb->len - skip : size; + +@@ -2414,7 +2461,7 @@ unlock: + /* It is questionable, see note in unix_dgram_recvmsg. + */ + if (UNIXCB(skb).fp) +- scm.fp = scm_fp_dup(UNIXCB(skb).fp); ++ unix_peek_fds(&scm, skb); + + sk_peek_offset_fwd(sk, chunk); + |