mirror of
https://github.com/Fishwaldo/build.git
synced 2025-04-01 03:31:36 +00:00
1625 lines
52 KiB
Diff
1625 lines
52 KiB
Diff
diff --git a/Makefile b/Makefile
|
|
index b8b8d33eab55..8d891c66803c 100644
|
|
--- a/Makefile
|
|
+++ b/Makefile
|
|
@@ -1,6 +1,6 @@
|
|
VERSION = 3
|
|
PATCHLEVEL = 10
|
|
-SUBLEVEL = 49
|
|
+SUBLEVEL = 50
|
|
EXTRAVERSION =
|
|
NAME = TOSSUG Baby Fish
|
|
|
|
diff --git a/arch/arc/include/uapi/asm/ptrace.h b/arch/arc/include/uapi/asm/ptrace.h
|
|
index 30333cec0fef..ef9d79a3db25 100644
|
|
--- a/arch/arc/include/uapi/asm/ptrace.h
|
|
+++ b/arch/arc/include/uapi/asm/ptrace.h
|
|
@@ -11,6 +11,7 @@
|
|
#ifndef _UAPI__ASM_ARC_PTRACE_H
|
|
#define _UAPI__ASM_ARC_PTRACE_H
|
|
|
|
+#define PTRACE_GET_THREAD_AREA 25
|
|
|
|
#ifndef __ASSEMBLY__
|
|
/*
|
|
diff --git a/arch/arc/kernel/ptrace.c b/arch/arc/kernel/ptrace.c
|
|
index 0851604bb9cd..f8a36ed9e0d5 100644
|
|
--- a/arch/arc/kernel/ptrace.c
|
|
+++ b/arch/arc/kernel/ptrace.c
|
|
@@ -136,6 +136,10 @@ long arch_ptrace(struct task_struct *child, long request,
|
|
pr_debug("REQ=%ld: ADDR =0x%lx, DATA=0x%lx)\n", request, addr, data);
|
|
|
|
switch (request) {
|
|
+ case PTRACE_GET_THREAD_AREA:
|
|
+ ret = put_user(task_thread_info(child)->thr_ptr,
|
|
+ (unsigned long __user *)data);
|
|
+ break;
|
|
default:
|
|
ret = ptrace_request(child, request, addr, data);
|
|
break;
|
|
diff --git a/arch/arm/Kconfig b/arch/arm/Kconfig
|
|
index 18a9f5ef643a..d41951246cd6 100644
|
|
--- a/arch/arm/Kconfig
|
|
+++ b/arch/arm/Kconfig
|
|
@@ -4,6 +4,7 @@ config ARM
|
|
select ARCH_BINFMT_ELF_RANDOMIZE_PIE
|
|
select ARCH_HAS_ATOMIC64_DEC_IF_POSITIVE
|
|
select ARCH_HAVE_CUSTOM_GPIO_H
|
|
+ select ARCH_SUPPORTS_ATOMIC_RMW
|
|
select ARCH_HAS_TICK_BROADCAST if GENERIC_CLOCKEVENTS_BROADCAST
|
|
select ARCH_WANT_IPC_PARSE_VERSION
|
|
select BUILDTIME_EXTABLE_SORT if MMU
|
|
diff --git a/arch/powerpc/Kconfig b/arch/powerpc/Kconfig
|
|
index fe404e77246e..7f656f119ea6 100644
|
|
--- a/arch/powerpc/Kconfig
|
|
+++ b/arch/powerpc/Kconfig
|
|
@@ -138,6 +138,7 @@ config PPC
|
|
select ARCH_USE_BUILTIN_BSWAP
|
|
select OLD_SIGSUSPEND
|
|
select OLD_SIGACTION if PPC32
|
|
+ select ARCH_SUPPORTS_ATOMIC_RMW
|
|
|
|
config EARLY_PRINTK
|
|
bool
|
|
diff --git a/arch/sparc/Kconfig b/arch/sparc/Kconfig
|
|
index 2668b3142fa2..03a1bc3c3dde 100644
|
|
--- a/arch/sparc/Kconfig
|
|
+++ b/arch/sparc/Kconfig
|
|
@@ -77,6 +77,7 @@ config SPARC64
|
|
select ARCH_HAVE_NMI_SAFE_CMPXCHG
|
|
select HAVE_C_RECORDMCOUNT
|
|
select NO_BOOTMEM
|
|
+ select ARCH_SUPPORTS_ATOMIC_RMW
|
|
|
|
config ARCH_DEFCONFIG
|
|
string
|
|
diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig
|
|
index fe120da25625..af88b27ce313 100644
|
|
--- a/arch/x86/Kconfig
|
|
+++ b/arch/x86/Kconfig
|
|
@@ -121,6 +121,7 @@ config X86
|
|
select OLD_SIGACTION if X86_32
|
|
select COMPAT_OLD_SIGACTION if IA32_EMULATION
|
|
select RTC_LIB
|
|
+ select ARCH_SUPPORTS_ATOMIC_RMW
|
|
|
|
config INSTRUCTION_DECODER
|
|
def_bool y
|
|
diff --git a/arch/x86/kernel/cpu/perf_event_intel.c b/arch/x86/kernel/cpu/perf_event_intel.c
|
|
index a9e22073bd56..b45ac6affa9c 100644
|
|
--- a/arch/x86/kernel/cpu/perf_event_intel.c
|
|
+++ b/arch/x86/kernel/cpu/perf_event_intel.c
|
|
@@ -1199,6 +1199,15 @@ again:
|
|
intel_pmu_lbr_read();
|
|
|
|
/*
|
|
+ * CondChgd bit 63 doesn't mean any overflow status. Ignore
|
|
+ * and clear the bit.
|
|
+ */
|
|
+ if (__test_and_clear_bit(63, (unsigned long *)&status)) {
|
|
+ if (!status)
|
|
+ goto done;
|
|
+ }
|
|
+
|
|
+ /*
|
|
* PEBS overflow sets bit 62 in the global status register
|
|
*/
|
|
if (__test_and_clear_bit(62, (unsigned long *)&status)) {
|
|
diff --git a/drivers/bluetooth/hci_h5.c b/drivers/bluetooth/hci_h5.c
|
|
index b6154d5a07a5..db0be2fb05fe 100644
|
|
--- a/drivers/bluetooth/hci_h5.c
|
|
+++ b/drivers/bluetooth/hci_h5.c
|
|
@@ -406,6 +406,7 @@ static int h5_rx_3wire_hdr(struct hci_uart *hu, unsigned char c)
|
|
H5_HDR_PKT_TYPE(hdr) != HCI_3WIRE_LINK_PKT) {
|
|
BT_ERR("Non-link packet received in non-active state");
|
|
h5_reset_rx(h5);
|
|
+ return 0;
|
|
}
|
|
|
|
h5->rx_func = h5_rx_payload;
|
|
diff --git a/drivers/gpu/drm/qxl/qxl_irq.c b/drivers/gpu/drm/qxl/qxl_irq.c
|
|
index 21393dc4700a..f4b6b89b98f3 100644
|
|
--- a/drivers/gpu/drm/qxl/qxl_irq.c
|
|
+++ b/drivers/gpu/drm/qxl/qxl_irq.c
|
|
@@ -33,6 +33,9 @@ irqreturn_t qxl_irq_handler(DRM_IRQ_ARGS)
|
|
|
|
pending = xchg(&qdev->ram_header->int_pending, 0);
|
|
|
|
+ if (!pending)
|
|
+ return IRQ_NONE;
|
|
+
|
|
atomic_inc(&qdev->irq_received);
|
|
|
|
if (pending & QXL_INTERRUPT_DISPLAY) {
|
|
diff --git a/drivers/gpu/drm/radeon/atombios_encoders.c b/drivers/gpu/drm/radeon/atombios_encoders.c
|
|
index 5802d7486354..1b564d7e4191 100644
|
|
--- a/drivers/gpu/drm/radeon/atombios_encoders.c
|
|
+++ b/drivers/gpu/drm/radeon/atombios_encoders.c
|
|
@@ -183,7 +183,6 @@ void radeon_atom_backlight_init(struct radeon_encoder *radeon_encoder,
|
|
struct backlight_properties props;
|
|
struct radeon_backlight_privdata *pdata;
|
|
struct radeon_encoder_atom_dig *dig;
|
|
- u8 backlight_level;
|
|
char bl_name[16];
|
|
|
|
/* Mac laptops with multiple GPUs use the gmux driver for backlight
|
|
@@ -222,12 +221,17 @@ void radeon_atom_backlight_init(struct radeon_encoder *radeon_encoder,
|
|
|
|
pdata->encoder = radeon_encoder;
|
|
|
|
- backlight_level = radeon_atom_get_backlight_level_from_reg(rdev);
|
|
-
|
|
dig = radeon_encoder->enc_priv;
|
|
dig->bl_dev = bd;
|
|
|
|
bd->props.brightness = radeon_atom_backlight_get_brightness(bd);
|
|
+ /* Set a reasonable default here if the level is 0 otherwise
|
|
+ * fbdev will attempt to turn the backlight on after console
|
|
+ * unblanking and it will try and restore 0 which turns the backlight
|
|
+ * off again.
|
|
+ */
|
|
+ if (bd->props.brightness == 0)
|
|
+ bd->props.brightness = RADEON_MAX_BL_LEVEL;
|
|
bd->props.power = FB_BLANK_UNBLANK;
|
|
backlight_update_status(bd);
|
|
|
|
diff --git a/drivers/gpu/drm/radeon/radeon_display.c b/drivers/gpu/drm/radeon/radeon_display.c
|
|
index 06ccfe477650..a84de32a91f5 100644
|
|
--- a/drivers/gpu/drm/radeon/radeon_display.c
|
|
+++ b/drivers/gpu/drm/radeon/radeon_display.c
|
|
@@ -688,6 +688,10 @@ int radeon_ddc_get_modes(struct radeon_connector *radeon_connector)
|
|
struct radeon_device *rdev = dev->dev_private;
|
|
int ret = 0;
|
|
|
|
+ /* don't leak the edid if we already fetched it in detect() */
|
|
+ if (radeon_connector->edid)
|
|
+ goto got_edid;
|
|
+
|
|
/* on hw with routers, select right port */
|
|
if (radeon_connector->router.ddc_valid)
|
|
radeon_router_select_ddc_port(radeon_connector);
|
|
@@ -727,6 +731,7 @@ int radeon_ddc_get_modes(struct radeon_connector *radeon_connector)
|
|
radeon_connector->edid = radeon_bios_get_hardcoded_edid(rdev);
|
|
}
|
|
if (radeon_connector->edid) {
|
|
+got_edid:
|
|
drm_mode_connector_update_edid_property(&radeon_connector->base, radeon_connector->edid);
|
|
ret = drm_add_edid_modes(&radeon_connector->base, radeon_connector->edid);
|
|
drm_edid_to_eld(&radeon_connector->base, radeon_connector->edid);
|
|
diff --git a/drivers/hv/hv_kvp.c b/drivers/hv/hv_kvp.c
|
|
index ed50e9e83c61..0e8c1ea4dd53 100644
|
|
--- a/drivers/hv/hv_kvp.c
|
|
+++ b/drivers/hv/hv_kvp.c
|
|
@@ -111,6 +111,15 @@ kvp_work_func(struct work_struct *dummy)
|
|
kvp_respond_to_host(NULL, HV_E_FAIL);
|
|
}
|
|
|
|
+static void poll_channel(struct vmbus_channel *channel)
|
|
+{
|
|
+ unsigned long flags;
|
|
+
|
|
+ spin_lock_irqsave(&channel->inbound_lock, flags);
|
|
+ hv_kvp_onchannelcallback(channel);
|
|
+ spin_unlock_irqrestore(&channel->inbound_lock, flags);
|
|
+}
|
|
+
|
|
static int kvp_handle_handshake(struct hv_kvp_msg *msg)
|
|
{
|
|
int ret = 1;
|
|
@@ -139,7 +148,7 @@ static int kvp_handle_handshake(struct hv_kvp_msg *msg)
|
|
kvp_register(dm_reg_value);
|
|
kvp_transaction.active = false;
|
|
if (kvp_transaction.kvp_context)
|
|
- hv_kvp_onchannelcallback(kvp_transaction.kvp_context);
|
|
+ poll_channel(kvp_transaction.kvp_context);
|
|
}
|
|
return ret;
|
|
}
|
|
@@ -552,6 +561,7 @@ response_done:
|
|
|
|
vmbus_sendpacket(channel, recv_buffer, buf_len, req_id,
|
|
VM_PKT_DATA_INBAND, 0);
|
|
+ poll_channel(channel);
|
|
|
|
}
|
|
|
|
@@ -585,7 +595,7 @@ void hv_kvp_onchannelcallback(void *context)
|
|
return;
|
|
}
|
|
|
|
- vmbus_recvpacket(channel, recv_buffer, PAGE_SIZE * 2, &recvlen,
|
|
+ vmbus_recvpacket(channel, recv_buffer, PAGE_SIZE * 4, &recvlen,
|
|
&requestid);
|
|
|
|
if (recvlen > 0) {
|
|
diff --git a/drivers/hv/hv_util.c b/drivers/hv/hv_util.c
|
|
index 2f561c5dfe24..64c778f7756f 100644
|
|
--- a/drivers/hv/hv_util.c
|
|
+++ b/drivers/hv/hv_util.c
|
|
@@ -279,7 +279,7 @@ static int util_probe(struct hv_device *dev,
|
|
(struct hv_util_service *)dev_id->driver_data;
|
|
int ret;
|
|
|
|
- srv->recv_buffer = kmalloc(PAGE_SIZE * 2, GFP_KERNEL);
|
|
+ srv->recv_buffer = kmalloc(PAGE_SIZE * 4, GFP_KERNEL);
|
|
if (!srv->recv_buffer)
|
|
return -ENOMEM;
|
|
if (srv->util_init) {
|
|
diff --git a/drivers/hwmon/adt7470.c b/drivers/hwmon/adt7470.c
|
|
index 58637355c1f6..79610bdf1d35 100644
|
|
--- a/drivers/hwmon/adt7470.c
|
|
+++ b/drivers/hwmon/adt7470.c
|
|
@@ -515,7 +515,7 @@ static ssize_t set_temp_min(struct device *dev,
|
|
return -EINVAL;
|
|
|
|
temp = DIV_ROUND_CLOSEST(temp, 1000);
|
|
- temp = clamp_val(temp, 0, 255);
|
|
+ temp = clamp_val(temp, -128, 127);
|
|
|
|
mutex_lock(&data->lock);
|
|
data->temp_min[attr->index] = temp;
|
|
@@ -549,7 +549,7 @@ static ssize_t set_temp_max(struct device *dev,
|
|
return -EINVAL;
|
|
|
|
temp = DIV_ROUND_CLOSEST(temp, 1000);
|
|
- temp = clamp_val(temp, 0, 255);
|
|
+ temp = clamp_val(temp, -128, 127);
|
|
|
|
mutex_lock(&data->lock);
|
|
data->temp_max[attr->index] = temp;
|
|
@@ -826,7 +826,7 @@ static ssize_t set_pwm_tmin(struct device *dev,
|
|
return -EINVAL;
|
|
|
|
temp = DIV_ROUND_CLOSEST(temp, 1000);
|
|
- temp = clamp_val(temp, 0, 255);
|
|
+ temp = clamp_val(temp, -128, 127);
|
|
|
|
mutex_lock(&data->lock);
|
|
data->pwm_tmin[attr->index] = temp;
|
|
diff --git a/drivers/hwmon/da9052-hwmon.c b/drivers/hwmon/da9052-hwmon.c
|
|
index 960fac3fb166..48044b044b7a 100644
|
|
--- a/drivers/hwmon/da9052-hwmon.c
|
|
+++ b/drivers/hwmon/da9052-hwmon.c
|
|
@@ -194,7 +194,7 @@ static ssize_t da9052_hwmon_show_name(struct device *dev,
|
|
struct device_attribute *devattr,
|
|
char *buf)
|
|
{
|
|
- return sprintf(buf, "da9052-hwmon\n");
|
|
+ return sprintf(buf, "da9052\n");
|
|
}
|
|
|
|
static ssize_t show_label(struct device *dev,
|
|
diff --git a/drivers/hwmon/da9055-hwmon.c b/drivers/hwmon/da9055-hwmon.c
|
|
index 029ecabc4380..1b275a2881d6 100644
|
|
--- a/drivers/hwmon/da9055-hwmon.c
|
|
+++ b/drivers/hwmon/da9055-hwmon.c
|
|
@@ -204,7 +204,7 @@ static ssize_t da9055_hwmon_show_name(struct device *dev,
|
|
struct device_attribute *devattr,
|
|
char *buf)
|
|
{
|
|
- return sprintf(buf, "da9055-hwmon\n");
|
|
+ return sprintf(buf, "da9055\n");
|
|
}
|
|
|
|
static ssize_t show_label(struct device *dev,
|
|
diff --git a/drivers/irqchip/irq-gic.c b/drivers/irqchip/irq-gic.c
|
|
index 4e11218d644e..c8ee1cb023b8 100644
|
|
--- a/drivers/irqchip/irq-gic.c
|
|
+++ b/drivers/irqchip/irq-gic.c
|
|
@@ -42,6 +42,7 @@
|
|
#include <linux/irqchip/chained_irq.h>
|
|
#include <linux/irqchip/arm-gic.h>
|
|
|
|
+#include <asm/cputype.h>
|
|
#include <asm/irq.h>
|
|
#include <asm/exception.h>
|
|
#include <asm/smp_plat.h>
|
|
@@ -754,7 +755,9 @@ void __init gic_init_bases(unsigned int gic_nr, int irq_start,
|
|
}
|
|
|
|
for_each_possible_cpu(cpu) {
|
|
- unsigned long offset = percpu_offset * cpu_logical_map(cpu);
|
|
+ u32 mpidr = cpu_logical_map(cpu);
|
|
+ u32 core_id = MPIDR_AFFINITY_LEVEL(mpidr, 0);
|
|
+ unsigned long offset = percpu_offset * core_id;
|
|
*per_cpu_ptr(gic->dist_base.percpu_base, cpu) = dist_base + offset;
|
|
*per_cpu_ptr(gic->cpu_base.percpu_base, cpu) = cpu_base + offset;
|
|
}
|
|
@@ -858,6 +861,7 @@ int __init gic_of_init(struct device_node *node, struct device_node *parent)
|
|
}
|
|
IRQCHIP_DECLARE(cortex_a15_gic, "arm,cortex-a15-gic", gic_of_init);
|
|
IRQCHIP_DECLARE(cortex_a9_gic, "arm,cortex-a9-gic", gic_of_init);
|
|
+IRQCHIP_DECLARE(cortex_a7_gic, "arm,cortex-a7-gic", gic_of_init);
|
|
IRQCHIP_DECLARE(msm_8660_qgic, "qcom,msm-8660-qgic", gic_of_init);
|
|
IRQCHIP_DECLARE(msm_qgic2, "qcom,msm-qgic2", gic_of_init);
|
|
|
|
diff --git a/drivers/md/dm-cache-metadata.c b/drivers/md/dm-cache-metadata.c
|
|
index a33e07f4222e..de737ba1d351 100644
|
|
--- a/drivers/md/dm-cache-metadata.c
|
|
+++ b/drivers/md/dm-cache-metadata.c
|
|
@@ -384,6 +384,15 @@ static int __open_metadata(struct dm_cache_metadata *cmd)
|
|
|
|
disk_super = dm_block_data(sblock);
|
|
|
|
+ /* Verify the data block size hasn't changed */
|
|
+ if (le32_to_cpu(disk_super->data_block_size) != cmd->data_block_size) {
|
|
+ DMERR("changing the data block size (from %u to %llu) is not supported",
|
|
+ le32_to_cpu(disk_super->data_block_size),
|
|
+ (unsigned long long)cmd->data_block_size);
|
|
+ r = -EINVAL;
|
|
+ goto bad;
|
|
+ }
|
|
+
|
|
r = __check_incompat_features(disk_super, cmd);
|
|
if (r < 0)
|
|
goto bad;
|
|
diff --git a/drivers/md/dm-thin-metadata.c b/drivers/md/dm-thin-metadata.c
|
|
index 5f49d704f275..3b1503dc1f13 100644
|
|
--- a/drivers/md/dm-thin-metadata.c
|
|
+++ b/drivers/md/dm-thin-metadata.c
|
|
@@ -591,6 +591,15 @@ static int __open_metadata(struct dm_pool_metadata *pmd)
|
|
|
|
disk_super = dm_block_data(sblock);
|
|
|
|
+ /* Verify the data block size hasn't changed */
|
|
+ if (le32_to_cpu(disk_super->data_block_size) != pmd->data_block_size) {
|
|
+ DMERR("changing the data block size (from %u to %llu) is not supported",
|
|
+ le32_to_cpu(disk_super->data_block_size),
|
|
+ (unsigned long long)pmd->data_block_size);
|
|
+ r = -EINVAL;
|
|
+ goto bad_unlock_sblock;
|
|
+ }
|
|
+
|
|
r = __check_incompat_features(disk_super, pmd);
|
|
if (r < 0)
|
|
goto bad_unlock_sblock;
|
|
diff --git a/drivers/media/usb/gspca/pac7302.c b/drivers/media/usb/gspca/pac7302.c
|
|
index 6008c8d546a3..20d9c15a305d 100644
|
|
--- a/drivers/media/usb/gspca/pac7302.c
|
|
+++ b/drivers/media/usb/gspca/pac7302.c
|
|
@@ -945,6 +945,7 @@ static const struct usb_device_id device_table[] = {
|
|
{USB_DEVICE(0x093a, 0x2620)},
|
|
{USB_DEVICE(0x093a, 0x2621)},
|
|
{USB_DEVICE(0x093a, 0x2622), .driver_info = FL_VFLIP},
|
|
+ {USB_DEVICE(0x093a, 0x2623), .driver_info = FL_VFLIP},
|
|
{USB_DEVICE(0x093a, 0x2624), .driver_info = FL_VFLIP},
|
|
{USB_DEVICE(0x093a, 0x2625)},
|
|
{USB_DEVICE(0x093a, 0x2626)},
|
|
diff --git a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
|
|
index 70be100feeb4..b04f7f128f49 100644
|
|
--- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
|
|
+++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
|
|
@@ -745,7 +745,8 @@ static void bnx2x_tpa_stop(struct bnx2x *bp, struct bnx2x_fastpath *fp,
|
|
|
|
return;
|
|
}
|
|
- bnx2x_frag_free(fp, new_data);
|
|
+ if (new_data)
|
|
+ bnx2x_frag_free(fp, new_data);
|
|
drop:
|
|
/* drop the packet and keep the buffer in the bin */
|
|
DP(NETIF_MSG_RX_STATUS,
|
|
diff --git a/drivers/net/ethernet/emulex/benet/be_main.c b/drivers/net/ethernet/emulex/benet/be_main.c
|
|
index 7371626c56a1..d81a7dbfeef6 100644
|
|
--- a/drivers/net/ethernet/emulex/benet/be_main.c
|
|
+++ b/drivers/net/ethernet/emulex/benet/be_main.c
|
|
@@ -2663,7 +2663,7 @@ static int be_open(struct net_device *netdev)
|
|
|
|
for_all_evt_queues(adapter, eqo, i) {
|
|
napi_enable(&eqo->napi);
|
|
- be_eq_notify(adapter, eqo->q.id, true, false, 0);
|
|
+ be_eq_notify(adapter, eqo->q.id, true, true, 0);
|
|
}
|
|
adapter->flags |= BE_FLAGS_NAPI_ENABLED;
|
|
|
|
diff --git a/drivers/net/ethernet/intel/igb/igb_main.c b/drivers/net/ethernet/intel/igb/igb_main.c
|
|
index 64cbe0dfe043..4d3c8122e2aa 100644
|
|
--- a/drivers/net/ethernet/intel/igb/igb_main.c
|
|
+++ b/drivers/net/ethernet/intel/igb/igb_main.c
|
|
@@ -7229,6 +7229,8 @@ static int igb_sriov_reinit(struct pci_dev *dev)
|
|
|
|
if (netif_running(netdev))
|
|
igb_close(netdev);
|
|
+ else
|
|
+ igb_reset(adapter);
|
|
|
|
igb_clear_interrupt_scheme(adapter);
|
|
|
|
diff --git a/drivers/net/ethernet/marvell/mvneta.c b/drivers/net/ethernet/marvell/mvneta.c
|
|
index a602aeeb3acb..658613021919 100644
|
|
--- a/drivers/net/ethernet/marvell/mvneta.c
|
|
+++ b/drivers/net/ethernet/marvell/mvneta.c
|
|
@@ -1145,7 +1145,7 @@ static u32 mvneta_txq_desc_csum(int l3_offs, int l3_proto,
|
|
command = l3_offs << MVNETA_TX_L3_OFF_SHIFT;
|
|
command |= ip_hdr_len << MVNETA_TX_IP_HLEN_SHIFT;
|
|
|
|
- if (l3_proto == swab16(ETH_P_IP))
|
|
+ if (l3_proto == htons(ETH_P_IP))
|
|
command |= MVNETA_TXD_IP_CSUM;
|
|
else
|
|
command |= MVNETA_TX_L3_IP6;
|
|
@@ -2306,7 +2306,7 @@ static void mvneta_adjust_link(struct net_device *ndev)
|
|
|
|
if (phydev->speed == SPEED_1000)
|
|
val |= MVNETA_GMAC_CONFIG_GMII_SPEED;
|
|
- else
|
|
+ else if (phydev->speed == SPEED_100)
|
|
val |= MVNETA_GMAC_CONFIG_MII_SPEED;
|
|
|
|
mvreg_write(pp, MVNETA_GMAC_AUTONEG_CONFIG, val);
|
|
diff --git a/drivers/net/ethernet/sun/sunvnet.c b/drivers/net/ethernet/sun/sunvnet.c
|
|
index 3df56840a3b9..398faff8be7a 100644
|
|
--- a/drivers/net/ethernet/sun/sunvnet.c
|
|
+++ b/drivers/net/ethernet/sun/sunvnet.c
|
|
@@ -1083,6 +1083,24 @@ static struct vnet *vnet_find_or_create(const u64 *local_mac)
|
|
return vp;
|
|
}
|
|
|
|
+static void vnet_cleanup(void)
|
|
+{
|
|
+ struct vnet *vp;
|
|
+ struct net_device *dev;
|
|
+
|
|
+ mutex_lock(&vnet_list_mutex);
|
|
+ while (!list_empty(&vnet_list)) {
|
|
+ vp = list_first_entry(&vnet_list, struct vnet, list);
|
|
+ list_del(&vp->list);
|
|
+ dev = vp->dev;
|
|
+ /* vio_unregister_driver() should have cleaned up port_list */
|
|
+ BUG_ON(!list_empty(&vp->port_list));
|
|
+ unregister_netdev(dev);
|
|
+ free_netdev(dev);
|
|
+ }
|
|
+ mutex_unlock(&vnet_list_mutex);
|
|
+}
|
|
+
|
|
static const char *local_mac_prop = "local-mac-address";
|
|
|
|
static struct vnet *vnet_find_parent(struct mdesc_handle *hp,
|
|
@@ -1240,7 +1258,6 @@ static int vnet_port_remove(struct vio_dev *vdev)
|
|
|
|
kfree(port);
|
|
|
|
- unregister_netdev(vp->dev);
|
|
}
|
|
return 0;
|
|
}
|
|
@@ -1268,6 +1285,7 @@ static int __init vnet_init(void)
|
|
static void __exit vnet_exit(void)
|
|
{
|
|
vio_unregister_driver(&vnet_port_driver);
|
|
+ vnet_cleanup();
|
|
}
|
|
|
|
module_init(vnet_init);
|
|
diff --git a/drivers/net/ppp/pppoe.c b/drivers/net/ppp/pppoe.c
|
|
index 6839fb07a4c9..becfa3ef7fdc 100644
|
|
--- a/drivers/net/ppp/pppoe.c
|
|
+++ b/drivers/net/ppp/pppoe.c
|
|
@@ -675,7 +675,7 @@ static int pppoe_connect(struct socket *sock, struct sockaddr *uservaddr,
|
|
po->chan.hdrlen = (sizeof(struct pppoe_hdr) +
|
|
dev->hard_header_len);
|
|
|
|
- po->chan.mtu = dev->mtu - sizeof(struct pppoe_hdr);
|
|
+ po->chan.mtu = dev->mtu - sizeof(struct pppoe_hdr) - 2;
|
|
po->chan.private = sk;
|
|
po->chan.ops = &pppoe_chan_ops;
|
|
|
|
diff --git a/drivers/net/usb/qmi_wwan.c b/drivers/net/usb/qmi_wwan.c
|
|
index 6fb0082b3308..6c584f8a2268 100644
|
|
--- a/drivers/net/usb/qmi_wwan.c
|
|
+++ b/drivers/net/usb/qmi_wwan.c
|
|
@@ -647,6 +647,7 @@ static const struct usb_device_id products[] = {
|
|
{QMI_FIXED_INTF(0x05c6, 0x9084, 4)},
|
|
{QMI_FIXED_INTF(0x05c6, 0x920d, 0)},
|
|
{QMI_FIXED_INTF(0x05c6, 0x920d, 5)},
|
|
+ {QMI_FIXED_INTF(0x0846, 0x68a2, 8)},
|
|
{QMI_FIXED_INTF(0x12d1, 0x140c, 1)}, /* Huawei E173 */
|
|
{QMI_FIXED_INTF(0x12d1, 0x14ac, 1)}, /* Huawei E1820 */
|
|
{QMI_FIXED_INTF(0x16d8, 0x6003, 0)}, /* CMOTech 6003 */
|
|
@@ -721,6 +722,7 @@ static const struct usb_device_id products[] = {
|
|
{QMI_FIXED_INTF(0x19d2, 0x1424, 2)},
|
|
{QMI_FIXED_INTF(0x19d2, 0x1425, 2)},
|
|
{QMI_FIXED_INTF(0x19d2, 0x1426, 2)}, /* ZTE MF91 */
|
|
+ {QMI_FIXED_INTF(0x19d2, 0x1428, 2)}, /* Telewell TW-LTE 4G v2 */
|
|
{QMI_FIXED_INTF(0x19d2, 0x2002, 4)}, /* ZTE (Vodafone) K3765-Z */
|
|
{QMI_FIXED_INTF(0x0f3d, 0x68a2, 8)}, /* Sierra Wireless MC7700 */
|
|
{QMI_FIXED_INTF(0x114f, 0x68a2, 8)}, /* Sierra Wireless MC7750 */
|
|
@@ -733,6 +735,7 @@ static const struct usb_device_id products[] = {
|
|
{QMI_FIXED_INTF(0x1199, 0x901f, 8)}, /* Sierra Wireless EM7355 */
|
|
{QMI_FIXED_INTF(0x1199, 0x9041, 8)}, /* Sierra Wireless MC7305/MC7355 */
|
|
{QMI_FIXED_INTF(0x1199, 0x9051, 8)}, /* Netgear AirCard 340U */
|
|
+ {QMI_FIXED_INTF(0x1199, 0x9057, 8)},
|
|
{QMI_FIXED_INTF(0x1bbb, 0x011e, 4)}, /* Telekom Speedstick LTE II (Alcatel One Touch L100V LTE) */
|
|
{QMI_FIXED_INTF(0x1bbb, 0x0203, 2)}, /* Alcatel L800MA */
|
|
{QMI_FIXED_INTF(0x2357, 0x0201, 4)}, /* TP-LINK HSUPA Modem MA180 */
|
|
diff --git a/drivers/net/wireless/iwlwifi/dvm/rxon.c b/drivers/net/wireless/iwlwifi/dvm/rxon.c
|
|
index cd1ad0019185..ca17e4c9eca2 100644
|
|
--- a/drivers/net/wireless/iwlwifi/dvm/rxon.c
|
|
+++ b/drivers/net/wireless/iwlwifi/dvm/rxon.c
|
|
@@ -1072,13 +1072,6 @@ int iwlagn_commit_rxon(struct iwl_priv *priv, struct iwl_rxon_context *ctx)
|
|
/* recalculate basic rates */
|
|
iwl_calc_basic_rates(priv, ctx);
|
|
|
|
- /*
|
|
- * force CTS-to-self frames protection if RTS-CTS is not preferred
|
|
- * one aggregation protection method
|
|
- */
|
|
- if (!priv->hw_params.use_rts_for_aggregation)
|
|
- ctx->staging.flags |= RXON_FLG_SELF_CTS_EN;
|
|
-
|
|
if ((ctx->vif && ctx->vif->bss_conf.use_short_slot) ||
|
|
!(ctx->staging.flags & RXON_FLG_BAND_24G_MSK))
|
|
ctx->staging.flags |= RXON_FLG_SHORT_SLOT_MSK;
|
|
@@ -1484,11 +1477,6 @@ void iwlagn_bss_info_changed(struct ieee80211_hw *hw,
|
|
else
|
|
ctx->staging.flags &= ~RXON_FLG_TGG_PROTECT_MSK;
|
|
|
|
- if (bss_conf->use_cts_prot)
|
|
- ctx->staging.flags |= RXON_FLG_SELF_CTS_EN;
|
|
- else
|
|
- ctx->staging.flags &= ~RXON_FLG_SELF_CTS_EN;
|
|
-
|
|
memcpy(ctx->staging.bssid_addr, bss_conf->bssid, ETH_ALEN);
|
|
|
|
if (vif->type == NL80211_IFTYPE_AP ||
|
|
diff --git a/drivers/net/wireless/mwifiex/main.c b/drivers/net/wireless/mwifiex/main.c
|
|
index fc3fe8ddcf62..83c61964d082 100644
|
|
--- a/drivers/net/wireless/mwifiex/main.c
|
|
+++ b/drivers/net/wireless/mwifiex/main.c
|
|
@@ -501,6 +501,7 @@ mwifiex_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
|
}
|
|
|
|
tx_info = MWIFIEX_SKB_TXCB(skb);
|
|
+ memset(tx_info, 0, sizeof(*tx_info));
|
|
tx_info->bss_num = priv->bss_num;
|
|
tx_info->bss_type = priv->bss_type;
|
|
|
|
diff --git a/drivers/usb/core/hub.c b/drivers/usb/core/hub.c
|
|
index 46efdca96952..63c217053668 100644
|
|
--- a/drivers/usb/core/hub.c
|
|
+++ b/drivers/usb/core/hub.c
|
|
@@ -887,6 +887,25 @@ static int hub_usb3_port_disable(struct usb_hub *hub, int port1)
|
|
if (!hub_is_superspeed(hub->hdev))
|
|
return -EINVAL;
|
|
|
|
+ ret = hub_port_status(hub, port1, &portstatus, &portchange);
|
|
+ if (ret < 0)
|
|
+ return ret;
|
|
+
|
|
+ /*
|
|
+ * USB controller Advanced Micro Devices, Inc. [AMD] FCH USB XHCI
|
|
+ * Controller [1022:7814] will have spurious result making the following
|
|
+ * usb 3.0 device hotplugging route to the 2.0 root hub and recognized
|
|
+ * as high-speed device if we set the usb 3.0 port link state to
|
|
+ * Disabled. Since it's already in USB_SS_PORT_LS_RX_DETECT state, we
|
|
+ * check the state here to avoid the bug.
|
|
+ */
|
|
+ if ((portstatus & USB_PORT_STAT_LINK_STATE) ==
|
|
+ USB_SS_PORT_LS_RX_DETECT) {
|
|
+ dev_dbg(&hub->ports[port1 - 1]->dev,
|
|
+ "Not disabling port; link state is RxDetect\n");
|
|
+ return ret;
|
|
+ }
|
|
+
|
|
ret = hub_set_port_link_state(hub, port1, USB_SS_PORT_LS_SS_DISABLED);
|
|
if (ret)
|
|
return ret;
|
|
diff --git a/fs/fuse/inode.c b/fs/fuse/inode.c
|
|
index b5718516825b..39a986e1da9e 100644
|
|
--- a/fs/fuse/inode.c
|
|
+++ b/fs/fuse/inode.c
|
|
@@ -461,6 +461,17 @@ static const match_table_t tokens = {
|
|
{OPT_ERR, NULL}
|
|
};
|
|
|
|
+static int fuse_match_uint(substring_t *s, unsigned int *res)
|
|
+{
|
|
+ int err = -ENOMEM;
|
|
+ char *buf = match_strdup(s);
|
|
+ if (buf) {
|
|
+ err = kstrtouint(buf, 10, res);
|
|
+ kfree(buf);
|
|
+ }
|
|
+ return err;
|
|
+}
|
|
+
|
|
static int parse_fuse_opt(char *opt, struct fuse_mount_data *d, int is_bdev)
|
|
{
|
|
char *p;
|
|
@@ -471,6 +482,7 @@ static int parse_fuse_opt(char *opt, struct fuse_mount_data *d, int is_bdev)
|
|
while ((p = strsep(&opt, ",")) != NULL) {
|
|
int token;
|
|
int value;
|
|
+ unsigned uv;
|
|
substring_t args[MAX_OPT_ARGS];
|
|
if (!*p)
|
|
continue;
|
|
@@ -494,18 +506,18 @@ static int parse_fuse_opt(char *opt, struct fuse_mount_data *d, int is_bdev)
|
|
break;
|
|
|
|
case OPT_USER_ID:
|
|
- if (match_int(&args[0], &value))
|
|
+ if (fuse_match_uint(&args[0], &uv))
|
|
return 0;
|
|
- d->user_id = make_kuid(current_user_ns(), value);
|
|
+ d->user_id = make_kuid(current_user_ns(), uv);
|
|
if (!uid_valid(d->user_id))
|
|
return 0;
|
|
d->user_id_present = 1;
|
|
break;
|
|
|
|
case OPT_GROUP_ID:
|
|
- if (match_int(&args[0], &value))
|
|
+ if (fuse_match_uint(&args[0], &uv))
|
|
return 0;
|
|
- d->group_id = make_kgid(current_user_ns(), value);
|
|
+ d->group_id = make_kgid(current_user_ns(), uv);
|
|
if (!gid_valid(d->group_id))
|
|
return 0;
|
|
d->group_id_present = 1;
|
|
diff --git a/include/net/sock.h b/include/net/sock.h
|
|
index 72f710d2f75a..26b15c0780be 100644
|
|
--- a/include/net/sock.h
|
|
+++ b/include/net/sock.h
|
|
@@ -1727,8 +1727,8 @@ sk_dst_get(struct sock *sk)
|
|
|
|
rcu_read_lock();
|
|
dst = rcu_dereference(sk->sk_dst_cache);
|
|
- if (dst)
|
|
- dst_hold(dst);
|
|
+ if (dst && !atomic_inc_not_zero(&dst->__refcnt))
|
|
+ dst = NULL;
|
|
rcu_read_unlock();
|
|
return dst;
|
|
}
|
|
@@ -1767,9 +1767,11 @@ __sk_dst_set(struct sock *sk, struct dst_entry *dst)
|
|
static inline void
|
|
sk_dst_set(struct sock *sk, struct dst_entry *dst)
|
|
{
|
|
- spin_lock(&sk->sk_dst_lock);
|
|
- __sk_dst_set(sk, dst);
|
|
- spin_unlock(&sk->sk_dst_lock);
|
|
+ struct dst_entry *old_dst;
|
|
+
|
|
+ sk_tx_queue_clear(sk);
|
|
+ old_dst = xchg((__force struct dst_entry **)&sk->sk_dst_cache, dst);
|
|
+ dst_release(old_dst);
|
|
}
|
|
|
|
static inline void
|
|
@@ -1781,9 +1783,7 @@ __sk_dst_reset(struct sock *sk)
|
|
static inline void
|
|
sk_dst_reset(struct sock *sk)
|
|
{
|
|
- spin_lock(&sk->sk_dst_lock);
|
|
- __sk_dst_reset(sk);
|
|
- spin_unlock(&sk->sk_dst_lock);
|
|
+ sk_dst_set(sk, NULL);
|
|
}
|
|
|
|
extern struct dst_entry *__sk_dst_check(struct sock *sk, u32 cookie);
|
|
diff --git a/kernel/Kconfig.locks b/kernel/Kconfig.locks
|
|
index 44511d100eaa..e4d30533c562 100644
|
|
--- a/kernel/Kconfig.locks
|
|
+++ b/kernel/Kconfig.locks
|
|
@@ -220,6 +220,9 @@ config INLINE_WRITE_UNLOCK_IRQRESTORE
|
|
|
|
endif
|
|
|
|
+config ARCH_SUPPORTS_ATOMIC_RMW
|
|
+ bool
|
|
+
|
|
config MUTEX_SPIN_ON_OWNER
|
|
def_bool y
|
|
- depends on SMP && !DEBUG_MUTEXES
|
|
+ depends on SMP && !DEBUG_MUTEXES && ARCH_SUPPORTS_ATOMIC_RMW
|
|
diff --git a/kernel/power/process.c b/kernel/power/process.c
|
|
index 98088e0e71e8..1b212bee1510 100644
|
|
--- a/kernel/power/process.c
|
|
+++ b/kernel/power/process.c
|
|
@@ -174,6 +174,7 @@ void thaw_processes(void)
|
|
|
|
printk("Restarting tasks ... ");
|
|
|
|
+ __usermodehelper_set_disable_depth(UMH_FREEZING);
|
|
thaw_workqueues();
|
|
|
|
read_lock(&tasklist_lock);
|
|
diff --git a/kernel/sched/debug.c b/kernel/sched/debug.c
|
|
index e745a1548367..701b6c8a4b12 100644
|
|
--- a/kernel/sched/debug.c
|
|
+++ b/kernel/sched/debug.c
|
|
@@ -551,7 +551,7 @@ void proc_sched_show_task(struct task_struct *p, struct seq_file *m)
|
|
|
|
avg_atom = p->se.sum_exec_runtime;
|
|
if (nr_switches)
|
|
- do_div(avg_atom, nr_switches);
|
|
+ avg_atom = div64_ul(avg_atom, nr_switches);
|
|
else
|
|
avg_atom = -1LL;
|
|
|
|
diff --git a/kernel/time/alarmtimer.c b/kernel/time/alarmtimer.c
|
|
index a8f5084dcde7..294bf4ef1f47 100644
|
|
--- a/kernel/time/alarmtimer.c
|
|
+++ b/kernel/time/alarmtimer.c
|
|
@@ -540,9 +540,14 @@ static int alarm_timer_set(struct k_itimer *timr, int flags,
|
|
struct itimerspec *new_setting,
|
|
struct itimerspec *old_setting)
|
|
{
|
|
+ ktime_t exp;
|
|
+
|
|
if (!rtcdev)
|
|
return -ENOTSUPP;
|
|
|
|
+ if (flags & ~TIMER_ABSTIME)
|
|
+ return -EINVAL;
|
|
+
|
|
if (old_setting)
|
|
alarm_timer_get(timr, old_setting);
|
|
|
|
@@ -552,8 +557,16 @@ static int alarm_timer_set(struct k_itimer *timr, int flags,
|
|
|
|
/* start the timer */
|
|
timr->it.alarm.interval = timespec_to_ktime(new_setting->it_interval);
|
|
- alarm_start(&timr->it.alarm.alarmtimer,
|
|
- timespec_to_ktime(new_setting->it_value));
|
|
+ exp = timespec_to_ktime(new_setting->it_value);
|
|
+ /* Convert (if necessary) to absolute time */
|
|
+ if (flags != TIMER_ABSTIME) {
|
|
+ ktime_t now;
|
|
+
|
|
+ now = alarm_bases[timr->it.alarm.alarmtimer.type].gettime();
|
|
+ exp = ktime_add(now, exp);
|
|
+ }
|
|
+
|
|
+ alarm_start(&timr->it.alarm.alarmtimer, exp);
|
|
return 0;
|
|
}
|
|
|
|
@@ -685,6 +698,9 @@ static int alarm_timer_nsleep(const clockid_t which_clock, int flags,
|
|
if (!alarmtimer_get_rtcdev())
|
|
return -ENOTSUPP;
|
|
|
|
+ if (flags & ~TIMER_ABSTIME)
|
|
+ return -EINVAL;
|
|
+
|
|
if (!capable(CAP_WAKE_ALARM))
|
|
return -EPERM;
|
|
|
|
diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c
|
|
index 797d3b91a30b..401d9bd1fe42 100644
|
|
--- a/kernel/trace/ftrace.c
|
|
+++ b/kernel/trace/ftrace.c
|
|
@@ -331,12 +331,12 @@ static void update_ftrace_function(void)
|
|
func = ftrace_ops_list_func;
|
|
}
|
|
|
|
+ update_function_graph_func();
|
|
+
|
|
/* If there's no change, then do nothing more here */
|
|
if (ftrace_trace_function == func)
|
|
return;
|
|
|
|
- update_function_graph_func();
|
|
-
|
|
/*
|
|
* If we are using the list function, it doesn't care
|
|
* about the function_trace_ops.
|
|
diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c
|
|
index 8e94c1102636..4063d5fe5e44 100644
|
|
--- a/kernel/trace/ring_buffer.c
|
|
+++ b/kernel/trace/ring_buffer.c
|
|
@@ -616,10 +616,6 @@ int ring_buffer_poll_wait(struct ring_buffer *buffer, int cpu,
|
|
struct ring_buffer_per_cpu *cpu_buffer;
|
|
struct rb_irq_work *work;
|
|
|
|
- if ((cpu == RING_BUFFER_ALL_CPUS && !ring_buffer_empty(buffer)) ||
|
|
- (cpu != RING_BUFFER_ALL_CPUS && !ring_buffer_empty_cpu(buffer, cpu)))
|
|
- return POLLIN | POLLRDNORM;
|
|
-
|
|
if (cpu == RING_BUFFER_ALL_CPUS)
|
|
work = &buffer->irq_work;
|
|
else {
|
|
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c
|
|
index 8fe92ce43f39..98a830d079b9 100644
|
|
--- a/kernel/trace/trace.c
|
|
+++ b/kernel/trace/trace.c
|
|
@@ -423,6 +423,9 @@ int __trace_puts(unsigned long ip, const char *str, int size)
|
|
struct print_entry *entry;
|
|
unsigned long irq_flags;
|
|
int alloc;
|
|
+ int pc;
|
|
+
|
|
+ pc = preempt_count();
|
|
|
|
if (unlikely(tracing_selftest_running || tracing_disabled))
|
|
return 0;
|
|
@@ -432,7 +435,7 @@ int __trace_puts(unsigned long ip, const char *str, int size)
|
|
local_save_flags(irq_flags);
|
|
buffer = global_trace.trace_buffer.buffer;
|
|
event = trace_buffer_lock_reserve(buffer, TRACE_PRINT, alloc,
|
|
- irq_flags, preempt_count());
|
|
+ irq_flags, pc);
|
|
if (!event)
|
|
return 0;
|
|
|
|
@@ -449,6 +452,7 @@ int __trace_puts(unsigned long ip, const char *str, int size)
|
|
entry->buf[size] = '\0';
|
|
|
|
__buffer_unlock_commit(buffer, event);
|
|
+ ftrace_trace_stack(buffer, irq_flags, 4, pc);
|
|
|
|
return size;
|
|
}
|
|
@@ -466,6 +470,9 @@ int __trace_bputs(unsigned long ip, const char *str)
|
|
struct bputs_entry *entry;
|
|
unsigned long irq_flags;
|
|
int size = sizeof(struct bputs_entry);
|
|
+ int pc;
|
|
+
|
|
+ pc = preempt_count();
|
|
|
|
if (unlikely(tracing_selftest_running || tracing_disabled))
|
|
return 0;
|
|
@@ -473,7 +480,7 @@ int __trace_bputs(unsigned long ip, const char *str)
|
|
local_save_flags(irq_flags);
|
|
buffer = global_trace.trace_buffer.buffer;
|
|
event = trace_buffer_lock_reserve(buffer, TRACE_BPUTS, size,
|
|
- irq_flags, preempt_count());
|
|
+ irq_flags, pc);
|
|
if (!event)
|
|
return 0;
|
|
|
|
@@ -482,6 +489,7 @@ int __trace_bputs(unsigned long ip, const char *str)
|
|
entry->str = str;
|
|
|
|
__buffer_unlock_commit(buffer, event);
|
|
+ ftrace_trace_stack(buffer, irq_flags, 4, pc);
|
|
|
|
return 1;
|
|
}
|
|
diff --git a/mm/shmem.c b/mm/shmem.c
|
|
index 509b393eceeb..16cc1d77f70a 100644
|
|
--- a/mm/shmem.c
|
|
+++ b/mm/shmem.c
|
|
@@ -80,11 +80,12 @@ static struct vfsmount *shm_mnt;
|
|
#define SHORT_SYMLINK_LEN 128
|
|
|
|
/*
|
|
- * shmem_fallocate and shmem_writepage communicate via inode->i_private
|
|
- * (with i_mutex making sure that it has only one user at a time):
|
|
- * we would prefer not to enlarge the shmem inode just for that.
|
|
+ * shmem_fallocate communicates with shmem_fault or shmem_writepage via
|
|
+ * inode->i_private (with i_mutex making sure that it has only one user at
|
|
+ * a time): we would prefer not to enlarge the shmem inode just for that.
|
|
*/
|
|
struct shmem_falloc {
|
|
+ wait_queue_head_t *waitq; /* faults into hole wait for punch to end */
|
|
pgoff_t start; /* start of range currently being fallocated */
|
|
pgoff_t next; /* the next page offset to be fallocated */
|
|
pgoff_t nr_falloced; /* how many new pages have been fallocated */
|
|
@@ -533,22 +534,19 @@ static void shmem_undo_range(struct inode *inode, loff_t lstart, loff_t lend,
|
|
return;
|
|
|
|
index = start;
|
|
- for ( ; ; ) {
|
|
+ while (index < end) {
|
|
cond_resched();
|
|
pvec.nr = shmem_find_get_pages_and_swap(mapping, index,
|
|
min(end - index, (pgoff_t)PAGEVEC_SIZE),
|
|
pvec.pages, indices);
|
|
if (!pvec.nr) {
|
|
- if (index == start || unfalloc)
|
|
+ /* If all gone or hole-punch or unfalloc, we're done */
|
|
+ if (index == start || end != -1)
|
|
break;
|
|
+ /* But if truncating, restart to make sure all gone */
|
|
index = start;
|
|
continue;
|
|
}
|
|
- if ((index == start || unfalloc) && indices[0] >= end) {
|
|
- shmem_deswap_pagevec(&pvec);
|
|
- pagevec_release(&pvec);
|
|
- break;
|
|
- }
|
|
mem_cgroup_uncharge_start();
|
|
for (i = 0; i < pagevec_count(&pvec); i++) {
|
|
struct page *page = pvec.pages[i];
|
|
@@ -560,8 +558,12 @@ static void shmem_undo_range(struct inode *inode, loff_t lstart, loff_t lend,
|
|
if (radix_tree_exceptional_entry(page)) {
|
|
if (unfalloc)
|
|
continue;
|
|
- nr_swaps_freed += !shmem_free_swap(mapping,
|
|
- index, page);
|
|
+ if (shmem_free_swap(mapping, index, page)) {
|
|
+ /* Swap was replaced by page: retry */
|
|
+ index--;
|
|
+ break;
|
|
+ }
|
|
+ nr_swaps_freed++;
|
|
continue;
|
|
}
|
|
|
|
@@ -570,6 +572,11 @@ static void shmem_undo_range(struct inode *inode, loff_t lstart, loff_t lend,
|
|
if (page->mapping == mapping) {
|
|
VM_BUG_ON(PageWriteback(page));
|
|
truncate_inode_page(mapping, page);
|
|
+ } else {
|
|
+ /* Page was replaced by swap: retry */
|
|
+ unlock_page(page);
|
|
+ index--;
|
|
+ break;
|
|
}
|
|
}
|
|
unlock_page(page);
|
|
@@ -826,6 +833,7 @@ static int shmem_writepage(struct page *page, struct writeback_control *wbc)
|
|
spin_lock(&inode->i_lock);
|
|
shmem_falloc = inode->i_private;
|
|
if (shmem_falloc &&
|
|
+ !shmem_falloc->waitq &&
|
|
index >= shmem_falloc->start &&
|
|
index < shmem_falloc->next)
|
|
shmem_falloc->nr_unswapped++;
|
|
@@ -1300,6 +1308,64 @@ static int shmem_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
|
|
int error;
|
|
int ret = VM_FAULT_LOCKED;
|
|
|
|
+ /*
|
|
+ * Trinity finds that probing a hole which tmpfs is punching can
|
|
+ * prevent the hole-punch from ever completing: which in turn
|
|
+ * locks writers out with its hold on i_mutex. So refrain from
|
|
+ * faulting pages into the hole while it's being punched. Although
|
|
+ * shmem_undo_range() does remove the additions, it may be unable to
|
|
+ * keep up, as each new page needs its own unmap_mapping_range() call,
|
|
+ * and the i_mmap tree grows ever slower to scan if new vmas are added.
|
|
+ *
|
|
+ * It does not matter if we sometimes reach this check just before the
|
|
+ * hole-punch begins, so that one fault then races with the punch:
|
|
+ * we just need to make racing faults a rare case.
|
|
+ *
|
|
+ * The implementation below would be much simpler if we just used a
|
|
+ * standard mutex or completion: but we cannot take i_mutex in fault,
|
|
+ * and bloating every shmem inode for this unlikely case would be sad.
|
|
+ */
|
|
+ if (unlikely(inode->i_private)) {
|
|
+ struct shmem_falloc *shmem_falloc;
|
|
+
|
|
+ spin_lock(&inode->i_lock);
|
|
+ shmem_falloc = inode->i_private;
|
|
+ if (shmem_falloc &&
|
|
+ shmem_falloc->waitq &&
|
|
+ vmf->pgoff >= shmem_falloc->start &&
|
|
+ vmf->pgoff < shmem_falloc->next) {
|
|
+ wait_queue_head_t *shmem_falloc_waitq;
|
|
+ DEFINE_WAIT(shmem_fault_wait);
|
|
+
|
|
+ ret = VM_FAULT_NOPAGE;
|
|
+ if ((vmf->flags & FAULT_FLAG_ALLOW_RETRY) &&
|
|
+ !(vmf->flags & FAULT_FLAG_RETRY_NOWAIT)) {
|
|
+ /* It's polite to up mmap_sem if we can */
|
|
+ up_read(&vma->vm_mm->mmap_sem);
|
|
+ ret = VM_FAULT_RETRY;
|
|
+ }
|
|
+
|
|
+ shmem_falloc_waitq = shmem_falloc->waitq;
|
|
+ prepare_to_wait(shmem_falloc_waitq, &shmem_fault_wait,
|
|
+ TASK_UNINTERRUPTIBLE);
|
|
+ spin_unlock(&inode->i_lock);
|
|
+ schedule();
|
|
+
|
|
+ /*
|
|
+ * shmem_falloc_waitq points into the shmem_fallocate()
|
|
+ * stack of the hole-punching task: shmem_falloc_waitq
|
|
+ * is usually invalid by the time we reach here, but
|
|
+ * finish_wait() does not dereference it in that case;
|
|
+ * though i_lock needed lest racing with wake_up_all().
|
|
+ */
|
|
+ spin_lock(&inode->i_lock);
|
|
+ finish_wait(shmem_falloc_waitq, &shmem_fault_wait);
|
|
+ spin_unlock(&inode->i_lock);
|
|
+ return ret;
|
|
+ }
|
|
+ spin_unlock(&inode->i_lock);
|
|
+ }
|
|
+
|
|
error = shmem_getpage(inode, vmf->pgoff, &vmf->page, SGP_CACHE, &ret);
|
|
if (error)
|
|
return ((error == -ENOMEM) ? VM_FAULT_OOM : VM_FAULT_SIGBUS);
|
|
@@ -1821,12 +1887,25 @@ static long shmem_fallocate(struct file *file, int mode, loff_t offset,
|
|
struct address_space *mapping = file->f_mapping;
|
|
loff_t unmap_start = round_up(offset, PAGE_SIZE);
|
|
loff_t unmap_end = round_down(offset + len, PAGE_SIZE) - 1;
|
|
+ DECLARE_WAIT_QUEUE_HEAD_ONSTACK(shmem_falloc_waitq);
|
|
+
|
|
+ shmem_falloc.waitq = &shmem_falloc_waitq;
|
|
+ shmem_falloc.start = unmap_start >> PAGE_SHIFT;
|
|
+ shmem_falloc.next = (unmap_end + 1) >> PAGE_SHIFT;
|
|
+ spin_lock(&inode->i_lock);
|
|
+ inode->i_private = &shmem_falloc;
|
|
+ spin_unlock(&inode->i_lock);
|
|
|
|
if ((u64)unmap_end > (u64)unmap_start)
|
|
unmap_mapping_range(mapping, unmap_start,
|
|
1 + unmap_end - unmap_start, 0);
|
|
shmem_truncate_range(inode, offset, offset + len - 1);
|
|
/* No need to unmap again: hole-punching leaves COWed pages */
|
|
+
|
|
+ spin_lock(&inode->i_lock);
|
|
+ inode->i_private = NULL;
|
|
+ wake_up_all(&shmem_falloc_waitq);
|
|
+ spin_unlock(&inode->i_lock);
|
|
error = 0;
|
|
goto out;
|
|
}
|
|
@@ -1844,6 +1923,7 @@ static long shmem_fallocate(struct file *file, int mode, loff_t offset,
|
|
goto out;
|
|
}
|
|
|
|
+ shmem_falloc.waitq = NULL;
|
|
shmem_falloc.start = start;
|
|
shmem_falloc.next = start;
|
|
shmem_falloc.nr_falloced = 0;
|
|
diff --git a/net/8021q/vlan_core.c b/net/8021q/vlan_core.c
|
|
index 4a78c4de9f20..42ef36a85e69 100644
|
|
--- a/net/8021q/vlan_core.c
|
|
+++ b/net/8021q/vlan_core.c
|
|
@@ -103,8 +103,11 @@ EXPORT_SYMBOL(vlan_dev_vlan_id);
|
|
|
|
static struct sk_buff *vlan_reorder_header(struct sk_buff *skb)
|
|
{
|
|
- if (skb_cow(skb, skb_headroom(skb)) < 0)
|
|
+ if (skb_cow(skb, skb_headroom(skb)) < 0) {
|
|
+ kfree_skb(skb);
|
|
return NULL;
|
|
+ }
|
|
+
|
|
memmove(skb->data - ETH_HLEN, skb->data - VLAN_ETH_HLEN, 2 * ETH_ALEN);
|
|
skb->mac_header += VLAN_HLEN;
|
|
return skb;
|
|
diff --git a/net/appletalk/ddp.c b/net/appletalk/ddp.c
|
|
index 0018daccdea9..8799e171addf 100644
|
|
--- a/net/appletalk/ddp.c
|
|
+++ b/net/appletalk/ddp.c
|
|
@@ -1489,8 +1489,6 @@ static int atalk_rcv(struct sk_buff *skb, struct net_device *dev,
|
|
goto drop;
|
|
|
|
/* Queue packet (standard) */
|
|
- skb->sk = sock;
|
|
-
|
|
if (sock_queue_rcv_skb(sock, skb) < 0)
|
|
goto drop;
|
|
|
|
@@ -1644,7 +1642,6 @@ static int atalk_sendmsg(struct kiocb *iocb, struct socket *sock, struct msghdr
|
|
if (!skb)
|
|
goto out;
|
|
|
|
- skb->sk = sk;
|
|
skb_reserve(skb, ddp_dl->header_length);
|
|
skb_reserve(skb, dev->hard_header_len);
|
|
skb->dev = dev;
|
|
diff --git a/net/core/dst.c b/net/core/dst.c
|
|
index df9cc810ec8e..c0e021871df8 100644
|
|
--- a/net/core/dst.c
|
|
+++ b/net/core/dst.c
|
|
@@ -267,6 +267,15 @@ again:
|
|
}
|
|
EXPORT_SYMBOL(dst_destroy);
|
|
|
|
+static void dst_destroy_rcu(struct rcu_head *head)
|
|
+{
|
|
+ struct dst_entry *dst = container_of(head, struct dst_entry, rcu_head);
|
|
+
|
|
+ dst = dst_destroy(dst);
|
|
+ if (dst)
|
|
+ __dst_free(dst);
|
|
+}
|
|
+
|
|
void dst_release(struct dst_entry *dst)
|
|
{
|
|
if (dst) {
|
|
@@ -274,11 +283,8 @@ void dst_release(struct dst_entry *dst)
|
|
|
|
newrefcnt = atomic_dec_return(&dst->__refcnt);
|
|
WARN_ON(newrefcnt < 0);
|
|
- if (unlikely(dst->flags & DST_NOCACHE) && !newrefcnt) {
|
|
- dst = dst_destroy(dst);
|
|
- if (dst)
|
|
- __dst_free(dst);
|
|
- }
|
|
+ if (unlikely(dst->flags & DST_NOCACHE) && !newrefcnt)
|
|
+ call_rcu(&dst->rcu_head, dst_destroy_rcu);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(dst_release);
|
|
diff --git a/net/dns_resolver/dns_query.c b/net/dns_resolver/dns_query.c
|
|
index c32be292c7e3..2022b46ab38f 100644
|
|
--- a/net/dns_resolver/dns_query.c
|
|
+++ b/net/dns_resolver/dns_query.c
|
|
@@ -150,7 +150,9 @@ int dns_query(const char *type, const char *name, size_t namelen,
|
|
if (!*_result)
|
|
goto put;
|
|
|
|
- memcpy(*_result, upayload->data, len + 1);
|
|
+ memcpy(*_result, upayload->data, len);
|
|
+ (*_result)[len] = '\0';
|
|
+
|
|
if (_expiry)
|
|
*_expiry = rkey->expiry;
|
|
|
|
diff --git a/net/ipv4/icmp.c b/net/ipv4/icmp.c
|
|
index 76e10b47e053..ea78ef5ac352 100644
|
|
--- a/net/ipv4/icmp.c
|
|
+++ b/net/ipv4/icmp.c
|
|
@@ -697,8 +697,6 @@ static void icmp_unreach(struct sk_buff *skb)
|
|
&iph->daddr);
|
|
} else {
|
|
info = ntohs(icmph->un.frag.mtu);
|
|
- if (!info)
|
|
- goto out;
|
|
}
|
|
break;
|
|
case ICMP_SR_FAILED:
|
|
diff --git a/net/ipv4/igmp.c b/net/ipv4/igmp.c
|
|
index 089b4af4fecc..38d63ca8a6b5 100644
|
|
--- a/net/ipv4/igmp.c
|
|
+++ b/net/ipv4/igmp.c
|
|
@@ -1874,6 +1874,10 @@ int ip_mc_leave_group(struct sock *sk, struct ip_mreqn *imr)
|
|
|
|
rtnl_lock();
|
|
in_dev = ip_mc_find_dev(net, imr);
|
|
+ if (!in_dev) {
|
|
+ ret = -ENODEV;
|
|
+ goto out;
|
|
+ }
|
|
ifindex = imr->imr_ifindex;
|
|
for (imlp = &inet->mc_list;
|
|
(iml = rtnl_dereference(*imlp)) != NULL;
|
|
@@ -1891,16 +1895,14 @@ int ip_mc_leave_group(struct sock *sk, struct ip_mreqn *imr)
|
|
|
|
*imlp = iml->next_rcu;
|
|
|
|
- if (in_dev)
|
|
- ip_mc_dec_group(in_dev, group);
|
|
+ ip_mc_dec_group(in_dev, group);
|
|
rtnl_unlock();
|
|
/* decrease mem now to avoid the memleak warning */
|
|
atomic_sub(sizeof(*iml), &sk->sk_omem_alloc);
|
|
kfree_rcu(iml, rcu);
|
|
return 0;
|
|
}
|
|
- if (!in_dev)
|
|
- ret = -ENODEV;
|
|
+out:
|
|
rtnl_unlock();
|
|
return ret;
|
|
}
|
|
diff --git a/net/ipv4/ip_options.c b/net/ipv4/ip_options.c
|
|
index ec7264514a82..089ed81d1878 100644
|
|
--- a/net/ipv4/ip_options.c
|
|
+++ b/net/ipv4/ip_options.c
|
|
@@ -288,6 +288,10 @@ int ip_options_compile(struct net *net,
|
|
optptr++;
|
|
continue;
|
|
}
|
|
+ if (unlikely(l < 2)) {
|
|
+ pp_ptr = optptr;
|
|
+ goto error;
|
|
+ }
|
|
optlen = optptr[1];
|
|
if (optlen<2 || optlen>l) {
|
|
pp_ptr = optptr;
|
|
diff --git a/net/ipv4/ip_tunnel.c b/net/ipv4/ip_tunnel.c
|
|
index fa6573264c8a..5642374cb751 100644
|
|
--- a/net/ipv4/ip_tunnel.c
|
|
+++ b/net/ipv4/ip_tunnel.c
|
|
@@ -166,6 +166,7 @@ struct ip_tunnel *ip_tunnel_lookup(struct ip_tunnel_net *itn,
|
|
|
|
hlist_for_each_entry_rcu(t, head, hash_node) {
|
|
if (remote != t->parms.iph.daddr ||
|
|
+ t->parms.iph.saddr != 0 ||
|
|
!(t->dev->flags & IFF_UP))
|
|
continue;
|
|
|
|
@@ -182,10 +183,11 @@ struct ip_tunnel *ip_tunnel_lookup(struct ip_tunnel_net *itn,
|
|
head = &itn->tunnels[hash];
|
|
|
|
hlist_for_each_entry_rcu(t, head, hash_node) {
|
|
- if ((local != t->parms.iph.saddr &&
|
|
- (local != t->parms.iph.daddr ||
|
|
- !ipv4_is_multicast(local))) ||
|
|
- !(t->dev->flags & IFF_UP))
|
|
+ if ((local != t->parms.iph.saddr || t->parms.iph.daddr != 0) &&
|
|
+ (local != t->parms.iph.daddr || !ipv4_is_multicast(local)))
|
|
+ continue;
|
|
+
|
|
+ if (!(t->dev->flags & IFF_UP))
|
|
continue;
|
|
|
|
if (!ip_tunnel_key_match(&t->parms, flags, key))
|
|
@@ -202,6 +204,8 @@ struct ip_tunnel *ip_tunnel_lookup(struct ip_tunnel_net *itn,
|
|
|
|
hlist_for_each_entry_rcu(t, head, hash_node) {
|
|
if (t->parms.i_key != key ||
|
|
+ t->parms.iph.saddr != 0 ||
|
|
+ t->parms.iph.daddr != 0 ||
|
|
!(t->dev->flags & IFF_UP))
|
|
continue;
|
|
|
|
diff --git a/net/ipv4/route.c b/net/ipv4/route.c
|
|
index 7256eef088b2..2b9887becb5c 100644
|
|
--- a/net/ipv4/route.c
|
|
+++ b/net/ipv4/route.c
|
|
@@ -985,20 +985,21 @@ void ipv4_sk_update_pmtu(struct sk_buff *skb, struct sock *sk, u32 mtu)
|
|
const struct iphdr *iph = (const struct iphdr *) skb->data;
|
|
struct flowi4 fl4;
|
|
struct rtable *rt;
|
|
- struct dst_entry *dst;
|
|
+ struct dst_entry *odst = NULL;
|
|
bool new = false;
|
|
|
|
bh_lock_sock(sk);
|
|
- rt = (struct rtable *) __sk_dst_get(sk);
|
|
+ odst = sk_dst_get(sk);
|
|
|
|
- if (sock_owned_by_user(sk) || !rt) {
|
|
+ if (sock_owned_by_user(sk) || !odst) {
|
|
__ipv4_sk_update_pmtu(skb, sk, mtu);
|
|
goto out;
|
|
}
|
|
|
|
__build_flow_key(&fl4, sk, iph, 0, 0, 0, 0, 0);
|
|
|
|
- if (!__sk_dst_check(sk, 0)) {
|
|
+ rt = (struct rtable *)odst;
|
|
+ if (odst->obsolete && odst->ops->check(odst, 0) == NULL) {
|
|
rt = ip_route_output_flow(sock_net(sk), &fl4, sk);
|
|
if (IS_ERR(rt))
|
|
goto out;
|
|
@@ -1008,8 +1009,7 @@ void ipv4_sk_update_pmtu(struct sk_buff *skb, struct sock *sk, u32 mtu)
|
|
|
|
__ip_rt_update_pmtu((struct rtable *) rt->dst.path, &fl4, mtu);
|
|
|
|
- dst = dst_check(&rt->dst, 0);
|
|
- if (!dst) {
|
|
+ if (!dst_check(&rt->dst, 0)) {
|
|
if (new)
|
|
dst_release(&rt->dst);
|
|
|
|
@@ -1021,10 +1021,11 @@ void ipv4_sk_update_pmtu(struct sk_buff *skb, struct sock *sk, u32 mtu)
|
|
}
|
|
|
|
if (new)
|
|
- __sk_dst_set(sk, &rt->dst);
|
|
+ sk_dst_set(sk, &rt->dst);
|
|
|
|
out:
|
|
bh_unlock_sock(sk);
|
|
+ dst_release(odst);
|
|
}
|
|
EXPORT_SYMBOL_GPL(ipv4_sk_update_pmtu);
|
|
|
|
diff --git a/net/ipv4/tcp.c b/net/ipv4/tcp.c
|
|
index 39bdb14b3214..5d4bd6ca3ab1 100644
|
|
--- a/net/ipv4/tcp.c
|
|
+++ b/net/ipv4/tcp.c
|
|
@@ -1065,7 +1065,7 @@ int tcp_sendmsg(struct kiocb *iocb, struct sock *sk, struct msghdr *msg,
|
|
if (unlikely(tp->repair)) {
|
|
if (tp->repair_queue == TCP_RECV_QUEUE) {
|
|
copied = tcp_send_rcvq(sk, msg, size);
|
|
- goto out;
|
|
+ goto out_nopush;
|
|
}
|
|
|
|
err = -EINVAL;
|
|
@@ -1238,6 +1238,7 @@ wait_for_memory:
|
|
out:
|
|
if (copied)
|
|
tcp_push(sk, flags, mss_now, tp->nonagle);
|
|
+out_nopush:
|
|
release_sock(sk);
|
|
return copied + copied_syn;
|
|
|
|
diff --git a/net/ipv4/tcp_input.c b/net/ipv4/tcp_input.c
|
|
index ba7d2b7ad9f9..ea7f52f3062d 100644
|
|
--- a/net/ipv4/tcp_input.c
|
|
+++ b/net/ipv4/tcp_input.c
|
|
@@ -1075,7 +1075,7 @@ static bool tcp_check_dsack(struct sock *sk, const struct sk_buff *ack_skb,
|
|
}
|
|
|
|
/* D-SACK for already forgotten data... Do dumb counting. */
|
|
- if (dup_sack && tp->undo_marker && tp->undo_retrans &&
|
|
+ if (dup_sack && tp->undo_marker && tp->undo_retrans > 0 &&
|
|
!after(end_seq_0, prior_snd_una) &&
|
|
after(end_seq_0, tp->undo_marker))
|
|
tp->undo_retrans--;
|
|
@@ -1130,7 +1130,7 @@ static int tcp_match_skb_to_sack(struct sock *sk, struct sk_buff *skb,
|
|
unsigned int new_len = (pkt_len / mss) * mss;
|
|
if (!in_sack && new_len < pkt_len) {
|
|
new_len += mss;
|
|
- if (new_len > skb->len)
|
|
+ if (new_len >= skb->len)
|
|
return 0;
|
|
}
|
|
pkt_len = new_len;
|
|
@@ -1154,7 +1154,7 @@ static u8 tcp_sacktag_one(struct sock *sk,
|
|
|
|
/* Account D-SACK for retransmitted packet. */
|
|
if (dup_sack && (sacked & TCPCB_RETRANS)) {
|
|
- if (tp->undo_marker && tp->undo_retrans &&
|
|
+ if (tp->undo_marker && tp->undo_retrans > 0 &&
|
|
after(end_seq, tp->undo_marker))
|
|
tp->undo_retrans--;
|
|
if (sacked & TCPCB_SACKED_ACKED)
|
|
@@ -1850,7 +1850,7 @@ static void tcp_clear_retrans_partial(struct tcp_sock *tp)
|
|
tp->lost_out = 0;
|
|
|
|
tp->undo_marker = 0;
|
|
- tp->undo_retrans = 0;
|
|
+ tp->undo_retrans = -1;
|
|
}
|
|
|
|
void tcp_clear_retrans(struct tcp_sock *tp)
|
|
@@ -2700,7 +2700,7 @@ static void tcp_enter_recovery(struct sock *sk, bool ece_ack)
|
|
|
|
tp->prior_ssthresh = 0;
|
|
tp->undo_marker = tp->snd_una;
|
|
- tp->undo_retrans = tp->retrans_out;
|
|
+ tp->undo_retrans = tp->retrans_out ? : -1;
|
|
|
|
if (inet_csk(sk)->icsk_ca_state < TCP_CA_CWR) {
|
|
if (!ece_ack)
|
|
diff --git a/net/ipv4/tcp_output.c b/net/ipv4/tcp_output.c
|
|
index 4a4e8746d1b2..56e29f0e230e 100644
|
|
--- a/net/ipv4/tcp_output.c
|
|
+++ b/net/ipv4/tcp_output.c
|
|
@@ -2428,13 +2428,15 @@ int tcp_retransmit_skb(struct sock *sk, struct sk_buff *skb)
|
|
if (!tp->retrans_stamp)
|
|
tp->retrans_stamp = TCP_SKB_CB(skb)->when;
|
|
|
|
- tp->undo_retrans += tcp_skb_pcount(skb);
|
|
-
|
|
/* snd_nxt is stored to detect loss of retransmitted segment,
|
|
* see tcp_input.c tcp_sacktag_write_queue().
|
|
*/
|
|
TCP_SKB_CB(skb)->ack_seq = tp->snd_nxt;
|
|
}
|
|
+
|
|
+ if (tp->undo_retrans < 0)
|
|
+ tp->undo_retrans = 0;
|
|
+ tp->undo_retrans += tcp_skb_pcount(skb);
|
|
return err;
|
|
}
|
|
|
|
diff --git a/net/netlink/af_netlink.c b/net/netlink/af_netlink.c
|
|
index be34adde692f..5ed562dfe743 100644
|
|
--- a/net/netlink/af_netlink.c
|
|
+++ b/net/netlink/af_netlink.c
|
|
@@ -500,7 +500,7 @@ static unsigned int netlink_poll(struct file *file, struct socket *sock,
|
|
while (nlk->cb != NULL && netlink_dump_space(nlk)) {
|
|
err = netlink_dump(sk);
|
|
if (err < 0) {
|
|
- sk->sk_err = err;
|
|
+ sk->sk_err = -err;
|
|
sk->sk_error_report(sk);
|
|
break;
|
|
}
|
|
@@ -2272,7 +2272,7 @@ static int netlink_recvmsg(struct kiocb *kiocb, struct socket *sock,
|
|
if (nlk->cb && atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf / 2) {
|
|
ret = netlink_dump(sk);
|
|
if (ret) {
|
|
- sk->sk_err = ret;
|
|
+ sk->sk_err = -ret;
|
|
sk->sk_error_report(sk);
|
|
}
|
|
}
|
|
diff --git a/net/sctp/sysctl.c b/net/sctp/sysctl.c
|
|
index fe0ba7488bdf..29299dcabfbb 100644
|
|
--- a/net/sctp/sysctl.c
|
|
+++ b/net/sctp/sysctl.c
|
|
@@ -368,8 +368,7 @@ static int proc_sctp_do_auth(struct ctl_table *ctl, int write,
|
|
tbl.data = &net->sctp.auth_enable;
|
|
|
|
ret = proc_dointvec(&tbl, write, buffer, lenp, ppos);
|
|
-
|
|
- if (write) {
|
|
+ if (write && ret == 0) {
|
|
struct sock *sk = net->sctp.ctl_sock;
|
|
|
|
net->sctp.auth_enable = new_value;
|
|
diff --git a/net/sctp/ulpevent.c b/net/sctp/ulpevent.c
|
|
index 10c018a5b9fe..ca907f2f5e5a 100644
|
|
--- a/net/sctp/ulpevent.c
|
|
+++ b/net/sctp/ulpevent.c
|
|
@@ -373,9 +373,10 @@ fail:
|
|
* specification [SCTP] and any extensions for a list of possible
|
|
* error formats.
|
|
*/
|
|
-struct sctp_ulpevent *sctp_ulpevent_make_remote_error(
|
|
- const struct sctp_association *asoc, struct sctp_chunk *chunk,
|
|
- __u16 flags, gfp_t gfp)
|
|
+struct sctp_ulpevent *
|
|
+sctp_ulpevent_make_remote_error(const struct sctp_association *asoc,
|
|
+ struct sctp_chunk *chunk, __u16 flags,
|
|
+ gfp_t gfp)
|
|
{
|
|
struct sctp_ulpevent *event;
|
|
struct sctp_remote_error *sre;
|
|
@@ -394,8 +395,7 @@ struct sctp_ulpevent *sctp_ulpevent_make_remote_error(
|
|
/* Copy the skb to a new skb with room for us to prepend
|
|
* notification with.
|
|
*/
|
|
- skb = skb_copy_expand(chunk->skb, sizeof(struct sctp_remote_error),
|
|
- 0, gfp);
|
|
+ skb = skb_copy_expand(chunk->skb, sizeof(*sre), 0, gfp);
|
|
|
|
/* Pull off the rest of the cause TLV from the chunk. */
|
|
skb_pull(chunk->skb, elen);
|
|
@@ -406,62 +406,21 @@ struct sctp_ulpevent *sctp_ulpevent_make_remote_error(
|
|
event = sctp_skb2event(skb);
|
|
sctp_ulpevent_init(event, MSG_NOTIFICATION, skb->truesize);
|
|
|
|
- sre = (struct sctp_remote_error *)
|
|
- skb_push(skb, sizeof(struct sctp_remote_error));
|
|
+ sre = (struct sctp_remote_error *) skb_push(skb, sizeof(*sre));
|
|
|
|
/* Trim the buffer to the right length. */
|
|
- skb_trim(skb, sizeof(struct sctp_remote_error) + elen);
|
|
+ skb_trim(skb, sizeof(*sre) + elen);
|
|
|
|
- /* Socket Extensions for SCTP
|
|
- * 5.3.1.3 SCTP_REMOTE_ERROR
|
|
- *
|
|
- * sre_type:
|
|
- * It should be SCTP_REMOTE_ERROR.
|
|
- */
|
|
+ /* RFC6458, Section 6.1.3. SCTP_REMOTE_ERROR */
|
|
+ memset(sre, 0, sizeof(*sre));
|
|
sre->sre_type = SCTP_REMOTE_ERROR;
|
|
-
|
|
- /*
|
|
- * Socket Extensions for SCTP
|
|
- * 5.3.1.3 SCTP_REMOTE_ERROR
|
|
- *
|
|
- * sre_flags: 16 bits (unsigned integer)
|
|
- * Currently unused.
|
|
- */
|
|
sre->sre_flags = 0;
|
|
-
|
|
- /* Socket Extensions for SCTP
|
|
- * 5.3.1.3 SCTP_REMOTE_ERROR
|
|
- *
|
|
- * sre_length: sizeof (__u32)
|
|
- *
|
|
- * This field is the total length of the notification data,
|
|
- * including the notification header.
|
|
- */
|
|
sre->sre_length = skb->len;
|
|
-
|
|
- /* Socket Extensions for SCTP
|
|
- * 5.3.1.3 SCTP_REMOTE_ERROR
|
|
- *
|
|
- * sre_error: 16 bits (unsigned integer)
|
|
- * This value represents one of the Operational Error causes defined in
|
|
- * the SCTP specification, in network byte order.
|
|
- */
|
|
sre->sre_error = cause;
|
|
-
|
|
- /* Socket Extensions for SCTP
|
|
- * 5.3.1.3 SCTP_REMOTE_ERROR
|
|
- *
|
|
- * sre_assoc_id: sizeof (sctp_assoc_t)
|
|
- *
|
|
- * The association id field, holds the identifier for the association.
|
|
- * All notifications for a given association have the same association
|
|
- * identifier. For TCP style socket, this field is ignored.
|
|
- */
|
|
sctp_ulpevent_set_owner(event, asoc);
|
|
sre->sre_assoc_id = sctp_assoc2id(asoc);
|
|
|
|
return event;
|
|
-
|
|
fail:
|
|
return NULL;
|
|
}
|
|
@@ -906,7 +865,9 @@ __u16 sctp_ulpevent_get_notification_type(const struct sctp_ulpevent *event)
|
|
return notification->sn_header.sn_type;
|
|
}
|
|
|
|
-/* Copy out the sndrcvinfo into a msghdr. */
|
|
+/* RFC6458, Section 5.3.2. SCTP Header Information Structure
|
|
+ * (SCTP_SNDRCV, DEPRECATED)
|
|
+ */
|
|
void sctp_ulpevent_read_sndrcvinfo(const struct sctp_ulpevent *event,
|
|
struct msghdr *msghdr)
|
|
{
|
|
@@ -915,74 +876,21 @@ void sctp_ulpevent_read_sndrcvinfo(const struct sctp_ulpevent *event,
|
|
if (sctp_ulpevent_is_notification(event))
|
|
return;
|
|
|
|
- /* Sockets API Extensions for SCTP
|
|
- * Section 5.2.2 SCTP Header Information Structure (SCTP_SNDRCV)
|
|
- *
|
|
- * sinfo_stream: 16 bits (unsigned integer)
|
|
- *
|
|
- * For recvmsg() the SCTP stack places the message's stream number in
|
|
- * this value.
|
|
- */
|
|
+ memset(&sinfo, 0, sizeof(sinfo));
|
|
sinfo.sinfo_stream = event->stream;
|
|
- /* sinfo_ssn: 16 bits (unsigned integer)
|
|
- *
|
|
- * For recvmsg() this value contains the stream sequence number that
|
|
- * the remote endpoint placed in the DATA chunk. For fragmented
|
|
- * messages this is the same number for all deliveries of the message
|
|
- * (if more than one recvmsg() is needed to read the message).
|
|
- */
|
|
sinfo.sinfo_ssn = event->ssn;
|
|
- /* sinfo_ppid: 32 bits (unsigned integer)
|
|
- *
|
|
- * In recvmsg() this value is
|
|
- * the same information that was passed by the upper layer in the peer
|
|
- * application. Please note that byte order issues are NOT accounted
|
|
- * for and this information is passed opaquely by the SCTP stack from
|
|
- * one end to the other.
|
|
- */
|
|
sinfo.sinfo_ppid = event->ppid;
|
|
- /* sinfo_flags: 16 bits (unsigned integer)
|
|
- *
|
|
- * This field may contain any of the following flags and is composed of
|
|
- * a bitwise OR of these values.
|
|
- *
|
|
- * recvmsg() flags:
|
|
- *
|
|
- * SCTP_UNORDERED - This flag is present when the message was sent
|
|
- * non-ordered.
|
|
- */
|
|
sinfo.sinfo_flags = event->flags;
|
|
- /* sinfo_tsn: 32 bit (unsigned integer)
|
|
- *
|
|
- * For the receiving side, this field holds a TSN that was
|
|
- * assigned to one of the SCTP Data Chunks.
|
|
- */
|
|
sinfo.sinfo_tsn = event->tsn;
|
|
- /* sinfo_cumtsn: 32 bit (unsigned integer)
|
|
- *
|
|
- * This field will hold the current cumulative TSN as
|
|
- * known by the underlying SCTP layer. Note this field is
|
|
- * ignored when sending and only valid for a receive
|
|
- * operation when sinfo_flags are set to SCTP_UNORDERED.
|
|
- */
|
|
sinfo.sinfo_cumtsn = event->cumtsn;
|
|
- /* sinfo_assoc_id: sizeof (sctp_assoc_t)
|
|
- *
|
|
- * The association handle field, sinfo_assoc_id, holds the identifier
|
|
- * for the association announced in the COMMUNICATION_UP notification.
|
|
- * All notifications for a given association have the same identifier.
|
|
- * Ignored for one-to-one style sockets.
|
|
- */
|
|
sinfo.sinfo_assoc_id = sctp_assoc2id(event->asoc);
|
|
-
|
|
- /* context value that is set via SCTP_CONTEXT socket option. */
|
|
+ /* Context value that is set via SCTP_CONTEXT socket option. */
|
|
sinfo.sinfo_context = event->asoc->default_rcv_context;
|
|
-
|
|
/* These fields are not used while receiving. */
|
|
sinfo.sinfo_timetolive = 0;
|
|
|
|
put_cmsg(msghdr, IPPROTO_SCTP, SCTP_SNDRCV,
|
|
- sizeof(struct sctp_sndrcvinfo), (void *)&sinfo);
|
|
+ sizeof(sinfo), &sinfo);
|
|
}
|
|
|
|
/* Do accounting for bytes received and hold a reference to the association
|
|
diff --git a/net/tipc/bcast.c b/net/tipc/bcast.c
|
|
index e5f3da507823..bf2755419ec6 100644
|
|
--- a/net/tipc/bcast.c
|
|
+++ b/net/tipc/bcast.c
|
|
@@ -531,6 +531,7 @@ receive:
|
|
|
|
buf = node->bclink.deferred_head;
|
|
node->bclink.deferred_head = buf->next;
|
|
+ buf->next = NULL;
|
|
node->bclink.deferred_size--;
|
|
goto receive;
|
|
}
|