mirror of
https://github.com/Fishwaldo/build.git
synced 2025-03-27 01:02:19 +00:00
254 lines
8.9 KiB
Diff
254 lines
8.9 KiB
Diff
diff --git a/Makefile b/Makefile
|
|
index 030f5af05f4e..ee92a12e3a4b 100644
|
|
--- a/Makefile
|
|
+++ b/Makefile
|
|
@@ -1,6 +1,6 @@
|
|
VERSION = 4
|
|
PATCHLEVEL = 4
|
|
-SUBLEVEL = 146
|
|
+SUBLEVEL = 147
|
|
EXTRAVERSION =
|
|
NAME = Blurry Fish Butt
|
|
|
|
diff --git a/drivers/i2c/busses/i2c-imx.c b/drivers/i2c/busses/i2c-imx.c
|
|
index d4d853680ae4..a4abf7dc9576 100644
|
|
--- a/drivers/i2c/busses/i2c-imx.c
|
|
+++ b/drivers/i2c/busses/i2c-imx.c
|
|
@@ -382,6 +382,7 @@ static int i2c_imx_dma_xfer(struct imx_i2c_struct *i2c_imx,
|
|
goto err_desc;
|
|
}
|
|
|
|
+ reinit_completion(&dma->cmd_complete);
|
|
txdesc->callback = i2c_imx_dma_callback;
|
|
txdesc->callback_param = i2c_imx;
|
|
if (dma_submit_error(dmaengine_submit(txdesc))) {
|
|
@@ -631,7 +632,6 @@ static int i2c_imx_dma_write(struct imx_i2c_struct *i2c_imx,
|
|
* The first byte must be transmitted by the CPU.
|
|
*/
|
|
imx_i2c_write_reg(msgs->addr << 1, i2c_imx, IMX_I2C_I2DR);
|
|
- reinit_completion(&i2c_imx->dma->cmd_complete);
|
|
time_left = wait_for_completion_timeout(
|
|
&i2c_imx->dma->cmd_complete,
|
|
msecs_to_jiffies(DMA_TIMEOUT));
|
|
@@ -690,7 +690,6 @@ static int i2c_imx_dma_read(struct imx_i2c_struct *i2c_imx,
|
|
if (result)
|
|
return result;
|
|
|
|
- reinit_completion(&i2c_imx->dma->cmd_complete);
|
|
time_left = wait_for_completion_timeout(
|
|
&i2c_imx->dma->cmd_complete,
|
|
msecs_to_jiffies(DMA_TIMEOUT));
|
|
diff --git a/drivers/pci/pci-acpi.c b/drivers/pci/pci-acpi.c
|
|
index a32ba753e413..afaf13474796 100644
|
|
--- a/drivers/pci/pci-acpi.c
|
|
+++ b/drivers/pci/pci-acpi.c
|
|
@@ -543,7 +543,7 @@ void acpi_pci_add_bus(struct pci_bus *bus)
|
|
union acpi_object *obj;
|
|
struct pci_host_bridge *bridge;
|
|
|
|
- if (acpi_pci_disabled || !bus->bridge)
|
|
+ if (acpi_pci_disabled || !bus->bridge || !ACPI_HANDLE(bus->bridge))
|
|
return;
|
|
|
|
acpi_pci_slot_enumerate(bus);
|
|
diff --git a/drivers/scsi/qla2xxx/qla_init.c b/drivers/scsi/qla2xxx/qla_init.c
|
|
index a9eb3cd453be..41a646696bab 100644
|
|
--- a/drivers/scsi/qla2xxx/qla_init.c
|
|
+++ b/drivers/scsi/qla2xxx/qla_init.c
|
|
@@ -325,11 +325,10 @@ qla2x00_async_tm_cmd(fc_port_t *fcport, uint32_t flags, uint32_t lun,
|
|
|
|
wait_for_completion(&tm_iocb->u.tmf.comp);
|
|
|
|
- rval = tm_iocb->u.tmf.comp_status == CS_COMPLETE ?
|
|
- QLA_SUCCESS : QLA_FUNCTION_FAILED;
|
|
+ rval = tm_iocb->u.tmf.data;
|
|
|
|
- if ((rval != QLA_SUCCESS) || tm_iocb->u.tmf.data) {
|
|
- ql_dbg(ql_dbg_taskm, vha, 0x8030,
|
|
+ if (rval != QLA_SUCCESS) {
|
|
+ ql_log(ql_log_warn, vha, 0x8030,
|
|
"TM IOCB failed (%x).\n", rval);
|
|
}
|
|
|
|
diff --git a/drivers/scsi/qla2xxx/qla_os.c b/drivers/scsi/qla2xxx/qla_os.c
|
|
index 5cbf20ab94aa..18b19744398a 100644
|
|
--- a/drivers/scsi/qla2xxx/qla_os.c
|
|
+++ b/drivers/scsi/qla2xxx/qla_os.c
|
|
@@ -4938,8 +4938,9 @@ qla2x00_do_dpc(void *data)
|
|
}
|
|
}
|
|
|
|
- if (test_and_clear_bit(ISP_ABORT_NEEDED,
|
|
- &base_vha->dpc_flags)) {
|
|
+ if (test_and_clear_bit
|
|
+ (ISP_ABORT_NEEDED, &base_vha->dpc_flags) &&
|
|
+ !test_bit(UNLOADING, &base_vha->dpc_flags)) {
|
|
|
|
ql_dbg(ql_dbg_dpc, base_vha, 0x4007,
|
|
"ISP abort scheduled.\n");
|
|
diff --git a/fs/ext4/super.c b/fs/ext4/super.c
|
|
index 49af3c50b263..3e4d8ac1974e 100644
|
|
--- a/fs/ext4/super.c
|
|
+++ b/fs/ext4/super.c
|
|
@@ -2102,7 +2102,7 @@ static int ext4_check_descriptors(struct super_block *sb,
|
|
struct ext4_sb_info *sbi = EXT4_SB(sb);
|
|
ext4_fsblk_t first_block = le32_to_cpu(sbi->s_es->s_first_data_block);
|
|
ext4_fsblk_t last_block;
|
|
- ext4_fsblk_t last_bg_block = sb_block + ext4_bg_num_gdb(sb, 0) + 1;
|
|
+ ext4_fsblk_t last_bg_block = sb_block + ext4_bg_num_gdb(sb, 0);
|
|
ext4_fsblk_t block_bitmap;
|
|
ext4_fsblk_t inode_bitmap;
|
|
ext4_fsblk_t inode_table;
|
|
@@ -3777,13 +3777,13 @@ static int ext4_fill_super(struct super_block *sb, void *data, int silent)
|
|
goto failed_mount2;
|
|
}
|
|
}
|
|
+ sbi->s_gdb_count = db_count;
|
|
if (!ext4_check_descriptors(sb, logical_sb_block, &first_not_zeroed)) {
|
|
ext4_msg(sb, KERN_ERR, "group descriptors corrupted!");
|
|
ret = -EFSCORRUPTED;
|
|
goto failed_mount2;
|
|
}
|
|
|
|
- sbi->s_gdb_count = db_count;
|
|
get_random_bytes(&sbi->s_next_generation, sizeof(u32));
|
|
spin_lock_init(&sbi->s_next_gen_lock);
|
|
|
|
diff --git a/fs/jfs/xattr.c b/fs/jfs/xattr.c
|
|
index 48b15a6e5558..40a26a542341 100644
|
|
--- a/fs/jfs/xattr.c
|
|
+++ b/fs/jfs/xattr.c
|
|
@@ -493,15 +493,17 @@ static int ea_get(struct inode *inode, struct ea_buffer *ea_buf, int min_size)
|
|
if (size > PSIZE) {
|
|
/*
|
|
* To keep the rest of the code simple. Allocate a
|
|
- * contiguous buffer to work with
|
|
+ * contiguous buffer to work with. Make the buffer large
|
|
+ * enough to make use of the whole extent.
|
|
*/
|
|
- ea_buf->xattr = kmalloc(size, GFP_KERNEL);
|
|
+ ea_buf->max_size = (size + sb->s_blocksize - 1) &
|
|
+ ~(sb->s_blocksize - 1);
|
|
+
|
|
+ ea_buf->xattr = kmalloc(ea_buf->max_size, GFP_KERNEL);
|
|
if (ea_buf->xattr == NULL)
|
|
return -ENOMEM;
|
|
|
|
ea_buf->flag = EA_MALLOC;
|
|
- ea_buf->max_size = (size + sb->s_blocksize - 1) &
|
|
- ~(sb->s_blocksize - 1);
|
|
|
|
if (ea_size == 0)
|
|
return 0;
|
|
diff --git a/include/linux/ring_buffer.h b/include/linux/ring_buffer.h
|
|
index 4acc552e9279..19d0778ec382 100644
|
|
--- a/include/linux/ring_buffer.h
|
|
+++ b/include/linux/ring_buffer.h
|
|
@@ -162,6 +162,7 @@ void ring_buffer_record_enable(struct ring_buffer *buffer);
|
|
void ring_buffer_record_off(struct ring_buffer *buffer);
|
|
void ring_buffer_record_on(struct ring_buffer *buffer);
|
|
int ring_buffer_record_is_on(struct ring_buffer *buffer);
|
|
+int ring_buffer_record_is_set_on(struct ring_buffer *buffer);
|
|
void ring_buffer_record_disable_cpu(struct ring_buffer *buffer, int cpu);
|
|
void ring_buffer_record_enable_cpu(struct ring_buffer *buffer, int cpu);
|
|
|
|
diff --git a/kernel/irq/manage.c b/kernel/irq/manage.c
|
|
index 5f55a8bf5264..0df2b44dac7c 100644
|
|
--- a/kernel/irq/manage.c
|
|
+++ b/kernel/irq/manage.c
|
|
@@ -1012,6 +1012,13 @@ static int irq_setup_forced_threading(struct irqaction *new)
|
|
if (new->flags & (IRQF_NO_THREAD | IRQF_PERCPU | IRQF_ONESHOT))
|
|
return 0;
|
|
|
|
+ /*
|
|
+ * No further action required for interrupts which are requested as
|
|
+ * threaded interrupts already
|
|
+ */
|
|
+ if (new->handler == irq_default_primary_handler)
|
|
+ return 0;
|
|
+
|
|
new->flags |= IRQF_ONESHOT;
|
|
|
|
/*
|
|
@@ -1019,7 +1026,7 @@ static int irq_setup_forced_threading(struct irqaction *new)
|
|
* thread handler. We force thread them as well by creating a
|
|
* secondary action.
|
|
*/
|
|
- if (new->handler != irq_default_primary_handler && new->thread_fn) {
|
|
+ if (new->handler && new->thread_fn) {
|
|
/* Allocate the secondary action */
|
|
new->secondary = kzalloc(sizeof(struct irqaction), GFP_KERNEL);
|
|
if (!new->secondary)
|
|
diff --git a/kernel/time/tick-sched.c b/kernel/time/tick-sched.c
|
|
index e5d228f7224c..5ad2e852e9f6 100644
|
|
--- a/kernel/time/tick-sched.c
|
|
+++ b/kernel/time/tick-sched.c
|
|
@@ -570,7 +570,7 @@ static void tick_nohz_restart(struct tick_sched *ts, ktime_t now)
|
|
|
|
static inline bool local_timer_softirq_pending(void)
|
|
{
|
|
- return local_softirq_pending() & TIMER_SOFTIRQ;
|
|
+ return local_softirq_pending() & BIT(TIMER_SOFTIRQ);
|
|
}
|
|
|
|
static ktime_t tick_nohz_stop_sched_tick(struct tick_sched *ts,
|
|
diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c
|
|
index d9cd6191760b..fdaa88f38aec 100644
|
|
--- a/kernel/trace/ring_buffer.c
|
|
+++ b/kernel/trace/ring_buffer.c
|
|
@@ -3141,6 +3141,22 @@ int ring_buffer_record_is_on(struct ring_buffer *buffer)
|
|
return !atomic_read(&buffer->record_disabled);
|
|
}
|
|
|
|
+/**
|
|
+ * ring_buffer_record_is_set_on - return true if the ring buffer is set writable
|
|
+ * @buffer: The ring buffer to see if write is set enabled
|
|
+ *
|
|
+ * Returns true if the ring buffer is set writable by ring_buffer_record_on().
|
|
+ * Note that this does NOT mean it is in a writable state.
|
|
+ *
|
|
+ * It may return true when the ring buffer has been disabled by
|
|
+ * ring_buffer_record_disable(), as that is a temporary disabling of
|
|
+ * the ring buffer.
|
|
+ */
|
|
+int ring_buffer_record_is_set_on(struct ring_buffer *buffer)
|
|
+{
|
|
+ return !(atomic_read(&buffer->record_disabled) & RB_BUFFER_OFF);
|
|
+}
|
|
+
|
|
/**
|
|
* ring_buffer_record_disable_cpu - stop all writes into the cpu_buffer
|
|
* @buffer: The ring buffer to stop writes to.
|
|
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c
|
|
index 8aef4e63ac57..1b980a8ef791 100644
|
|
--- a/kernel/trace/trace.c
|
|
+++ b/kernel/trace/trace.c
|
|
@@ -1088,6 +1088,12 @@ update_max_tr(struct trace_array *tr, struct task_struct *tsk, int cpu)
|
|
|
|
arch_spin_lock(&tr->max_lock);
|
|
|
|
+ /* Inherit the recordable setting from trace_buffer */
|
|
+ if (ring_buffer_record_is_set_on(tr->trace_buffer.buffer))
|
|
+ ring_buffer_record_on(tr->max_buffer.buffer);
|
|
+ else
|
|
+ ring_buffer_record_off(tr->max_buffer.buffer);
|
|
+
|
|
buf = tr->trace_buffer.buffer;
|
|
tr->trace_buffer.buffer = tr->max_buffer.buffer;
|
|
tr->max_buffer.buffer = buf;
|
|
diff --git a/net/netlink/af_netlink.c b/net/netlink/af_netlink.c
|
|
index 9708fff318d5..bf292010760a 100644
|
|
--- a/net/netlink/af_netlink.c
|
|
+++ b/net/netlink/af_netlink.c
|
|
@@ -986,6 +986,11 @@ static int netlink_bind(struct socket *sock, struct sockaddr *addr,
|
|
return err;
|
|
}
|
|
|
|
+ if (nlk->ngroups == 0)
|
|
+ groups = 0;
|
|
+ else if (nlk->ngroups < 8*sizeof(groups))
|
|
+ groups &= (1UL << nlk->ngroups) - 1;
|
|
+
|
|
bound = nlk->bound;
|
|
if (bound) {
|
|
/* Ensure nlk->portid is up-to-date. */
|