mirror of
https://github.com/Fishwaldo/linux-bl808.git
synced 2025-04-05 05:44:43 +00:00
ioatdma: fix race between updating ioat->head and IOAT_COMPLETION_PENDING
There is a race that can hit during __cleanup() when the ioat->head pointer is incremented during descriptor submission. The __cleanup() can clear the PENDING flag when it does not see any active descriptors. This causes new submitted descriptors to be ignored because the COMPLETION_PENDING flag is cleared. This was introduced when code was adapted from ioatdma v1 to ioatdma v2. For v2 and v3, IOAT_COMPLETION_PENDING flag will be abandoned and a new flag IOAT_CHAN_ACTIVE will be utilized. This flag will also be protected under the prep_lock when being modified in order to avoid the race. Signed-off-by: Dave Jiang <dave.jiang@intel.com> Reviewed-by: Dan Williams <djbw@fb.com> Signed-off-by: Vinod Koul <vinod.koul@intel.com>
This commit is contained in:
parent
cfdf5b6cc5
commit
4dec23d771
3 changed files with 128 additions and 97 deletions
|
@ -97,6 +97,7 @@ struct ioat_chan_common {
|
||||||
#define IOAT_KOBJ_INIT_FAIL 3
|
#define IOAT_KOBJ_INIT_FAIL 3
|
||||||
#define IOAT_RESHAPE_PENDING 4
|
#define IOAT_RESHAPE_PENDING 4
|
||||||
#define IOAT_RUN 5
|
#define IOAT_RUN 5
|
||||||
|
#define IOAT_CHAN_ACTIVE 6
|
||||||
struct timer_list timer;
|
struct timer_list timer;
|
||||||
#define COMPLETION_TIMEOUT msecs_to_jiffies(100)
|
#define COMPLETION_TIMEOUT msecs_to_jiffies(100)
|
||||||
#define IDLE_TIMEOUT msecs_to_jiffies(2000)
|
#define IDLE_TIMEOUT msecs_to_jiffies(2000)
|
||||||
|
|
|
@ -269,12 +269,36 @@ static void ioat2_restart_channel(struct ioat2_dma_chan *ioat)
|
||||||
__ioat2_restart_chan(ioat);
|
__ioat2_restart_chan(ioat);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void check_active(struct ioat2_dma_chan *ioat)
|
||||||
|
{
|
||||||
|
struct ioat_chan_common *chan = &ioat->base;
|
||||||
|
|
||||||
|
if (ioat2_ring_active(ioat)) {
|
||||||
|
mod_timer(&chan->timer, jiffies + COMPLETION_TIMEOUT);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (test_and_clear_bit(IOAT_CHAN_ACTIVE, &chan->state))
|
||||||
|
mod_timer(&chan->timer, jiffies + IDLE_TIMEOUT);
|
||||||
|
else if (ioat->alloc_order > ioat_get_alloc_order()) {
|
||||||
|
/* if the ring is idle, empty, and oversized try to step
|
||||||
|
* down the size
|
||||||
|
*/
|
||||||
|
reshape_ring(ioat, ioat->alloc_order - 1);
|
||||||
|
|
||||||
|
/* keep shrinking until we get back to our minimum
|
||||||
|
* default size
|
||||||
|
*/
|
||||||
|
if (ioat->alloc_order > ioat_get_alloc_order())
|
||||||
|
mod_timer(&chan->timer, jiffies + IDLE_TIMEOUT);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
void ioat2_timer_event(unsigned long data)
|
void ioat2_timer_event(unsigned long data)
|
||||||
{
|
{
|
||||||
struct ioat2_dma_chan *ioat = to_ioat2_chan((void *) data);
|
struct ioat2_dma_chan *ioat = to_ioat2_chan((void *) data);
|
||||||
struct ioat_chan_common *chan = &ioat->base;
|
struct ioat_chan_common *chan = &ioat->base;
|
||||||
|
|
||||||
if (test_bit(IOAT_COMPLETION_PENDING, &chan->state)) {
|
|
||||||
dma_addr_t phys_complete;
|
dma_addr_t phys_complete;
|
||||||
u64 status;
|
u64 status;
|
||||||
|
|
||||||
|
@ -300,37 +324,28 @@ void ioat2_timer_event(unsigned long data)
|
||||||
* forceful with a restart
|
* forceful with a restart
|
||||||
*/
|
*/
|
||||||
spin_lock_bh(&chan->cleanup_lock);
|
spin_lock_bh(&chan->cleanup_lock);
|
||||||
if (ioat_cleanup_preamble(chan, &phys_complete)) {
|
if (ioat_cleanup_preamble(chan, &phys_complete))
|
||||||
__cleanup(ioat, phys_complete);
|
__cleanup(ioat, phys_complete);
|
||||||
} else if (test_bit(IOAT_COMPLETION_ACK, &chan->state)) {
|
else if (test_bit(IOAT_COMPLETION_ACK, &chan->state)) {
|
||||||
spin_lock_bh(&ioat->prep_lock);
|
spin_lock_bh(&ioat->prep_lock);
|
||||||
ioat2_restart_channel(ioat);
|
ioat2_restart_channel(ioat);
|
||||||
spin_unlock_bh(&ioat->prep_lock);
|
spin_unlock_bh(&ioat->prep_lock);
|
||||||
|
spin_unlock_bh(&chan->cleanup_lock);
|
||||||
|
return;
|
||||||
} else {
|
} else {
|
||||||
set_bit(IOAT_COMPLETION_ACK, &chan->state);
|
set_bit(IOAT_COMPLETION_ACK, &chan->state);
|
||||||
mod_timer(&chan->timer, jiffies + COMPLETION_TIMEOUT);
|
mod_timer(&chan->timer, jiffies + COMPLETION_TIMEOUT);
|
||||||
}
|
}
|
||||||
spin_unlock_bh(&chan->cleanup_lock);
|
|
||||||
} else {
|
|
||||||
u16 active;
|
|
||||||
|
|
||||||
/* if the ring is idle, empty, and oversized try to step
|
|
||||||
* down the size
|
if (ioat2_ring_active(ioat))
|
||||||
*/
|
mod_timer(&chan->timer, jiffies + COMPLETION_TIMEOUT);
|
||||||
spin_lock_bh(&chan->cleanup_lock);
|
else {
|
||||||
spin_lock_bh(&ioat->prep_lock);
|
spin_lock_bh(&ioat->prep_lock);
|
||||||
active = ioat2_ring_active(ioat);
|
check_active(ioat);
|
||||||
if (active == 0 && ioat->alloc_order > ioat_get_alloc_order())
|
|
||||||
reshape_ring(ioat, ioat->alloc_order-1);
|
|
||||||
spin_unlock_bh(&ioat->prep_lock);
|
spin_unlock_bh(&ioat->prep_lock);
|
||||||
spin_unlock_bh(&chan->cleanup_lock);
|
|
||||||
|
|
||||||
/* keep shrinking until we get back to our minimum
|
|
||||||
* default size
|
|
||||||
*/
|
|
||||||
if (ioat->alloc_order > ioat_get_alloc_order())
|
|
||||||
mod_timer(&chan->timer, jiffies + IDLE_TIMEOUT);
|
|
||||||
}
|
}
|
||||||
|
spin_unlock_bh(&chan->cleanup_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int ioat2_reset_hw(struct ioat_chan_common *chan)
|
static int ioat2_reset_hw(struct ioat_chan_common *chan)
|
||||||
|
@ -404,7 +419,7 @@ static dma_cookie_t ioat2_tx_submit_unlock(struct dma_async_tx_descriptor *tx)
|
||||||
cookie = dma_cookie_assign(tx);
|
cookie = dma_cookie_assign(tx);
|
||||||
dev_dbg(to_dev(&ioat->base), "%s: cookie: %d\n", __func__, cookie);
|
dev_dbg(to_dev(&ioat->base), "%s: cookie: %d\n", __func__, cookie);
|
||||||
|
|
||||||
if (!test_and_set_bit(IOAT_COMPLETION_PENDING, &chan->state))
|
if (!test_and_set_bit(IOAT_CHAN_ACTIVE, &chan->state))
|
||||||
mod_timer(&chan->timer, jiffies + COMPLETION_TIMEOUT);
|
mod_timer(&chan->timer, jiffies + COMPLETION_TIMEOUT);
|
||||||
|
|
||||||
/* make descriptor updates visible before advancing ioat->head,
|
/* make descriptor updates visible before advancing ioat->head,
|
||||||
|
|
|
@ -342,12 +342,36 @@ static void ioat3_restart_channel(struct ioat2_dma_chan *ioat)
|
||||||
__ioat2_restart_chan(ioat);
|
__ioat2_restart_chan(ioat);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void ioat3_timer_event(unsigned long data)
|
static void check_active(struct ioat2_dma_chan *ioat)
|
||||||
|
{
|
||||||
|
struct ioat_chan_common *chan = &ioat->base;
|
||||||
|
|
||||||
|
if (ioat2_ring_active(ioat)) {
|
||||||
|
mod_timer(&chan->timer, jiffies + COMPLETION_TIMEOUT);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (test_and_clear_bit(IOAT_CHAN_ACTIVE, &chan->state))
|
||||||
|
mod_timer(&chan->timer, jiffies + IDLE_TIMEOUT);
|
||||||
|
else if (ioat->alloc_order > ioat_get_alloc_order()) {
|
||||||
|
/* if the ring is idle, empty, and oversized try to step
|
||||||
|
* down the size
|
||||||
|
*/
|
||||||
|
reshape_ring(ioat, ioat->alloc_order - 1);
|
||||||
|
|
||||||
|
/* keep shrinking until we get back to our minimum
|
||||||
|
* default size
|
||||||
|
*/
|
||||||
|
if (ioat->alloc_order > ioat_get_alloc_order())
|
||||||
|
mod_timer(&chan->timer, jiffies + IDLE_TIMEOUT);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
void ioat3_timer_event(unsigned long data)
|
||||||
{
|
{
|
||||||
struct ioat2_dma_chan *ioat = to_ioat2_chan((void *) data);
|
struct ioat2_dma_chan *ioat = to_ioat2_chan((void *) data);
|
||||||
struct ioat_chan_common *chan = &ioat->base;
|
struct ioat_chan_common *chan = &ioat->base;
|
||||||
|
|
||||||
if (test_bit(IOAT_COMPLETION_PENDING, &chan->state)) {
|
|
||||||
dma_addr_t phys_complete;
|
dma_addr_t phys_complete;
|
||||||
u64 status;
|
u64 status;
|
||||||
|
|
||||||
|
@ -379,31 +403,22 @@ static void ioat3_timer_event(unsigned long data)
|
||||||
spin_lock_bh(&ioat->prep_lock);
|
spin_lock_bh(&ioat->prep_lock);
|
||||||
ioat3_restart_channel(ioat);
|
ioat3_restart_channel(ioat);
|
||||||
spin_unlock_bh(&ioat->prep_lock);
|
spin_unlock_bh(&ioat->prep_lock);
|
||||||
|
spin_unlock_bh(&chan->cleanup_lock);
|
||||||
|
return;
|
||||||
} else {
|
} else {
|
||||||
set_bit(IOAT_COMPLETION_ACK, &chan->state);
|
set_bit(IOAT_COMPLETION_ACK, &chan->state);
|
||||||
mod_timer(&chan->timer, jiffies + COMPLETION_TIMEOUT);
|
mod_timer(&chan->timer, jiffies + COMPLETION_TIMEOUT);
|
||||||
}
|
}
|
||||||
spin_unlock_bh(&chan->cleanup_lock);
|
|
||||||
} else {
|
|
||||||
u16 active;
|
|
||||||
|
|
||||||
/* if the ring is idle, empty, and oversized try to step
|
|
||||||
* down the size
|
if (ioat2_ring_active(ioat))
|
||||||
*/
|
mod_timer(&chan->timer, jiffies + COMPLETION_TIMEOUT);
|
||||||
spin_lock_bh(&chan->cleanup_lock);
|
else {
|
||||||
spin_lock_bh(&ioat->prep_lock);
|
spin_lock_bh(&ioat->prep_lock);
|
||||||
active = ioat2_ring_active(ioat);
|
check_active(ioat);
|
||||||
if (active == 0 && ioat->alloc_order > ioat_get_alloc_order())
|
|
||||||
reshape_ring(ioat, ioat->alloc_order-1);
|
|
||||||
spin_unlock_bh(&ioat->prep_lock);
|
spin_unlock_bh(&ioat->prep_lock);
|
||||||
spin_unlock_bh(&chan->cleanup_lock);
|
|
||||||
|
|
||||||
/* keep shrinking until we get back to our minimum
|
|
||||||
* default size
|
|
||||||
*/
|
|
||||||
if (ioat->alloc_order > ioat_get_alloc_order())
|
|
||||||
mod_timer(&chan->timer, jiffies + IDLE_TIMEOUT);
|
|
||||||
}
|
}
|
||||||
|
spin_unlock_bh(&chan->cleanup_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
static enum dma_status
|
static enum dma_status
|
||||||
|
|
Loading…
Add table
Reference in a new issue