Commit 9a2c160a authored by Jan Glauber's avatar Jan Glauber Committed by Martin Schwidefsky

[S390] qdio: fix check for running under z/VM

The check whether qdio runs under z/VM was incorrect since SIGA-Sync is not
set if the device runs with QIOASSIST. Use MACHINE_IS_VM instead to prevent
polling under z/VM.

Merge qdio_inbound_q_done and tiqdio_is_inbound_q_done.
Signed-off-by: default avatarJan Glauber <jang@linux.vnet.ibm.com>
Signed-off-by: default avatarMartin Schwidefsky <schwidefsky@de.ibm.com>
parent 60b5df2f
...@@ -499,7 +499,7 @@ static int get_inbound_buffer_frontier(struct qdio_q *q) ...@@ -499,7 +499,7 @@ static int get_inbound_buffer_frontier(struct qdio_q *q)
/* /*
* No siga-sync needed for non-qebsm here, as the inbound queue * No siga-sync needed for non-qebsm here, as the inbound queue
* will be synced on the next siga-r, resp. * will be synced on the next siga-r, resp.
* tiqdio_is_inbound_q_done will do the siga-sync. * qdio_inbound_q_done will do the siga-sync.
*/ */
q->first_to_check = add_buf(q->first_to_check, count); q->first_to_check = add_buf(q->first_to_check, count);
atomic_sub(count, &q->nr_buf_used); atomic_sub(count, &q->nr_buf_used);
...@@ -530,35 +530,32 @@ static int qdio_inbound_q_moved(struct qdio_q *q) ...@@ -530,35 +530,32 @@ static int qdio_inbound_q_moved(struct qdio_q *q)
if ((bufnr != q->last_move) || q->qdio_error) { if ((bufnr != q->last_move) || q->qdio_error) {
q->last_move = bufnr; q->last_move = bufnr;
if (!need_siga_sync(q) && !pci_out_supported(q)) if (!is_thinint_irq(q->irq_ptr) && !MACHINE_IS_VM)
q->u.in.timestamp = get_usecs(); q->u.in.timestamp = get_usecs();
DBF_DEV_EVENT(DBF_INFO, q->irq_ptr, "in moved");
return 1; return 1;
} else } else
return 0; return 0;
} }
static int qdio_inbound_q_done(struct qdio_q *q) static inline int qdio_inbound_q_done(struct qdio_q *q)
{ {
unsigned char state = 0; unsigned char state = 0;
if (!atomic_read(&q->nr_buf_used)) if (!atomic_read(&q->nr_buf_used))
return 1; return 1;
/*
* We need that one for synchronization with the adapter, as it
* does a kind of PCI avoidance.
*/
qdio_siga_sync_q(q); qdio_siga_sync_q(q);
get_buf_state(q, q->first_to_check, &state, 0); get_buf_state(q, q->first_to_check, &state, 0);
if (state == SLSB_P_INPUT_PRIMED) if (state == SLSB_P_INPUT_PRIMED)
/* we got something to do */ /* more work coming */
return 0; return 0;
/* on VM, we don't poll, so the q is always done here */ if (is_thinint_irq(q->irq_ptr))
if (need_siga_sync(q) || pci_out_supported(q)) return 1;
/* don't poll under z/VM */
if (MACHINE_IS_VM)
return 1; return 1;
/* /*
...@@ -569,27 +566,8 @@ static int qdio_inbound_q_done(struct qdio_q *q) ...@@ -569,27 +566,8 @@ static int qdio_inbound_q_done(struct qdio_q *q)
DBF_DEV_EVENT(DBF_INFO, q->irq_ptr, "in done:%3d", DBF_DEV_EVENT(DBF_INFO, q->irq_ptr, "in done:%3d",
q->first_to_check); q->first_to_check);
return 1; return 1;
} else { } else
DBF_DEV_EVENT(DBF_INFO, q->irq_ptr, "in notd:%3d",
q->first_to_check);
return 0;
}
}
static inline int tiqdio_inbound_q_done(struct qdio_q *q)
{
unsigned char state = 0;
if (!atomic_read(&q->nr_buf_used))
return 1;
qdio_siga_sync_q(q);
get_buf_state(q, q->first_to_check, &state, 0);
if (state == SLSB_P_INPUT_PRIMED)
/* more work coming */
return 0; return 0;
return 1;
} }
static void qdio_kick_handler(struct qdio_q *q) static void qdio_kick_handler(struct qdio_q *q)
...@@ -847,7 +825,7 @@ static void __tiqdio_inbound_processing(struct qdio_q *q) ...@@ -847,7 +825,7 @@ static void __tiqdio_inbound_processing(struct qdio_q *q)
qdio_kick_handler(q); qdio_kick_handler(q);
if (!tiqdio_inbound_q_done(q)) { if (!qdio_inbound_q_done(q)) {
qdio_perf_stat_inc(&perf_stats.thinint_inbound_loop); qdio_perf_stat_inc(&perf_stats.thinint_inbound_loop);
if (likely(q->irq_ptr->state != QDIO_IRQ_STATE_STOPPED)) if (likely(q->irq_ptr->state != QDIO_IRQ_STATE_STOPPED))
tasklet_schedule(&q->tasklet); tasklet_schedule(&q->tasklet);
...@@ -858,7 +836,7 @@ static void __tiqdio_inbound_processing(struct qdio_q *q) ...@@ -858,7 +836,7 @@ static void __tiqdio_inbound_processing(struct qdio_q *q)
* We need to check again to not lose initiative after * We need to check again to not lose initiative after
* resetting the ACK state. * resetting the ACK state.
*/ */
if (!tiqdio_inbound_q_done(q)) { if (!qdio_inbound_q_done(q)) {
qdio_perf_stat_inc(&perf_stats.thinint_inbound_loop2); qdio_perf_stat_inc(&perf_stats.thinint_inbound_loop2);
if (likely(q->irq_ptr->state != QDIO_IRQ_STATE_STOPPED)) if (likely(q->irq_ptr->state != QDIO_IRQ_STATE_STOPPED))
tasklet_schedule(&q->tasklet); tasklet_schedule(&q->tasklet);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment