Commit 104ea556 authored by frank.blaschka@de.ibm.com's avatar frank.blaschka@de.ibm.com Committed by David S. Miller

qdio: support asynchronous delivery of storage blocks

This patch introduces support for asynchronous delivery of storage blocks for
Hipersockets. Upper layers may exploit this functionality to reuse SBALs for
which the delivery status is still pending.
Signed-off-by: default avatarEinar Lueck <elelueck@de.ibm.com>
Signed-off-by: default avatarJan Glauber <jang@linux.vnet.ibm.com>
Signed-off-by: default avatarFrank Blaschka <frank.blaschka@de.ibm.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent 3881ac44
...@@ -122,6 +122,40 @@ struct slibe { ...@@ -122,6 +122,40 @@ struct slibe {
u64 parms; u64 parms;
}; };
/**
* struct qaob - queue asynchronous operation block
* @res0: reserved parameters
* @res1: reserved parameter
* @res2: reserved parameter
* @res3: reserved parameter
* @aorc: asynchronous operation return code
* @flags: internal flags
* @cbtbs: control block type
* @sb_count: number of storage blocks
* @sba: storage block element addresses
* @dcount: size of storage block elements
* @user0: user defineable value
* @res4: reserved paramater
* @user1: user defineable value
* @user2: user defineable value
*/
struct qaob {
u64 res0[6];
u8 res1;
u8 res2;
u8 res3;
u8 aorc;
u8 flags;
u16 cbtbs;
u8 sb_count;
u64 sba[QDIO_MAX_ELEMENTS_PER_BUFFER];
u16 dcount[QDIO_MAX_ELEMENTS_PER_BUFFER];
u64 user0;
u64 res4[2];
u64 user1;
u64 user2;
} __attribute__ ((packed, aligned(256)));
/** /**
* struct slib - storage list information block (SLIB) * struct slib - storage list information block (SLIB)
* @nsliba: next SLIB address (if any) * @nsliba: next SLIB address (if any)
...@@ -225,6 +259,31 @@ struct slsb { ...@@ -225,6 +259,31 @@ struct slsb {
#define CHSC_AC2_DATA_DIV_AVAILABLE 0x0010 #define CHSC_AC2_DATA_DIV_AVAILABLE 0x0010
#define CHSC_AC2_DATA_DIV_ENABLED 0x0002 #define CHSC_AC2_DATA_DIV_ENABLED 0x0002
/**
* struct qdio_outbuf_state - SBAL related asynchronous operation information
* (for communication with upper layer programs)
* (only required for use with completion queues)
* @flags: flags indicating state of buffer
* @aob: pointer to QAOB used for the particular SBAL
* @user: pointer to upper layer program's state information related to SBAL
* (stored in user1 data of QAOB)
*/
struct qdio_outbuf_state {
u8 flags;
struct qaob *aob;
void *user;
};
#define QDIO_OUTBUF_STATE_FLAG_NONE 0x00
#define QDIO_OUTBUF_STATE_FLAG_PENDING 0x01
#define CHSC_AC1_INITIATE_INPUTQ 0x80
#define CHSC_AC2_DATA_DIV_AVAILABLE 0x0010
#define CHSC_AC2_DATA_DIV_ENABLED 0x0002
#define CHSC_AC3_FORMAT2_CQ_AVAILABLE 0x8000
struct qdio_ssqd_desc { struct qdio_ssqd_desc {
u8 flags; u8 flags;
u8:8; u8:8;
...@@ -243,8 +302,7 @@ struct qdio_ssqd_desc { ...@@ -243,8 +302,7 @@ struct qdio_ssqd_desc {
u64 sch_token; u64 sch_token;
u8 mro; u8 mro;
u8 mri; u8 mri;
u8:8; u16 qdioac3;
u8 sbalic;
u16:16; u16:16;
u8:8; u8:8;
u8 mmwc; u8 mmwc;
...@@ -280,9 +338,11 @@ typedef void qdio_handler_t(struct ccw_device *, unsigned int, int, ...@@ -280,9 +338,11 @@ typedef void qdio_handler_t(struct ccw_device *, unsigned int, int,
* @no_output_qs: number of output queues * @no_output_qs: number of output queues
* @input_handler: handler to be called for input queues * @input_handler: handler to be called for input queues
* @output_handler: handler to be called for output queues * @output_handler: handler to be called for output queues
* @queue_start_poll: polling handlers (one per input queue or NULL)
* @int_parm: interruption parameter * @int_parm: interruption parameter
* @input_sbal_addr_array: address of no_input_qs * 128 pointers * @input_sbal_addr_array: address of no_input_qs * 128 pointers
* @output_sbal_addr_array: address of no_output_qs * 128 pointers * @output_sbal_addr_array: address of no_output_qs * 128 pointers
* @output_sbal_state_array: no_output_qs * 128 state info (for CQ or NULL)
*/ */
struct qdio_initialize { struct qdio_initialize {
struct ccw_device *cdev; struct ccw_device *cdev;
...@@ -297,11 +357,12 @@ struct qdio_initialize { ...@@ -297,11 +357,12 @@ struct qdio_initialize {
unsigned int no_output_qs; unsigned int no_output_qs;
qdio_handler_t *input_handler; qdio_handler_t *input_handler;
qdio_handler_t *output_handler; qdio_handler_t *output_handler;
void (*queue_start_poll) (struct ccw_device *, int, unsigned long); void (**queue_start_poll) (struct ccw_device *, int, unsigned long);
int scan_threshold; int scan_threshold;
unsigned long int_parm; unsigned long int_parm;
void **input_sbal_addr_array; void **input_sbal_addr_array;
void **output_sbal_addr_array; void **output_sbal_addr_array;
struct qdio_outbuf_state *output_sbal_state_array;
}; };
#define QDIO_STATE_INACTIVE 0x00000002 /* after qdio_cleanup */ #define QDIO_STATE_INACTIVE 0x00000002 /* after qdio_cleanup */
...@@ -316,6 +377,7 @@ struct qdio_initialize { ...@@ -316,6 +377,7 @@ struct qdio_initialize {
extern int qdio_allocate(struct qdio_initialize *); extern int qdio_allocate(struct qdio_initialize *);
extern int qdio_establish(struct qdio_initialize *); extern int qdio_establish(struct qdio_initialize *);
extern int qdio_activate(struct ccw_device *); extern int qdio_activate(struct ccw_device *);
extern void qdio_release_aob(struct qaob *);
extern int do_QDIO(struct ccw_device *, unsigned int, int, unsigned int, extern int do_QDIO(struct ccw_device *, unsigned int, int, unsigned int,
unsigned int); unsigned int);
extern int qdio_start_irq(struct ccw_device *, int); extern int qdio_start_irq(struct ccw_device *, int);
......
...@@ -44,6 +44,7 @@ enum qdio_irq_states { ...@@ -44,6 +44,7 @@ enum qdio_irq_states {
#define SLSB_STATE_NOT_INIT 0x0 #define SLSB_STATE_NOT_INIT 0x0
#define SLSB_STATE_EMPTY 0x1 #define SLSB_STATE_EMPTY 0x1
#define SLSB_STATE_PRIMED 0x2 #define SLSB_STATE_PRIMED 0x2
#define SLSB_STATE_PENDING 0x3
#define SLSB_STATE_HALTED 0xe #define SLSB_STATE_HALTED 0xe
#define SLSB_STATE_ERROR 0xf #define SLSB_STATE_ERROR 0xf
#define SLSB_TYPE_INPUT 0x0 #define SLSB_TYPE_INPUT 0x0
...@@ -67,6 +68,8 @@ enum qdio_irq_states { ...@@ -67,6 +68,8 @@ enum qdio_irq_states {
(SLSB_OWNER_PROG | SLSB_TYPE_OUTPUT | SLSB_STATE_NOT_INIT) /* 0xa0 */ (SLSB_OWNER_PROG | SLSB_TYPE_OUTPUT | SLSB_STATE_NOT_INIT) /* 0xa0 */
#define SLSB_P_OUTPUT_EMPTY \ #define SLSB_P_OUTPUT_EMPTY \
(SLSB_OWNER_PROG | SLSB_TYPE_OUTPUT | SLSB_STATE_EMPTY) /* 0xa1 */ (SLSB_OWNER_PROG | SLSB_TYPE_OUTPUT | SLSB_STATE_EMPTY) /* 0xa1 */
#define SLSB_P_OUTPUT_PENDING \
(SLSB_OWNER_PROG | SLSB_TYPE_OUTPUT | SLSB_STATE_PENDING) /* 0xa3 */
#define SLSB_CU_OUTPUT_PRIMED \ #define SLSB_CU_OUTPUT_PRIMED \
(SLSB_OWNER_CU | SLSB_TYPE_OUTPUT | SLSB_STATE_PRIMED) /* 0x62 */ (SLSB_OWNER_CU | SLSB_TYPE_OUTPUT | SLSB_STATE_PRIMED) /* 0x62 */
#define SLSB_P_OUTPUT_HALTED \ #define SLSB_P_OUTPUT_HALTED \
...@@ -97,6 +100,7 @@ enum qdio_irq_states { ...@@ -97,6 +100,7 @@ enum qdio_irq_states {
#define QDIO_SIGA_WRITE 0x00 #define QDIO_SIGA_WRITE 0x00
#define QDIO_SIGA_READ 0x01 #define QDIO_SIGA_READ 0x01
#define QDIO_SIGA_SYNC 0x02 #define QDIO_SIGA_SYNC 0x02
#define QDIO_SIGA_WRITEQ 0x04
#define QDIO_SIGA_QEBSM_FLAG 0x80 #define QDIO_SIGA_QEBSM_FLAG 0x80
#ifdef CONFIG_64BIT #ifdef CONFIG_64BIT
...@@ -253,6 +257,12 @@ struct qdio_input_q { ...@@ -253,6 +257,12 @@ struct qdio_input_q {
struct qdio_output_q { struct qdio_output_q {
/* PCIs are enabled for the queue */ /* PCIs are enabled for the queue */
int pci_out_enabled; int pci_out_enabled;
/* cq: use asynchronous output buffers */
int use_cq;
/* cq: aobs used for particual SBAL */
struct qaob **aobs;
/* cq: sbal state related to asynchronous operation */
struct qdio_outbuf_state *sbal_state;
/* timer to check for more outbound work */ /* timer to check for more outbound work */
struct timer_list timer; struct timer_list timer;
/* used SBALs before tasklet schedule */ /* used SBALs before tasklet schedule */
...@@ -432,9 +442,20 @@ struct indicator_t { ...@@ -432,9 +442,20 @@ struct indicator_t {
extern struct indicator_t *q_indicators; extern struct indicator_t *q_indicators;
static inline int shared_ind(u32 *dsci) static inline int has_multiple_inq_on_dsci(struct qdio_irq *irq)
{ {
return dsci == &q_indicators[TIQDIO_SHARED_IND].ind; return irq->nr_input_qs > 1;
}
static inline int references_shared_dsci(struct qdio_irq *irq)
{
return irq->dsci == &q_indicators[TIQDIO_SHARED_IND].ind;
}
static inline int shared_ind(struct qdio_q *q)
{
struct qdio_irq *i = q->irq_ptr;
return references_shared_dsci(i) || has_multiple_inq_on_dsci(i);
} }
/* prototypes for thin interrupt */ /* prototypes for thin interrupt */
...@@ -449,6 +470,7 @@ void tiqdio_free_memory(void); ...@@ -449,6 +470,7 @@ void tiqdio_free_memory(void);
int tiqdio_register_thinints(void); int tiqdio_register_thinints(void);
void tiqdio_unregister_thinints(void); void tiqdio_unregister_thinints(void);
/* prototypes for setup */ /* prototypes for setup */
void qdio_inbound_processing(unsigned long data); void qdio_inbound_processing(unsigned long data);
void qdio_outbound_processing(unsigned long data); void qdio_outbound_processing(unsigned long data);
...@@ -469,6 +491,9 @@ int qdio_setup_create_sysfs(struct ccw_device *cdev); ...@@ -469,6 +491,9 @@ int qdio_setup_create_sysfs(struct ccw_device *cdev);
void qdio_setup_destroy_sysfs(struct ccw_device *cdev); void qdio_setup_destroy_sysfs(struct ccw_device *cdev);
int qdio_setup_init(void); int qdio_setup_init(void);
void qdio_setup_exit(void); void qdio_setup_exit(void);
int qdio_enable_async_operation(struct qdio_output_q *q);
void qdio_disable_async_operation(struct qdio_output_q *q);
struct qaob *qdio_allocate_aob(void);
int debug_get_buf_state(struct qdio_q *q, unsigned int bufnr, int debug_get_buf_state(struct qdio_q *q, unsigned int bufnr,
unsigned char *state); unsigned char *state);
......
...@@ -76,6 +76,9 @@ static int qstat_show(struct seq_file *m, void *v) ...@@ -76,6 +76,9 @@ static int qstat_show(struct seq_file *m, void *v)
case SLSB_P_OUTPUT_NOT_INIT: case SLSB_P_OUTPUT_NOT_INIT:
seq_printf(m, "N"); seq_printf(m, "N");
break; break;
case SLSB_P_OUTPUT_PENDING:
seq_printf(m, "P");
break;
case SLSB_P_INPUT_PRIMED: case SLSB_P_INPUT_PRIMED:
case SLSB_CU_OUTPUT_PRIMED: case SLSB_CU_OUTPUT_PRIMED:
seq_printf(m, "+"); seq_printf(m, "+");
......
...@@ -14,6 +14,7 @@ ...@@ -14,6 +14,7 @@
#include <linux/timer.h> #include <linux/timer.h>
#include <linux/delay.h> #include <linux/delay.h>
#include <linux/gfp.h> #include <linux/gfp.h>
#include <linux/io.h>
#include <linux/kernel_stat.h> #include <linux/kernel_stat.h>
#include <linux/atomic.h> #include <linux/atomic.h>
#include <asm/debug.h> #include <asm/debug.h>
...@@ -77,11 +78,13 @@ static inline int do_siga_input(unsigned long schid, unsigned int mask, ...@@ -77,11 +78,13 @@ static inline int do_siga_input(unsigned long schid, unsigned int mask,
* Note: For IQDC unicast queues only the highest priority queue is processed. * Note: For IQDC unicast queues only the highest priority queue is processed.
*/ */
static inline int do_siga_output(unsigned long schid, unsigned long mask, static inline int do_siga_output(unsigned long schid, unsigned long mask,
unsigned int *bb, unsigned int fc) unsigned int *bb, unsigned int fc,
unsigned long aob)
{ {
register unsigned long __fc asm("0") = fc; register unsigned long __fc asm("0") = fc;
register unsigned long __schid asm("1") = schid; register unsigned long __schid asm("1") = schid;
register unsigned long __mask asm("2") = mask; register unsigned long __mask asm("2") = mask;
register unsigned long __aob asm("3") = aob;
int cc = QDIO_ERROR_SIGA_ACCESS_EXCEPTION; int cc = QDIO_ERROR_SIGA_ACCESS_EXCEPTION;
asm volatile( asm volatile(
...@@ -90,7 +93,8 @@ static inline int do_siga_output(unsigned long schid, unsigned long mask, ...@@ -90,7 +93,8 @@ static inline int do_siga_output(unsigned long schid, unsigned long mask,
" srl %0,28\n" " srl %0,28\n"
"1:\n" "1:\n"
EX_TABLE(0b, 1b) EX_TABLE(0b, 1b)
: "+d" (cc), "+d" (__fc), "+d" (__schid), "+d" (__mask) : "+d" (cc), "+d" (__fc), "+d" (__schid), "+d" (__mask),
"+d" (__aob)
: : "cc", "memory"); : : "cc", "memory");
*bb = ((unsigned int) __fc) >> 31; *bb = ((unsigned int) __fc) >> 31;
return cc; return cc;
...@@ -212,7 +216,7 @@ static int qdio_do_sqbs(struct qdio_q *q, unsigned char state, int start, ...@@ -212,7 +216,7 @@ static int qdio_do_sqbs(struct qdio_q *q, unsigned char state, int start,
/* returns number of examined buffers and their common state in *state */ /* returns number of examined buffers and their common state in *state */
static inline int get_buf_states(struct qdio_q *q, unsigned int bufnr, static inline int get_buf_states(struct qdio_q *q, unsigned int bufnr,
unsigned char *state, unsigned int count, unsigned char *state, unsigned int count,
int auto_ack) int auto_ack, int merge_pending)
{ {
unsigned char __state = 0; unsigned char __state = 0;
int i; int i;
...@@ -224,9 +228,14 @@ static inline int get_buf_states(struct qdio_q *q, unsigned int bufnr, ...@@ -224,9 +228,14 @@ static inline int get_buf_states(struct qdio_q *q, unsigned int bufnr,
return qdio_do_eqbs(q, state, bufnr, count, auto_ack); return qdio_do_eqbs(q, state, bufnr, count, auto_ack);
for (i = 0; i < count; i++) { for (i = 0; i < count; i++) {
if (!__state) if (!__state) {
__state = q->slsb.val[bufnr]; __state = q->slsb.val[bufnr];
else if (q->slsb.val[bufnr] != __state) if (merge_pending && __state == SLSB_P_OUTPUT_PENDING)
__state = SLSB_P_OUTPUT_EMPTY;
} else if (merge_pending) {
if ((q->slsb.val[bufnr] & __state) != __state)
break;
} else if (q->slsb.val[bufnr] != __state)
break; break;
bufnr = next_buf(bufnr); bufnr = next_buf(bufnr);
} }
...@@ -237,7 +246,7 @@ static inline int get_buf_states(struct qdio_q *q, unsigned int bufnr, ...@@ -237,7 +246,7 @@ static inline int get_buf_states(struct qdio_q *q, unsigned int bufnr,
static inline int get_buf_state(struct qdio_q *q, unsigned int bufnr, static inline int get_buf_state(struct qdio_q *q, unsigned int bufnr,
unsigned char *state, int auto_ack) unsigned char *state, int auto_ack)
{ {
return get_buf_states(q, bufnr, state, 1, auto_ack); return get_buf_states(q, bufnr, state, 1, auto_ack, 0);
} }
/* wrap-around safe setting of slsb states, returns number of changed buffers */ /* wrap-around safe setting of slsb states, returns number of changed buffers */
...@@ -308,19 +317,28 @@ static inline int qdio_siga_sync_q(struct qdio_q *q) ...@@ -308,19 +317,28 @@ static inline int qdio_siga_sync_q(struct qdio_q *q)
return qdio_siga_sync(q, q->mask, 0); return qdio_siga_sync(q, q->mask, 0);
} }
static int qdio_siga_output(struct qdio_q *q, unsigned int *busy_bit) static int qdio_siga_output(struct qdio_q *q, unsigned int *busy_bit,
unsigned long aob)
{ {
unsigned long schid = *((u32 *) &q->irq_ptr->schid); unsigned long schid = *((u32 *) &q->irq_ptr->schid);
unsigned int fc = QDIO_SIGA_WRITE; unsigned int fc = QDIO_SIGA_WRITE;
u64 start_time = 0; u64 start_time = 0;
int retries = 0, cc; int retries = 0, cc;
unsigned long laob = 0;
if (q->u.out.use_cq && aob != 0) {
fc = QDIO_SIGA_WRITEQ;
laob = aob;
}
if (is_qebsm(q)) { if (is_qebsm(q)) {
schid = q->irq_ptr->sch_token; schid = q->irq_ptr->sch_token;
fc |= QDIO_SIGA_QEBSM_FLAG; fc |= QDIO_SIGA_QEBSM_FLAG;
} }
again: again:
cc = do_siga_output(schid, q->mask, busy_bit, fc); WARN_ON_ONCE((aob && queue_type(q) != QDIO_IQDIO_QFMT) ||
(aob && fc != QDIO_SIGA_WRITEQ));
cc = do_siga_output(schid, q->mask, busy_bit, fc, laob);
/* hipersocket busy condition */ /* hipersocket busy condition */
if (unlikely(*busy_bit)) { if (unlikely(*busy_bit)) {
...@@ -379,7 +397,7 @@ int debug_get_buf_state(struct qdio_q *q, unsigned int bufnr, ...@@ -379,7 +397,7 @@ int debug_get_buf_state(struct qdio_q *q, unsigned int bufnr,
{ {
if (need_siga_sync(q)) if (need_siga_sync(q))
qdio_siga_sync_q(q); qdio_siga_sync_q(q);
return get_buf_states(q, bufnr, state, 1, 0); return get_buf_states(q, bufnr, state, 1, 0, 0);
} }
static inline void qdio_stop_polling(struct qdio_q *q) static inline void qdio_stop_polling(struct qdio_q *q)
...@@ -507,7 +525,7 @@ static int get_inbound_buffer_frontier(struct qdio_q *q) ...@@ -507,7 +525,7 @@ static int get_inbound_buffer_frontier(struct qdio_q *q)
* No siga sync here, as a PCI or we after a thin interrupt * No siga sync here, as a PCI or we after a thin interrupt
* already sync'ed the queues. * already sync'ed the queues.
*/ */
count = get_buf_states(q, q->first_to_check, &state, count, 1); count = get_buf_states(q, q->first_to_check, &state, count, 1, 0);
if (!count) if (!count)
goto out; goto out;
...@@ -590,6 +608,107 @@ static inline int qdio_inbound_q_done(struct qdio_q *q) ...@@ -590,6 +608,107 @@ static inline int qdio_inbound_q_done(struct qdio_q *q)
return 0; return 0;
} }
static inline int contains_aobs(struct qdio_q *q)
{
return !q->is_input_q && q->u.out.use_cq;
}
static inline void qdio_trace_aob(struct qdio_irq *irq, struct qdio_q *q,
int i, struct qaob *aob)
{
int tmp;
DBF_DEV_EVENT(DBF_INFO, irq, "AOB%d:%lx", i,
(unsigned long) virt_to_phys(aob));
DBF_DEV_EVENT(DBF_INFO, irq, "RES00:%lx",
(unsigned long) aob->res0[0]);
DBF_DEV_EVENT(DBF_INFO, irq, "RES01:%lx",
(unsigned long) aob->res0[1]);
DBF_DEV_EVENT(DBF_INFO, irq, "RES02:%lx",
(unsigned long) aob->res0[2]);
DBF_DEV_EVENT(DBF_INFO, irq, "RES03:%lx",
(unsigned long) aob->res0[3]);
DBF_DEV_EVENT(DBF_INFO, irq, "RES04:%lx",
(unsigned long) aob->res0[4]);
DBF_DEV_EVENT(DBF_INFO, irq, "RES05:%lx",
(unsigned long) aob->res0[5]);
DBF_DEV_EVENT(DBF_INFO, irq, "RES1:%x", aob->res1);
DBF_DEV_EVENT(DBF_INFO, irq, "RES2:%x", aob->res2);
DBF_DEV_EVENT(DBF_INFO, irq, "RES3:%x", aob->res3);
DBF_DEV_EVENT(DBF_INFO, irq, "AORC:%u", aob->aorc);
DBF_DEV_EVENT(DBF_INFO, irq, "FLAGS:%u", aob->flags);
DBF_DEV_EVENT(DBF_INFO, irq, "CBTBS:%u", aob->cbtbs);
DBF_DEV_EVENT(DBF_INFO, irq, "SBC:%u", aob->sb_count);
for (tmp = 0; tmp < QDIO_MAX_ELEMENTS_PER_BUFFER; ++tmp) {
DBF_DEV_EVENT(DBF_INFO, irq, "SBA%d:%lx", tmp,
(unsigned long) aob->sba[tmp]);
DBF_DEV_EVENT(DBF_INFO, irq, "rSBA%d:%lx", tmp,
(unsigned long) q->sbal[i]->element[tmp].addr);
DBF_DEV_EVENT(DBF_INFO, irq, "DC%d:%u", tmp, aob->dcount[tmp]);
DBF_DEV_EVENT(DBF_INFO, irq, "rDC%d:%u", tmp,
q->sbal[i]->element[tmp].length);
}
DBF_DEV_EVENT(DBF_INFO, irq, "USER0:%lx", (unsigned long) aob->user0);
for (tmp = 0; tmp < 2; ++tmp) {
DBF_DEV_EVENT(DBF_INFO, irq, "RES4%d:%lx", tmp,
(unsigned long) aob->res4[tmp]);
}
DBF_DEV_EVENT(DBF_INFO, irq, "USER1:%lx", (unsigned long) aob->user1);
DBF_DEV_EVENT(DBF_INFO, irq, "USER2:%lx", (unsigned long) aob->user2);
}
static inline void qdio_handle_aobs(struct qdio_q *q, int start, int count)
{
unsigned char state = 0;
int j, b = start;
if (!contains_aobs(q))
return;
for (j = 0; j < count; ++j) {
get_buf_state(q, b, &state, 0);
if (state == SLSB_P_OUTPUT_PENDING) {
struct qaob *aob = q->u.out.aobs[b];
if (aob == NULL)
continue;
BUG_ON(q->u.out.sbal_state == NULL);
q->u.out.sbal_state[b].flags |=
QDIO_OUTBUF_STATE_FLAG_PENDING;
q->u.out.aobs[b] = NULL;
} else if (state == SLSB_P_OUTPUT_EMPTY) {
BUG_ON(q->u.out.sbal_state == NULL);
q->u.out.sbal_state[b].aob = NULL;
}
b = next_buf(b);
}
}
static inline unsigned long qdio_aob_for_buffer(struct qdio_output_q *q,
int bufnr)
{
unsigned long phys_aob = 0;
if (!q->use_cq)
goto out;
if (!q->aobs[bufnr]) {
struct qaob *aob = qdio_allocate_aob();
q->aobs[bufnr] = aob;
}
if (q->aobs[bufnr]) {
BUG_ON(q->sbal_state == NULL);
q->sbal_state[bufnr].flags = QDIO_OUTBUF_STATE_FLAG_NONE;
q->sbal_state[bufnr].aob = q->aobs[bufnr];
q->aobs[bufnr]->user1 = (u64) q->sbal_state[bufnr].user;
phys_aob = virt_to_phys(q->aobs[bufnr]);
BUG_ON(phys_aob & 0xFF);
}
out:
return phys_aob;
}
static void qdio_kick_handler(struct qdio_q *q) static void qdio_kick_handler(struct qdio_q *q)
{ {
int start = q->first_to_kick; int start = q->first_to_kick;
...@@ -610,6 +729,8 @@ static void qdio_kick_handler(struct qdio_q *q) ...@@ -610,6 +729,8 @@ static void qdio_kick_handler(struct qdio_q *q)
start, count); start, count);
} }
qdio_handle_aobs(q, start, count);
q->handler(q->irq_ptr->cdev, q->qdio_error, q->nr, start, count, q->handler(q->irq_ptr->cdev, q->qdio_error, q->nr, start, count,
q->irq_ptr->int_parm); q->irq_ptr->int_parm);
...@@ -672,23 +793,26 @@ static int get_outbound_buffer_frontier(struct qdio_q *q) ...@@ -672,23 +793,26 @@ static int get_outbound_buffer_frontier(struct qdio_q *q)
*/ */
count = min(atomic_read(&q->nr_buf_used), QDIO_MAX_BUFFERS_MASK); count = min(atomic_read(&q->nr_buf_used), QDIO_MAX_BUFFERS_MASK);
stop = add_buf(q->first_to_check, count); stop = add_buf(q->first_to_check, count);
if (q->first_to_check == stop) if (q->first_to_check == stop)
return q->first_to_check; goto out;
count = get_buf_states(q, q->first_to_check, &state, count, 0); count = get_buf_states(q, q->first_to_check, &state, count, 0, 1);
if (!count) if (!count)
return q->first_to_check; goto out;
switch (state) { switch (state) {
case SLSB_P_OUTPUT_PENDING:
BUG();
case SLSB_P_OUTPUT_EMPTY: case SLSB_P_OUTPUT_EMPTY:
/* the adapter got it */ /* the adapter got it */
DBF_DEV_EVENT(DBF_INFO, q->irq_ptr, "out empty:%1d %02x", q->nr, count); DBF_DEV_EVENT(DBF_INFO, q->irq_ptr,
"out empty:%1d %02x", q->nr, count);
atomic_sub(count, &q->nr_buf_used); atomic_sub(count, &q->nr_buf_used);
q->first_to_check = add_buf(q->first_to_check, count); q->first_to_check = add_buf(q->first_to_check, count);
if (q->irq_ptr->perf_stat_enabled) if (q->irq_ptr->perf_stat_enabled)
account_sbals(q, count); account_sbals(q, count);
break; break;
case SLSB_P_OUTPUT_ERROR: case SLSB_P_OUTPUT_ERROR:
process_buffer_error(q, count); process_buffer_error(q, count);
...@@ -701,7 +825,8 @@ static int get_outbound_buffer_frontier(struct qdio_q *q) ...@@ -701,7 +825,8 @@ static int get_outbound_buffer_frontier(struct qdio_q *q)
/* the adapter has not fetched the output yet */ /* the adapter has not fetched the output yet */
if (q->irq_ptr->perf_stat_enabled) if (q->irq_ptr->perf_stat_enabled)
q->q_stats.nr_sbal_nop++; q->q_stats.nr_sbal_nop++;
DBF_DEV_EVENT(DBF_INFO, q->irq_ptr, "out primed:%1d", q->nr); DBF_DEV_EVENT(DBF_INFO, q->irq_ptr, "out primed:%1d",
q->nr);
break; break;
case SLSB_P_OUTPUT_NOT_INIT: case SLSB_P_OUTPUT_NOT_INIT:
case SLSB_P_OUTPUT_HALTED: case SLSB_P_OUTPUT_HALTED:
...@@ -709,6 +834,8 @@ static int get_outbound_buffer_frontier(struct qdio_q *q) ...@@ -709,6 +834,8 @@ static int get_outbound_buffer_frontier(struct qdio_q *q)
default: default:
BUG(); BUG();
} }
out:
return q->first_to_check; return q->first_to_check;
} }
...@@ -732,7 +859,7 @@ static inline int qdio_outbound_q_moved(struct qdio_q *q) ...@@ -732,7 +859,7 @@ static inline int qdio_outbound_q_moved(struct qdio_q *q)
return 0; return 0;
} }
static int qdio_kick_outbound_q(struct qdio_q *q) static int qdio_kick_outbound_q(struct qdio_q *q, unsigned long aob)
{ {
int retries = 0, cc; int retries = 0, cc;
unsigned int busy_bit; unsigned int busy_bit;
...@@ -744,7 +871,7 @@ static int qdio_kick_outbound_q(struct qdio_q *q) ...@@ -744,7 +871,7 @@ static int qdio_kick_outbound_q(struct qdio_q *q)
retry: retry:
qperf_inc(q, siga_write); qperf_inc(q, siga_write);
cc = qdio_siga_output(q, &busy_bit); cc = qdio_siga_output(q, &busy_bit, aob);
switch (cc) { switch (cc) {
case 0: case 0:
break; break;
...@@ -921,9 +1048,10 @@ static void qdio_int_handler_pci(struct qdio_irq *irq_ptr) ...@@ -921,9 +1048,10 @@ static void qdio_int_handler_pci(struct qdio_irq *irq_ptr)
} }
q->u.in.queue_start_poll(q->irq_ptr->cdev, q->nr, q->u.in.queue_start_poll(q->irq_ptr->cdev, q->nr,
q->irq_ptr->int_parm); q->irq_ptr->int_parm);
} else } else {
tasklet_schedule(&q->tasklet); tasklet_schedule(&q->tasklet);
} }
}
if (!pci_out_supported(q)) if (!pci_out_supported(q))
return; return;
...@@ -1236,6 +1364,26 @@ int qdio_allocate(struct qdio_initialize *init_data) ...@@ -1236,6 +1364,26 @@ int qdio_allocate(struct qdio_initialize *init_data)
} }
EXPORT_SYMBOL_GPL(qdio_allocate); EXPORT_SYMBOL_GPL(qdio_allocate);
static void qdio_detect_hsicq(struct qdio_irq *irq_ptr)
{
struct qdio_q *q = irq_ptr->input_qs[0];
int i, use_cq = 0;
if (irq_ptr->nr_input_qs > 1 && queue_type(q) == QDIO_IQDIO_QFMT)
use_cq = 1;
for_each_output_queue(irq_ptr, q, i) {
if (use_cq) {
if (qdio_enable_async_operation(&q->u.out) < 0) {
use_cq = 0;
continue;
}
} else
qdio_disable_async_operation(&q->u.out);
}
DBF_EVENT("use_cq:%d", use_cq);
}
/** /**
* qdio_establish - establish queues on a qdio subchannel * qdio_establish - establish queues on a qdio subchannel
* @init_data: initialization data * @init_data: initialization data
...@@ -1301,6 +1449,8 @@ int qdio_establish(struct qdio_initialize *init_data) ...@@ -1301,6 +1449,8 @@ int qdio_establish(struct qdio_initialize *init_data)
qdio_setup_ssqd_info(irq_ptr); qdio_setup_ssqd_info(irq_ptr);
DBF_EVENT("qib ac:%4x", irq_ptr->qib.ac); DBF_EVENT("qib ac:%4x", irq_ptr->qib.ac);
qdio_detect_hsicq(irq_ptr);
/* qebsm is now setup if available, initialize buffer states */ /* qebsm is now setup if available, initialize buffer states */
qdio_init_buf_states(irq_ptr); qdio_init_buf_states(irq_ptr);
...@@ -1480,17 +1630,21 @@ static int handle_outbound(struct qdio_q *q, unsigned int callflags, ...@@ -1480,17 +1630,21 @@ static int handle_outbound(struct qdio_q *q, unsigned int callflags,
q->u.out.pci_out_enabled = 0; q->u.out.pci_out_enabled = 0;
if (queue_type(q) == QDIO_IQDIO_QFMT) { if (queue_type(q) == QDIO_IQDIO_QFMT) {
/* One SIGA-W per buffer required for unicast HiperSockets. */ unsigned long phys_aob = 0;
/* One SIGA-W per buffer required for unicast HSI */
WARN_ON_ONCE(count > 1 && !multicast_outbound(q)); WARN_ON_ONCE(count > 1 && !multicast_outbound(q));
rc = qdio_kick_outbound_q(q); phys_aob = qdio_aob_for_buffer(&q->u.out, bufnr);
rc = qdio_kick_outbound_q(q, phys_aob);
} else if (need_siga_sync(q)) { } else if (need_siga_sync(q)) {
rc = qdio_siga_sync_q(q); rc = qdio_siga_sync_q(q);
} else { } else {
/* try to fast requeue buffers */ /* try to fast requeue buffers */
get_buf_state(q, prev_buf(bufnr), &state, 0); get_buf_state(q, prev_buf(bufnr), &state, 0);
if (state != SLSB_CU_OUTPUT_PRIMED) if (state != SLSB_CU_OUTPUT_PRIMED)
rc = qdio_kick_outbound_q(q); rc = qdio_kick_outbound_q(q, 0);
else else
qperf_inc(q, fast_requeue); qperf_inc(q, fast_requeue);
} }
...@@ -1518,6 +1672,7 @@ int do_QDIO(struct ccw_device *cdev, unsigned int callflags, ...@@ -1518,6 +1672,7 @@ int do_QDIO(struct ccw_device *cdev, unsigned int callflags,
{ {
struct qdio_irq *irq_ptr; struct qdio_irq *irq_ptr;
if (bufnr >= QDIO_MAX_BUFFERS_PER_Q || count > QDIO_MAX_BUFFERS_PER_Q) if (bufnr >= QDIO_MAX_BUFFERS_PER_Q || count > QDIO_MAX_BUFFERS_PER_Q)
return -EINVAL; return -EINVAL;
...@@ -1562,7 +1717,7 @@ int qdio_start_irq(struct ccw_device *cdev, int nr) ...@@ -1562,7 +1717,7 @@ int qdio_start_irq(struct ccw_device *cdev, int nr)
WARN_ON(queue_irqs_enabled(q)); WARN_ON(queue_irqs_enabled(q));
if (!shared_ind(q->irq_ptr->dsci)) if (!shared_ind(q))
xchg(q->irq_ptr->dsci, 0); xchg(q->irq_ptr->dsci, 0);
qdio_stop_polling(q); qdio_stop_polling(q);
...@@ -1572,7 +1727,7 @@ int qdio_start_irq(struct ccw_device *cdev, int nr) ...@@ -1572,7 +1727,7 @@ int qdio_start_irq(struct ccw_device *cdev, int nr)
* We need to check again to not lose initiative after * We need to check again to not lose initiative after
* resetting the ACK state. * resetting the ACK state.
*/ */
if (!shared_ind(q->irq_ptr->dsci) && *q->irq_ptr->dsci) if (!shared_ind(q) && *q->irq_ptr->dsci)
goto rescan; goto rescan;
if (!qdio_inbound_q_done(q)) if (!qdio_inbound_q_done(q))
goto rescan; goto rescan;
......
...@@ -19,6 +19,22 @@ ...@@ -19,6 +19,22 @@
#include "qdio_debug.h" #include "qdio_debug.h"
static struct kmem_cache *qdio_q_cache; static struct kmem_cache *qdio_q_cache;
static struct kmem_cache *qdio_aob_cache;
struct qaob *qdio_allocate_aob()
{
struct qaob *aob;
aob = kmem_cache_zalloc(qdio_aob_cache, GFP_ATOMIC);
return aob;
}
EXPORT_SYMBOL_GPL(qdio_allocate_aob);
void qdio_release_aob(struct qaob *aob)
{
kmem_cache_free(qdio_aob_cache, aob);
}
EXPORT_SYMBOL_GPL(qdio_release_aob);
/* /*
* qebsm is only available under 64bit but the adapter sets the feature * qebsm is only available under 64bit but the adapter sets the feature
...@@ -154,29 +170,36 @@ static void setup_queues(struct qdio_irq *irq_ptr, ...@@ -154,29 +170,36 @@ static void setup_queues(struct qdio_irq *irq_ptr,
struct qdio_q *q; struct qdio_q *q;
void **input_sbal_array = qdio_init->input_sbal_addr_array; void **input_sbal_array = qdio_init->input_sbal_addr_array;
void **output_sbal_array = qdio_init->output_sbal_addr_array; void **output_sbal_array = qdio_init->output_sbal_addr_array;
struct qdio_outbuf_state *output_sbal_state_array =
qdio_init->output_sbal_state_array;
int i; int i;
for_each_input_queue(irq_ptr, q, i) { for_each_input_queue(irq_ptr, q, i) {
DBF_EVENT("in-q:%1d", i); DBF_EVENT("inq:%1d", i);
setup_queues_misc(q, irq_ptr, qdio_init->input_handler, i); setup_queues_misc(q, irq_ptr, qdio_init->input_handler, i);
q->is_input_q = 1; q->is_input_q = 1;
q->u.in.queue_start_poll = qdio_init->queue_start_poll; q->u.in.queue_start_poll = qdio_init->queue_start_poll[i];
setup_storage_lists(q, irq_ptr, input_sbal_array, i); setup_storage_lists(q, irq_ptr, input_sbal_array, i);
input_sbal_array += QDIO_MAX_BUFFERS_PER_Q; input_sbal_array += QDIO_MAX_BUFFERS_PER_Q;
if (is_thinint_irq(irq_ptr)) if (is_thinint_irq(irq_ptr)) {
tasklet_init(&q->tasklet, tiqdio_inbound_processing, tasklet_init(&q->tasklet, tiqdio_inbound_processing,
(unsigned long) q); (unsigned long) q);
else } else {
tasklet_init(&q->tasklet, qdio_inbound_processing, tasklet_init(&q->tasklet, qdio_inbound_processing,
(unsigned long) q); (unsigned long) q);
} }
}
for_each_output_queue(irq_ptr, q, i) { for_each_output_queue(irq_ptr, q, i) {
DBF_EVENT("outq:%1d", i); DBF_EVENT("outq:%1d", i);
setup_queues_misc(q, irq_ptr, qdio_init->output_handler, i); setup_queues_misc(q, irq_ptr, qdio_init->output_handler, i);
q->u.out.sbal_state = output_sbal_state_array;
output_sbal_state_array += QDIO_MAX_BUFFERS_PER_Q;
q->is_input_q = 0; q->is_input_q = 0;
q->u.out.scan_threshold = qdio_init->scan_threshold; q->u.out.scan_threshold = qdio_init->scan_threshold;
setup_storage_lists(q, irq_ptr, output_sbal_array, i); setup_storage_lists(q, irq_ptr, output_sbal_array, i);
...@@ -311,6 +334,19 @@ void qdio_release_memory(struct qdio_irq *irq_ptr) ...@@ -311,6 +334,19 @@ void qdio_release_memory(struct qdio_irq *irq_ptr)
for (i = 0; i < QDIO_MAX_QUEUES_PER_IRQ; i++) { for (i = 0; i < QDIO_MAX_QUEUES_PER_IRQ; i++) {
q = irq_ptr->output_qs[i]; q = irq_ptr->output_qs[i];
if (q) { if (q) {
if (q->u.out.use_cq) {
int n;
for (n = 0; n < QDIO_MAX_BUFFERS_PER_Q; ++n) {
struct qaob *aob = q->u.out.aobs[n];
if (aob) {
qdio_release_aob(aob);
q->u.out.aobs[n] = NULL;
}
}
qdio_disable_async_operation(&q->u.out);
}
free_page((unsigned long) q->slib); free_page((unsigned long) q->slib);
kmem_cache_free(qdio_q_cache, q); kmem_cache_free(qdio_q_cache, q);
} }
...@@ -465,23 +501,60 @@ void qdio_print_subchannel_info(struct qdio_irq *irq_ptr, ...@@ -465,23 +501,60 @@ void qdio_print_subchannel_info(struct qdio_irq *irq_ptr,
printk(KERN_INFO "%s", s); printk(KERN_INFO "%s", s);
} }
int qdio_enable_async_operation(struct qdio_output_q *outq)
{
outq->aobs = kzalloc(sizeof(struct qaob *) * QDIO_MAX_BUFFERS_PER_Q,
GFP_ATOMIC);
if (!outq->aobs) {
outq->use_cq = 0;
return -ENOMEM;
}
outq->use_cq = 1;
return 0;
}
void qdio_disable_async_operation(struct qdio_output_q *q)
{
kfree(q->aobs);
q->aobs = NULL;
q->use_cq = 0;
}
int __init qdio_setup_init(void) int __init qdio_setup_init(void)
{ {
int rc;
qdio_q_cache = kmem_cache_create("qdio_q", sizeof(struct qdio_q), qdio_q_cache = kmem_cache_create("qdio_q", sizeof(struct qdio_q),
256, 0, NULL); 256, 0, NULL);
if (!qdio_q_cache) if (!qdio_q_cache)
return -ENOMEM; return -ENOMEM;
qdio_aob_cache = kmem_cache_create("qdio_aob",
sizeof(struct qaob),
sizeof(struct qaob),
0,
NULL);
if (!qdio_aob_cache) {
rc = -ENOMEM;
goto free_qdio_q_cache;
}
/* Check for OSA/FCP thin interrupts (bit 67). */ /* Check for OSA/FCP thin interrupts (bit 67). */
DBF_EVENT("thinint:%1d", DBF_EVENT("thinint:%1d",
(css_general_characteristics.aif_osa) ? 1 : 0); (css_general_characteristics.aif_osa) ? 1 : 0);
/* Check for QEBSM support in general (bit 58). */ /* Check for QEBSM support in general (bit 58). */
DBF_EVENT("cssQEBSM:%1d", (qebsm_possible()) ? 1 : 0); DBF_EVENT("cssQEBSM:%1d", (qebsm_possible()) ? 1 : 0);
return 0; rc = 0;
out:
return rc;
free_qdio_q_cache:
kmem_cache_destroy(qdio_q_cache);
goto out;
} }
void qdio_setup_exit(void) void qdio_setup_exit(void)
{ {
kmem_cache_destroy(qdio_aob_cache);
kmem_cache_destroy(qdio_q_cache); kmem_cache_destroy(qdio_q_cache);
} }
...@@ -67,12 +67,9 @@ static void put_indicator(u32 *addr) ...@@ -67,12 +67,9 @@ static void put_indicator(u32 *addr)
void tiqdio_add_input_queues(struct qdio_irq *irq_ptr) void tiqdio_add_input_queues(struct qdio_irq *irq_ptr)
{ {
struct qdio_q *q;
int i;
mutex_lock(&tiq_list_lock); mutex_lock(&tiq_list_lock);
for_each_input_queue(irq_ptr, q, i) BUG_ON(irq_ptr->nr_input_qs < 1);
list_add_rcu(&q->entry, &tiq_list); list_add_rcu(&irq_ptr->input_qs[0]->entry, &tiq_list);
mutex_unlock(&tiq_list_lock); mutex_unlock(&tiq_list_lock);
xchg(irq_ptr->dsci, 1 << 7); xchg(irq_ptr->dsci, 1 << 7);
} }
...@@ -80,19 +77,17 @@ void tiqdio_add_input_queues(struct qdio_irq *irq_ptr) ...@@ -80,19 +77,17 @@ void tiqdio_add_input_queues(struct qdio_irq *irq_ptr)
void tiqdio_remove_input_queues(struct qdio_irq *irq_ptr) void tiqdio_remove_input_queues(struct qdio_irq *irq_ptr)
{ {
struct qdio_q *q; struct qdio_q *q;
int i;
for (i = 0; i < irq_ptr->nr_input_qs; i++) { BUG_ON(irq_ptr->nr_input_qs < 1);
q = irq_ptr->input_qs[i]; q = irq_ptr->input_qs[0];
/* if establish triggered an error */ /* if establish triggered an error */
if (!q || !q->entry.prev || !q->entry.next) if (!q || !q->entry.prev || !q->entry.next)
continue; return;
mutex_lock(&tiq_list_lock); mutex_lock(&tiq_list_lock);
list_del_rcu(&q->entry); list_del_rcu(&q->entry);
mutex_unlock(&tiq_list_lock); mutex_unlock(&tiq_list_lock);
synchronize_rcu(); synchronize_rcu();
}
} }
static inline u32 clear_shared_ind(void) static inline u32 clear_shared_ind(void)
...@@ -102,6 +97,40 @@ static inline u32 clear_shared_ind(void) ...@@ -102,6 +97,40 @@ static inline u32 clear_shared_ind(void)
return xchg(&q_indicators[TIQDIO_SHARED_IND].ind, 0); return xchg(&q_indicators[TIQDIO_SHARED_IND].ind, 0);
} }
static inline void tiqdio_call_inq_handlers(struct qdio_irq *irq)
{
struct qdio_q *q;
int i;
for_each_input_queue(irq, q, i) {
if (!references_shared_dsci(irq) &&
has_multiple_inq_on_dsci(irq))
xchg(q->irq_ptr->dsci, 0);
if (q->u.in.queue_start_poll) {
/* skip if polling is enabled or already in work */
if (test_and_set_bit(QDIO_QUEUE_IRQS_DISABLED,
&q->u.in.queue_irq_state)) {
qperf_inc(q, int_discarded);
continue;
}
/* avoid dsci clear here, done after processing */
q->u.in.queue_start_poll(q->irq_ptr->cdev, q->nr,
q->irq_ptr->int_parm);
} else {
if (!shared_ind(q))
xchg(q->irq_ptr->dsci, 0);
/*
* Call inbound processing but not directly
* since that could starve other thinint queues.
*/
tasklet_schedule(&q->tasklet);
}
}
}
/** /**
* tiqdio_thinint_handler - thin interrupt handler for qdio * tiqdio_thinint_handler - thin interrupt handler for qdio
* @alsi: pointer to adapter local summary indicator * @alsi: pointer to adapter local summary indicator
...@@ -120,35 +149,18 @@ static void tiqdio_thinint_handler(void *alsi, void *data) ...@@ -120,35 +149,18 @@ static void tiqdio_thinint_handler(void *alsi, void *data)
/* check for work on all inbound thinint queues */ /* check for work on all inbound thinint queues */
list_for_each_entry_rcu(q, &tiq_list, entry) { list_for_each_entry_rcu(q, &tiq_list, entry) {
struct qdio_irq *irq;
/* only process queues from changed sets */ /* only process queues from changed sets */
if (unlikely(shared_ind(q->irq_ptr->dsci))) { irq = q->irq_ptr;
if (unlikely(references_shared_dsci(irq))) {
if (!si_used) if (!si_used)
continue; continue;
} else if (!*q->irq_ptr->dsci) } else if (!*irq->dsci)
continue; continue;
if (q->u.in.queue_start_poll) { tiqdio_call_inq_handlers(irq);
/* skip if polling is enabled or already in work */
if (test_and_set_bit(QDIO_QUEUE_IRQS_DISABLED,
&q->u.in.queue_irq_state)) {
qperf_inc(q, int_discarded);
continue;
}
/* avoid dsci clear here, done after processing */
q->u.in.queue_start_poll(q->irq_ptr->cdev, q->nr,
q->irq_ptr->int_parm);
} else {
/* only clear it if the indicator is non-shared */
if (!shared_ind(q->irq_ptr->dsci))
xchg(q->irq_ptr->dsci, 0);
/*
* Call inbound processing but not directly
* since that could starve other thinint queues.
*/
tasklet_schedule(&q->tasklet);
}
qperf_inc(q, adapter_int); qperf_inc(q, adapter_int);
} }
rcu_read_unlock(); rcu_read_unlock();
......
...@@ -3939,6 +3939,7 @@ static int qeth_qdio_establish(struct qeth_card *card) ...@@ -3939,6 +3939,7 @@ static int qeth_qdio_establish(struct qeth_card *card)
struct qdio_initialize init_data; struct qdio_initialize init_data;
char *qib_param_field; char *qib_param_field;
struct qdio_buffer **in_sbal_ptrs; struct qdio_buffer **in_sbal_ptrs;
void (**queue_start_poll) (struct ccw_device *, int, unsigned long);
struct qdio_buffer **out_sbal_ptrs; struct qdio_buffer **out_sbal_ptrs;
int i, j, k; int i, j, k;
int rc = 0; int rc = 0;
...@@ -3947,8 +3948,10 @@ static int qeth_qdio_establish(struct qeth_card *card) ...@@ -3947,8 +3948,10 @@ static int qeth_qdio_establish(struct qeth_card *card)
qib_param_field = kzalloc(QDIO_MAX_BUFFERS_PER_Q * sizeof(char), qib_param_field = kzalloc(QDIO_MAX_BUFFERS_PER_Q * sizeof(char),
GFP_KERNEL); GFP_KERNEL);
if (!qib_param_field) if (!qib_param_field) {
return -ENOMEM; rc = -ENOMEM;
goto out_free_nothing;
}
qeth_create_qib_param_field(card, qib_param_field); qeth_create_qib_param_field(card, qib_param_field);
qeth_create_qib_param_field_blkt(card, qib_param_field); qeth_create_qib_param_field_blkt(card, qib_param_field);
...@@ -3956,20 +3959,26 @@ static int qeth_qdio_establish(struct qeth_card *card) ...@@ -3956,20 +3959,26 @@ static int qeth_qdio_establish(struct qeth_card *card)
in_sbal_ptrs = kmalloc(QDIO_MAX_BUFFERS_PER_Q * sizeof(void *), in_sbal_ptrs = kmalloc(QDIO_MAX_BUFFERS_PER_Q * sizeof(void *),
GFP_KERNEL); GFP_KERNEL);
if (!in_sbal_ptrs) { if (!in_sbal_ptrs) {
kfree(qib_param_field); rc = -ENOMEM;
return -ENOMEM; goto out_free_qib_param;
} }
for (i = 0; i < QDIO_MAX_BUFFERS_PER_Q; ++i) for (i = 0; i < QDIO_MAX_BUFFERS_PER_Q; ++i)
in_sbal_ptrs[i] = (struct qdio_buffer *) in_sbal_ptrs[i] = (struct qdio_buffer *)
virt_to_phys(card->qdio.in_q->bufs[i].buffer); virt_to_phys(card->qdio.in_q->bufs[i].buffer);
queue_start_poll = kmalloc(sizeof(void *) * 1, GFP_KERNEL);
if (!queue_start_poll) {
rc = -ENOMEM;
goto out_free_in_sbals;
}
queue_start_poll[0] = card->discipline.start_poll;
out_sbal_ptrs = out_sbal_ptrs =
kmalloc(card->qdio.no_out_queues * QDIO_MAX_BUFFERS_PER_Q * kmalloc(card->qdio.no_out_queues * QDIO_MAX_BUFFERS_PER_Q *
sizeof(void *), GFP_KERNEL); sizeof(void *), GFP_KERNEL);
if (!out_sbal_ptrs) { if (!out_sbal_ptrs) {
kfree(in_sbal_ptrs); rc = -ENOMEM;
kfree(qib_param_field); goto out_free_queue_start_poll;
return -ENOMEM;
} }
for (i = 0, k = 0; i < card->qdio.no_out_queues; ++i) for (i = 0, k = 0; i < card->qdio.no_out_queues; ++i)
for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j, ++k) { for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j, ++k) {
...@@ -3986,7 +3995,7 @@ static int qeth_qdio_establish(struct qeth_card *card) ...@@ -3986,7 +3995,7 @@ static int qeth_qdio_establish(struct qeth_card *card)
init_data.no_output_qs = card->qdio.no_out_queues; init_data.no_output_qs = card->qdio.no_out_queues;
init_data.input_handler = card->discipline.input_handler; init_data.input_handler = card->discipline.input_handler;
init_data.output_handler = card->discipline.output_handler; init_data.output_handler = card->discipline.output_handler;
init_data.queue_start_poll = card->discipline.start_poll; init_data.queue_start_poll = queue_start_poll;
init_data.int_parm = (unsigned long) card; init_data.int_parm = (unsigned long) card;
init_data.input_sbal_addr_array = (void **) in_sbal_ptrs; init_data.input_sbal_addr_array = (void **) in_sbal_ptrs;
init_data.output_sbal_addr_array = (void **) out_sbal_ptrs; init_data.output_sbal_addr_array = (void **) out_sbal_ptrs;
...@@ -4008,8 +4017,13 @@ static int qeth_qdio_establish(struct qeth_card *card) ...@@ -4008,8 +4017,13 @@ static int qeth_qdio_establish(struct qeth_card *card)
} }
out: out:
kfree(out_sbal_ptrs); kfree(out_sbal_ptrs);
out_free_queue_start_poll:
kfree(queue_start_poll);
out_free_in_sbals:
kfree(in_sbal_ptrs); kfree(in_sbal_ptrs);
out_free_qib_param:
kfree(qib_param_field); kfree(qib_param_field);
out_free_nothing:
return rc; return rc;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment