mirror of
https://github.com/FEX-Emu/linux.git
synced 2024-12-28 04:17:47 +00:00
432ac5e04b
Remove a memset hack that relied on the internal layout of the qdio_irq struct and move the per device statistics data into an own cache line to avoid cache line bashing between the inbound and the outbound queue tasklets. Also reduce the number of allocated queues from 32 to 4 which is the current maximum. That saves a cache line in struct qdio_irq. Signed-off-by: Jan Glauber <jang@linux.vnet.ibm.com> Signed-off-by: Martin Schwidefsky <schwidefsky@de.ibm.com>
487 lines
13 KiB
C
487 lines
13 KiB
C
/*
|
|
* driver/s390/cio/qdio_setup.c
|
|
*
|
|
* qdio queue initialization
|
|
*
|
|
* Copyright (C) IBM Corp. 2008
|
|
* Author(s): Jan Glauber <jang@linux.vnet.ibm.com>
|
|
*/
|
|
#include <linux/kernel.h>
|
|
#include <linux/slab.h>
|
|
#include <asm/qdio.h>
|
|
|
|
#include "cio.h"
|
|
#include "css.h"
|
|
#include "device.h"
|
|
#include "ioasm.h"
|
|
#include "chsc.h"
|
|
#include "qdio.h"
|
|
#include "qdio_debug.h"
|
|
|
|
static struct kmem_cache *qdio_q_cache;
|
|
|
|
/*
|
|
* qebsm is only available under 64bit but the adapter sets the feature
|
|
* flag anyway, so we manually override it.
|
|
*/
|
|
static inline int qebsm_possible(void)
|
|
{
|
|
#ifdef CONFIG_64BIT
|
|
return css_general_characteristics.qebsm;
|
|
#endif
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* qib_param_field: pointer to 128 bytes or NULL, if no param field
|
|
* nr_input_qs: pointer to nr_queues*128 words of data or NULL
|
|
*/
|
|
static void set_impl_params(struct qdio_irq *irq_ptr,
|
|
unsigned int qib_param_field_format,
|
|
unsigned char *qib_param_field,
|
|
unsigned long *input_slib_elements,
|
|
unsigned long *output_slib_elements)
|
|
{
|
|
struct qdio_q *q;
|
|
int i, j;
|
|
|
|
if (!irq_ptr)
|
|
return;
|
|
|
|
irq_ptr->qib.pfmt = qib_param_field_format;
|
|
if (qib_param_field)
|
|
memcpy(irq_ptr->qib.parm, qib_param_field,
|
|
QDIO_MAX_BUFFERS_PER_Q);
|
|
|
|
if (!input_slib_elements)
|
|
goto output;
|
|
|
|
for_each_input_queue(irq_ptr, q, i) {
|
|
for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++)
|
|
q->slib->slibe[j].parms =
|
|
input_slib_elements[i * QDIO_MAX_BUFFERS_PER_Q + j];
|
|
}
|
|
output:
|
|
if (!output_slib_elements)
|
|
return;
|
|
|
|
for_each_output_queue(irq_ptr, q, i) {
|
|
for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++)
|
|
q->slib->slibe[j].parms =
|
|
output_slib_elements[i * QDIO_MAX_BUFFERS_PER_Q + j];
|
|
}
|
|
}
|
|
|
|
static int __qdio_allocate_qs(struct qdio_q **irq_ptr_qs, int nr_queues)
|
|
{
|
|
struct qdio_q *q;
|
|
int i;
|
|
|
|
for (i = 0; i < nr_queues; i++) {
|
|
q = kmem_cache_alloc(qdio_q_cache, GFP_KERNEL);
|
|
if (!q)
|
|
return -ENOMEM;
|
|
|
|
q->slib = (struct slib *) __get_free_page(GFP_KERNEL);
|
|
if (!q->slib) {
|
|
kmem_cache_free(qdio_q_cache, q);
|
|
return -ENOMEM;
|
|
}
|
|
irq_ptr_qs[i] = q;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int qdio_allocate_qs(struct qdio_irq *irq_ptr, int nr_input_qs, int nr_output_qs)
|
|
{
|
|
int rc;
|
|
|
|
rc = __qdio_allocate_qs(irq_ptr->input_qs, nr_input_qs);
|
|
if (rc)
|
|
return rc;
|
|
rc = __qdio_allocate_qs(irq_ptr->output_qs, nr_output_qs);
|
|
return rc;
|
|
}
|
|
|
|
static void setup_queues_misc(struct qdio_q *q, struct qdio_irq *irq_ptr,
|
|
qdio_handler_t *handler, int i)
|
|
{
|
|
/* must be cleared by every qdio_establish */
|
|
memset(q, 0, ((char *)&q->slib) - ((char *)q));
|
|
memset(q->slib, 0, PAGE_SIZE);
|
|
|
|
q->irq_ptr = irq_ptr;
|
|
q->mask = 1 << (31 - i);
|
|
q->nr = i;
|
|
q->handler = handler;
|
|
}
|
|
|
|
static void setup_storage_lists(struct qdio_q *q, struct qdio_irq *irq_ptr,
|
|
void **sbals_array, int i)
|
|
{
|
|
struct qdio_q *prev;
|
|
int j;
|
|
|
|
DBF_HEX(&q, sizeof(void *));
|
|
q->sl = (struct sl *)((char *)q->slib + PAGE_SIZE / 2);
|
|
|
|
/* fill in sbal */
|
|
for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++) {
|
|
q->sbal[j] = *sbals_array++;
|
|
BUG_ON((unsigned long)q->sbal[j] & 0xff);
|
|
}
|
|
|
|
/* fill in slib */
|
|
if (i > 0) {
|
|
prev = (q->is_input_q) ? irq_ptr->input_qs[i - 1]
|
|
: irq_ptr->output_qs[i - 1];
|
|
prev->slib->nsliba = (unsigned long)q->slib;
|
|
}
|
|
|
|
q->slib->sla = (unsigned long)q->sl;
|
|
q->slib->slsba = (unsigned long)&q->slsb.val[0];
|
|
|
|
/* fill in sl */
|
|
for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; j++)
|
|
q->sl->element[j].sbal = (unsigned long)q->sbal[j];
|
|
}
|
|
|
|
static void setup_queues(struct qdio_irq *irq_ptr,
|
|
struct qdio_initialize *qdio_init)
|
|
{
|
|
struct qdio_q *q;
|
|
void **input_sbal_array = qdio_init->input_sbal_addr_array;
|
|
void **output_sbal_array = qdio_init->output_sbal_addr_array;
|
|
int i;
|
|
|
|
for_each_input_queue(irq_ptr, q, i) {
|
|
DBF_EVENT("in-q:%1d", i);
|
|
setup_queues_misc(q, irq_ptr, qdio_init->input_handler, i);
|
|
|
|
q->is_input_q = 1;
|
|
setup_storage_lists(q, irq_ptr, input_sbal_array, i);
|
|
input_sbal_array += QDIO_MAX_BUFFERS_PER_Q;
|
|
|
|
if (is_thinint_irq(irq_ptr))
|
|
tasklet_init(&q->tasklet, tiqdio_inbound_processing,
|
|
(unsigned long) q);
|
|
else
|
|
tasklet_init(&q->tasklet, qdio_inbound_processing,
|
|
(unsigned long) q);
|
|
}
|
|
|
|
for_each_output_queue(irq_ptr, q, i) {
|
|
DBF_EVENT("outq:%1d", i);
|
|
setup_queues_misc(q, irq_ptr, qdio_init->output_handler, i);
|
|
|
|
q->is_input_q = 0;
|
|
setup_storage_lists(q, irq_ptr, output_sbal_array, i);
|
|
output_sbal_array += QDIO_MAX_BUFFERS_PER_Q;
|
|
|
|
tasklet_init(&q->tasklet, qdio_outbound_processing,
|
|
(unsigned long) q);
|
|
setup_timer(&q->u.out.timer, (void(*)(unsigned long))
|
|
&qdio_outbound_timer, (unsigned long)q);
|
|
}
|
|
}
|
|
|
|
static void process_ac_flags(struct qdio_irq *irq_ptr, unsigned char qdioac)
|
|
{
|
|
if (qdioac & AC1_SIGA_INPUT_NEEDED)
|
|
irq_ptr->siga_flag.input = 1;
|
|
if (qdioac & AC1_SIGA_OUTPUT_NEEDED)
|
|
irq_ptr->siga_flag.output = 1;
|
|
if (qdioac & AC1_SIGA_SYNC_NEEDED)
|
|
irq_ptr->siga_flag.sync = 1;
|
|
if (qdioac & AC1_AUTOMATIC_SYNC_ON_THININT)
|
|
irq_ptr->siga_flag.no_sync_ti = 1;
|
|
if (qdioac & AC1_AUTOMATIC_SYNC_ON_OUT_PCI)
|
|
irq_ptr->siga_flag.no_sync_out_pci = 1;
|
|
|
|
if (irq_ptr->siga_flag.no_sync_out_pci &&
|
|
irq_ptr->siga_flag.no_sync_ti)
|
|
irq_ptr->siga_flag.no_sync_out_ti = 1;
|
|
}
|
|
|
|
static void check_and_setup_qebsm(struct qdio_irq *irq_ptr,
|
|
unsigned char qdioac, unsigned long token)
|
|
{
|
|
if (!(irq_ptr->qib.rflags & QIB_RFLAGS_ENABLE_QEBSM))
|
|
goto no_qebsm;
|
|
if (!(qdioac & AC1_SC_QEBSM_AVAILABLE) ||
|
|
(!(qdioac & AC1_SC_QEBSM_ENABLED)))
|
|
goto no_qebsm;
|
|
|
|
irq_ptr->sch_token = token;
|
|
|
|
DBF_EVENT("V=V:1");
|
|
DBF_EVENT("%8lx", irq_ptr->sch_token);
|
|
return;
|
|
|
|
no_qebsm:
|
|
irq_ptr->sch_token = 0;
|
|
irq_ptr->qib.rflags &= ~QIB_RFLAGS_ENABLE_QEBSM;
|
|
DBF_EVENT("noV=V");
|
|
}
|
|
|
|
/*
|
|
* If there is a qdio_irq we use the chsc_page and store the information
|
|
* in the qdio_irq, otherwise we copy it to the specified structure.
|
|
*/
|
|
int qdio_setup_get_ssqd(struct qdio_irq *irq_ptr,
|
|
struct subchannel_id *schid,
|
|
struct qdio_ssqd_desc *data)
|
|
{
|
|
struct chsc_ssqd_area *ssqd;
|
|
int rc;
|
|
|
|
DBF_EVENT("getssqd:%4x", schid->sch_no);
|
|
if (irq_ptr != NULL)
|
|
ssqd = (struct chsc_ssqd_area *)irq_ptr->chsc_page;
|
|
else
|
|
ssqd = (struct chsc_ssqd_area *)__get_free_page(GFP_KERNEL);
|
|
memset(ssqd, 0, PAGE_SIZE);
|
|
|
|
ssqd->request = (struct chsc_header) {
|
|
.length = 0x0010,
|
|
.code = 0x0024,
|
|
};
|
|
ssqd->first_sch = schid->sch_no;
|
|
ssqd->last_sch = schid->sch_no;
|
|
ssqd->ssid = schid->ssid;
|
|
|
|
if (chsc(ssqd))
|
|
return -EIO;
|
|
rc = chsc_error_from_response(ssqd->response.code);
|
|
if (rc)
|
|
return rc;
|
|
|
|
if (!(ssqd->qdio_ssqd.flags & CHSC_FLAG_QDIO_CAPABILITY) ||
|
|
!(ssqd->qdio_ssqd.flags & CHSC_FLAG_VALIDITY) ||
|
|
(ssqd->qdio_ssqd.sch != schid->sch_no))
|
|
return -EINVAL;
|
|
|
|
if (irq_ptr != NULL)
|
|
memcpy(&irq_ptr->ssqd_desc, &ssqd->qdio_ssqd,
|
|
sizeof(struct qdio_ssqd_desc));
|
|
else {
|
|
memcpy(data, &ssqd->qdio_ssqd,
|
|
sizeof(struct qdio_ssqd_desc));
|
|
free_page((unsigned long)ssqd);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
void qdio_setup_ssqd_info(struct qdio_irq *irq_ptr)
|
|
{
|
|
unsigned char qdioac;
|
|
int rc;
|
|
|
|
rc = qdio_setup_get_ssqd(irq_ptr, &irq_ptr->schid, NULL);
|
|
if (rc) {
|
|
DBF_ERROR("%4x ssqd ERR", irq_ptr->schid.sch_no);
|
|
DBF_ERROR("rc:%x", rc);
|
|
/* all flags set, worst case */
|
|
qdioac = AC1_SIGA_INPUT_NEEDED | AC1_SIGA_OUTPUT_NEEDED |
|
|
AC1_SIGA_SYNC_NEEDED;
|
|
} else
|
|
qdioac = irq_ptr->ssqd_desc.qdioac1;
|
|
|
|
check_and_setup_qebsm(irq_ptr, qdioac, irq_ptr->ssqd_desc.sch_token);
|
|
process_ac_flags(irq_ptr, qdioac);
|
|
DBF_EVENT("qdioac:%4x", qdioac);
|
|
}
|
|
|
|
void qdio_release_memory(struct qdio_irq *irq_ptr)
|
|
{
|
|
struct qdio_q *q;
|
|
int i;
|
|
|
|
/*
|
|
* Must check queue array manually since irq_ptr->nr_input_queues /
|
|
* irq_ptr->nr_input_queues may not yet be set.
|
|
*/
|
|
for (i = 0; i < QDIO_MAX_QUEUES_PER_IRQ; i++) {
|
|
q = irq_ptr->input_qs[i];
|
|
if (q) {
|
|
free_page((unsigned long) q->slib);
|
|
kmem_cache_free(qdio_q_cache, q);
|
|
}
|
|
}
|
|
for (i = 0; i < QDIO_MAX_QUEUES_PER_IRQ; i++) {
|
|
q = irq_ptr->output_qs[i];
|
|
if (q) {
|
|
free_page((unsigned long) q->slib);
|
|
kmem_cache_free(qdio_q_cache, q);
|
|
}
|
|
}
|
|
free_page((unsigned long) irq_ptr->qdr);
|
|
free_page(irq_ptr->chsc_page);
|
|
free_page((unsigned long) irq_ptr);
|
|
}
|
|
|
|
static void __qdio_allocate_fill_qdr(struct qdio_irq *irq_ptr,
|
|
struct qdio_q **irq_ptr_qs,
|
|
int i, int nr)
|
|
{
|
|
irq_ptr->qdr->qdf0[i + nr].sliba =
|
|
(unsigned long)irq_ptr_qs[i]->slib;
|
|
|
|
irq_ptr->qdr->qdf0[i + nr].sla =
|
|
(unsigned long)irq_ptr_qs[i]->sl;
|
|
|
|
irq_ptr->qdr->qdf0[i + nr].slsba =
|
|
(unsigned long)&irq_ptr_qs[i]->slsb.val[0];
|
|
|
|
irq_ptr->qdr->qdf0[i + nr].akey = PAGE_DEFAULT_KEY >> 4;
|
|
irq_ptr->qdr->qdf0[i + nr].bkey = PAGE_DEFAULT_KEY >> 4;
|
|
irq_ptr->qdr->qdf0[i + nr].ckey = PAGE_DEFAULT_KEY >> 4;
|
|
irq_ptr->qdr->qdf0[i + nr].dkey = PAGE_DEFAULT_KEY >> 4;
|
|
}
|
|
|
|
static void setup_qdr(struct qdio_irq *irq_ptr,
|
|
struct qdio_initialize *qdio_init)
|
|
{
|
|
int i;
|
|
|
|
irq_ptr->qdr->qfmt = qdio_init->q_format;
|
|
irq_ptr->qdr->iqdcnt = qdio_init->no_input_qs;
|
|
irq_ptr->qdr->oqdcnt = qdio_init->no_output_qs;
|
|
irq_ptr->qdr->iqdsz = sizeof(struct qdesfmt0) / 4; /* size in words */
|
|
irq_ptr->qdr->oqdsz = sizeof(struct qdesfmt0) / 4;
|
|
irq_ptr->qdr->qiba = (unsigned long)&irq_ptr->qib;
|
|
irq_ptr->qdr->qkey = PAGE_DEFAULT_KEY >> 4;
|
|
|
|
for (i = 0; i < qdio_init->no_input_qs; i++)
|
|
__qdio_allocate_fill_qdr(irq_ptr, irq_ptr->input_qs, i, 0);
|
|
|
|
for (i = 0; i < qdio_init->no_output_qs; i++)
|
|
__qdio_allocate_fill_qdr(irq_ptr, irq_ptr->output_qs, i,
|
|
qdio_init->no_input_qs);
|
|
}
|
|
|
|
static void setup_qib(struct qdio_irq *irq_ptr,
|
|
struct qdio_initialize *init_data)
|
|
{
|
|
if (qebsm_possible())
|
|
irq_ptr->qib.rflags |= QIB_RFLAGS_ENABLE_QEBSM;
|
|
|
|
irq_ptr->qib.qfmt = init_data->q_format;
|
|
if (init_data->no_input_qs)
|
|
irq_ptr->qib.isliba =
|
|
(unsigned long)(irq_ptr->input_qs[0]->slib);
|
|
if (init_data->no_output_qs)
|
|
irq_ptr->qib.osliba =
|
|
(unsigned long)(irq_ptr->output_qs[0]->slib);
|
|
memcpy(irq_ptr->qib.ebcnam, init_data->adapter_name, 8);
|
|
}
|
|
|
|
int qdio_setup_irq(struct qdio_initialize *init_data)
|
|
{
|
|
struct ciw *ciw;
|
|
struct qdio_irq *irq_ptr = init_data->cdev->private->qdio_data;
|
|
int rc;
|
|
|
|
memset(&irq_ptr->qib, 0, sizeof(irq_ptr->qib));
|
|
memset(&irq_ptr->siga_flag, 0, sizeof(irq_ptr->siga_flag));
|
|
memset(&irq_ptr->ccw, 0, sizeof(irq_ptr->ccw));
|
|
memset(&irq_ptr->ssqd_desc, 0, sizeof(irq_ptr->ssqd_desc));
|
|
memset(&irq_ptr->perf_stat, 0, sizeof(irq_ptr->perf_stat));
|
|
|
|
irq_ptr->debugfs_dev = irq_ptr->debugfs_perf = NULL;
|
|
irq_ptr->sch_token = irq_ptr->state = irq_ptr->perf_stat_enabled = 0;
|
|
|
|
/* wipes qib.ac, required by ar7063 */
|
|
memset(irq_ptr->qdr, 0, sizeof(struct qdr));
|
|
|
|
irq_ptr->int_parm = init_data->int_parm;
|
|
irq_ptr->nr_input_qs = init_data->no_input_qs;
|
|
irq_ptr->nr_output_qs = init_data->no_output_qs;
|
|
|
|
irq_ptr->schid = ccw_device_get_subchannel_id(init_data->cdev);
|
|
irq_ptr->cdev = init_data->cdev;
|
|
setup_queues(irq_ptr, init_data);
|
|
|
|
setup_qib(irq_ptr, init_data);
|
|
qdio_setup_thinint(irq_ptr);
|
|
set_impl_params(irq_ptr, init_data->qib_param_field_format,
|
|
init_data->qib_param_field,
|
|
init_data->input_slib_elements,
|
|
init_data->output_slib_elements);
|
|
|
|
/* fill input and output descriptors */
|
|
setup_qdr(irq_ptr, init_data);
|
|
|
|
/* qdr, qib, sls, slsbs, slibs, sbales are filled now */
|
|
|
|
/* get qdio commands */
|
|
ciw = ccw_device_get_ciw(init_data->cdev, CIW_TYPE_EQUEUE);
|
|
if (!ciw) {
|
|
DBF_ERROR("%4x NO EQ", irq_ptr->schid.sch_no);
|
|
rc = -EINVAL;
|
|
goto out_err;
|
|
}
|
|
irq_ptr->equeue = *ciw;
|
|
|
|
ciw = ccw_device_get_ciw(init_data->cdev, CIW_TYPE_AQUEUE);
|
|
if (!ciw) {
|
|
DBF_ERROR("%4x NO AQ", irq_ptr->schid.sch_no);
|
|
rc = -EINVAL;
|
|
goto out_err;
|
|
}
|
|
irq_ptr->aqueue = *ciw;
|
|
|
|
/* set new interrupt handler */
|
|
irq_ptr->orig_handler = init_data->cdev->handler;
|
|
init_data->cdev->handler = qdio_int_handler;
|
|
return 0;
|
|
out_err:
|
|
qdio_release_memory(irq_ptr);
|
|
return rc;
|
|
}
|
|
|
|
void qdio_print_subchannel_info(struct qdio_irq *irq_ptr,
|
|
struct ccw_device *cdev)
|
|
{
|
|
char s[80];
|
|
|
|
snprintf(s, 80, "qdio: %s %s on SC %x using "
|
|
"AI:%d QEBSM:%d PCI:%d TDD:%d SIGA:%s%s%s%s%s%s\n",
|
|
dev_name(&cdev->dev),
|
|
(irq_ptr->qib.qfmt == QDIO_QETH_QFMT) ? "OSA" :
|
|
((irq_ptr->qib.qfmt == QDIO_ZFCP_QFMT) ? "ZFCP" : "HS"),
|
|
irq_ptr->schid.sch_no,
|
|
is_thinint_irq(irq_ptr),
|
|
(irq_ptr->sch_token) ? 1 : 0,
|
|
(irq_ptr->qib.ac & QIB_AC_OUTBOUND_PCI_SUPPORTED) ? 1 : 0,
|
|
css_general_characteristics.aif_tdd,
|
|
(irq_ptr->siga_flag.input) ? "R" : " ",
|
|
(irq_ptr->siga_flag.output) ? "W" : " ",
|
|
(irq_ptr->siga_flag.sync) ? "S" : " ",
|
|
(!irq_ptr->siga_flag.no_sync_ti) ? "A" : " ",
|
|
(!irq_ptr->siga_flag.no_sync_out_ti) ? "O" : " ",
|
|
(!irq_ptr->siga_flag.no_sync_out_pci) ? "P" : " ");
|
|
printk(KERN_INFO "%s", s);
|
|
}
|
|
|
|
int __init qdio_setup_init(void)
|
|
{
|
|
qdio_q_cache = kmem_cache_create("qdio_q", sizeof(struct qdio_q),
|
|
256, 0, NULL);
|
|
if (!qdio_q_cache)
|
|
return -ENOMEM;
|
|
|
|
/* Check for OSA/FCP thin interrupts (bit 67). */
|
|
DBF_EVENT("thinint:%1d",
|
|
(css_general_characteristics.aif_osa) ? 1 : 0);
|
|
|
|
/* Check for QEBSM support in general (bit 58). */
|
|
DBF_EVENT("cssQEBSM:%1d", (qebsm_possible()) ? 1 : 0);
|
|
return 0;
|
|
}
|
|
|
|
void qdio_setup_exit(void)
|
|
{
|
|
kmem_cache_destroy(qdio_q_cache);
|
|
}
|