mirror of
https://github.com/FEX-Emu/linux.git
synced 2024-12-26 11:28:28 +00:00
6473d160b4
I noticed that many source files include <linux/pci.h> while they do not appear to need it. Here is an attempt to clean it all up. In order to find all possibly affected files, I searched for all files including <linux/pci.h> but without any other occurence of "pci" or "PCI". I removed the include statement from all of these, then I compiled an allmodconfig kernel on both i386 and x86_64 and fixed the false positives manually. My tests covered 66% of the affected files, so there could be false positives remaining. Untested files are: arch/alpha/kernel/err_common.c arch/alpha/kernel/err_ev6.c arch/alpha/kernel/err_ev7.c arch/ia64/sn/kernel/huberror.c arch/ia64/sn/kernel/xpnet.c arch/m68knommu/kernel/dma.c arch/mips/lib/iomap.c arch/powerpc/platforms/pseries/ras.c arch/ppc/8260_io/enet.c arch/ppc/8260_io/fcc_enet.c arch/ppc/8xx_io/enet.c arch/ppc/syslib/ppc4xx_sgdma.c arch/sh64/mach-cayman/iomap.c arch/xtensa/kernel/xtensa_ksyms.c arch/xtensa/platform-iss/setup.c drivers/i2c/busses/i2c-at91.c drivers/i2c/busses/i2c-mpc.c drivers/media/video/saa711x.c drivers/misc/hdpuftrs/hdpu_cpustate.c drivers/misc/hdpuftrs/hdpu_nexus.c drivers/net/au1000_eth.c drivers/net/fec_8xx/fec_main.c drivers/net/fec_8xx/fec_mii.c drivers/net/fs_enet/fs_enet-main.c drivers/net/fs_enet/mac-fcc.c drivers/net/fs_enet/mac-fec.c drivers/net/fs_enet/mac-scc.c drivers/net/fs_enet/mii-bitbang.c drivers/net/fs_enet/mii-fec.c drivers/net/ibm_emac/ibm_emac_core.c drivers/net/lasi_82596.c drivers/parisc/hppb.c drivers/sbus/sbus.c drivers/video/g364fb.c drivers/video/platinumfb.c drivers/video/stifb.c drivers/video/valkyriefb.c include/asm-arm/arch-ixp4xx/dma.h sound/oss/au1550_ac97.c I would welcome test reports for these files. I am fine with removing the untested files from the patch if the general opinion is that these changes aren't safe. The tested part would still be nice to have. Note that this patch depends on another header fixup patch I submitted to LKML yesterday: [PATCH] scatterlist.h needs types.h http://lkml.org/lkml/2007/3/01/141 Signed-off-by: Jean Delvare <khali@linux-fr.org> Cc: Badari Pulavarty <pbadari@us.ibm.com> Signed-off-by: Greg Kroah-Hartman <gregkh@suse.de>
366 lines
8.8 KiB
C
366 lines
8.8 KiB
C
/*
|
|
* Copyright (c) 2006 QLogic, Inc. All rights reserved.
|
|
* Copyright (c) 2003, 2004, 2005, 2006 PathScale, Inc. All rights reserved.
|
|
*
|
|
* This software is available to you under a choice of one of two
|
|
* licenses. You may choose to be licensed under the terms of the GNU
|
|
* General Public License (GPL) Version 2, available from the file
|
|
* COPYING in the main directory of this source tree, or the
|
|
* OpenIB.org BSD license below:
|
|
*
|
|
* Redistribution and use in source and binary forms, with or
|
|
* without modification, are permitted provided that the following
|
|
* conditions are met:
|
|
*
|
|
* - Redistributions of source code must retain the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer.
|
|
*
|
|
* - Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer in the documentation and/or other materials
|
|
* provided with the distribution.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*/
|
|
|
|
/*
|
|
* These are the routines used by layered drivers, currently just the
|
|
* layered ethernet driver and verbs layer.
|
|
*/
|
|
|
|
#include <linux/io.h>
|
|
#include <asm/byteorder.h>
|
|
|
|
#include "ipath_kernel.h"
|
|
#include "ipath_layer.h"
|
|
#include "ipath_verbs.h"
|
|
#include "ipath_common.h"
|
|
|
|
/* Acquire before ipath_devs_lock. */
|
|
static DEFINE_MUTEX(ipath_layer_mutex);
|
|
|
|
u16 ipath_layer_rcv_opcode;
|
|
|
|
static int (*layer_intr)(void *, u32);
|
|
static int (*layer_rcv)(void *, void *, struct sk_buff *);
|
|
static int (*layer_rcv_lid)(void *, void *);
|
|
|
|
static void *(*layer_add_one)(int, struct ipath_devdata *);
|
|
static void (*layer_remove_one)(void *);
|
|
|
|
int __ipath_layer_intr(struct ipath_devdata *dd, u32 arg)
|
|
{
|
|
int ret = -ENODEV;
|
|
|
|
if (dd->ipath_layer.l_arg && layer_intr)
|
|
ret = layer_intr(dd->ipath_layer.l_arg, arg);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int ipath_layer_intr(struct ipath_devdata *dd, u32 arg)
|
|
{
|
|
int ret;
|
|
|
|
mutex_lock(&ipath_layer_mutex);
|
|
|
|
ret = __ipath_layer_intr(dd, arg);
|
|
|
|
mutex_unlock(&ipath_layer_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int __ipath_layer_rcv(struct ipath_devdata *dd, void *hdr,
|
|
struct sk_buff *skb)
|
|
{
|
|
int ret = -ENODEV;
|
|
|
|
if (dd->ipath_layer.l_arg && layer_rcv)
|
|
ret = layer_rcv(dd->ipath_layer.l_arg, hdr, skb);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int __ipath_layer_rcv_lid(struct ipath_devdata *dd, void *hdr)
|
|
{
|
|
int ret = -ENODEV;
|
|
|
|
if (dd->ipath_layer.l_arg && layer_rcv_lid)
|
|
ret = layer_rcv_lid(dd->ipath_layer.l_arg, hdr);
|
|
|
|
return ret;
|
|
}
|
|
|
|
void ipath_layer_lid_changed(struct ipath_devdata *dd)
|
|
{
|
|
mutex_lock(&ipath_layer_mutex);
|
|
|
|
if (dd->ipath_layer.l_arg && layer_intr)
|
|
layer_intr(dd->ipath_layer.l_arg, IPATH_LAYER_INT_LID);
|
|
|
|
mutex_unlock(&ipath_layer_mutex);
|
|
}
|
|
|
|
void ipath_layer_add(struct ipath_devdata *dd)
|
|
{
|
|
mutex_lock(&ipath_layer_mutex);
|
|
|
|
if (layer_add_one)
|
|
dd->ipath_layer.l_arg =
|
|
layer_add_one(dd->ipath_unit, dd);
|
|
|
|
mutex_unlock(&ipath_layer_mutex);
|
|
}
|
|
|
|
void ipath_layer_remove(struct ipath_devdata *dd)
|
|
{
|
|
mutex_lock(&ipath_layer_mutex);
|
|
|
|
if (dd->ipath_layer.l_arg && layer_remove_one) {
|
|
layer_remove_one(dd->ipath_layer.l_arg);
|
|
dd->ipath_layer.l_arg = NULL;
|
|
}
|
|
|
|
mutex_unlock(&ipath_layer_mutex);
|
|
}
|
|
|
|
int ipath_layer_register(void *(*l_add)(int, struct ipath_devdata *),
|
|
void (*l_remove)(void *),
|
|
int (*l_intr)(void *, u32),
|
|
int (*l_rcv)(void *, void *, struct sk_buff *),
|
|
u16 l_rcv_opcode,
|
|
int (*l_rcv_lid)(void *, void *))
|
|
{
|
|
struct ipath_devdata *dd, *tmp;
|
|
unsigned long flags;
|
|
|
|
mutex_lock(&ipath_layer_mutex);
|
|
|
|
layer_add_one = l_add;
|
|
layer_remove_one = l_remove;
|
|
layer_intr = l_intr;
|
|
layer_rcv = l_rcv;
|
|
layer_rcv_lid = l_rcv_lid;
|
|
ipath_layer_rcv_opcode = l_rcv_opcode;
|
|
|
|
spin_lock_irqsave(&ipath_devs_lock, flags);
|
|
|
|
list_for_each_entry_safe(dd, tmp, &ipath_dev_list, ipath_list) {
|
|
if (!(dd->ipath_flags & IPATH_INITTED))
|
|
continue;
|
|
|
|
if (dd->ipath_layer.l_arg)
|
|
continue;
|
|
|
|
spin_unlock_irqrestore(&ipath_devs_lock, flags);
|
|
dd->ipath_layer.l_arg = l_add(dd->ipath_unit, dd);
|
|
spin_lock_irqsave(&ipath_devs_lock, flags);
|
|
}
|
|
|
|
spin_unlock_irqrestore(&ipath_devs_lock, flags);
|
|
mutex_unlock(&ipath_layer_mutex);
|
|
|
|
return 0;
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ipath_layer_register);
|
|
|
|
void ipath_layer_unregister(void)
|
|
{
|
|
struct ipath_devdata *dd, *tmp;
|
|
unsigned long flags;
|
|
|
|
mutex_lock(&ipath_layer_mutex);
|
|
spin_lock_irqsave(&ipath_devs_lock, flags);
|
|
|
|
list_for_each_entry_safe(dd, tmp, &ipath_dev_list, ipath_list) {
|
|
if (dd->ipath_layer.l_arg && layer_remove_one) {
|
|
spin_unlock_irqrestore(&ipath_devs_lock, flags);
|
|
layer_remove_one(dd->ipath_layer.l_arg);
|
|
spin_lock_irqsave(&ipath_devs_lock, flags);
|
|
dd->ipath_layer.l_arg = NULL;
|
|
}
|
|
}
|
|
|
|
spin_unlock_irqrestore(&ipath_devs_lock, flags);
|
|
|
|
layer_add_one = NULL;
|
|
layer_remove_one = NULL;
|
|
layer_intr = NULL;
|
|
layer_rcv = NULL;
|
|
layer_rcv_lid = NULL;
|
|
|
|
mutex_unlock(&ipath_layer_mutex);
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ipath_layer_unregister);
|
|
|
|
int ipath_layer_open(struct ipath_devdata *dd, u32 * pktmax)
|
|
{
|
|
int ret;
|
|
u32 intval = 0;
|
|
|
|
mutex_lock(&ipath_layer_mutex);
|
|
|
|
if (!dd->ipath_layer.l_arg) {
|
|
ret = -EINVAL;
|
|
goto bail;
|
|
}
|
|
|
|
ret = ipath_setrcvhdrsize(dd, IPATH_HEADER_QUEUE_WORDS);
|
|
|
|
if (ret < 0)
|
|
goto bail;
|
|
|
|
*pktmax = dd->ipath_ibmaxlen;
|
|
|
|
if (*dd->ipath_statusp & IPATH_STATUS_IB_READY)
|
|
intval |= IPATH_LAYER_INT_IF_UP;
|
|
if (dd->ipath_lid)
|
|
intval |= IPATH_LAYER_INT_LID;
|
|
if (dd->ipath_mlid)
|
|
intval |= IPATH_LAYER_INT_BCAST;
|
|
/*
|
|
* do this on open, in case low level is already up and
|
|
* just layered driver was reloaded, etc.
|
|
*/
|
|
if (intval)
|
|
layer_intr(dd->ipath_layer.l_arg, intval);
|
|
|
|
ret = 0;
|
|
bail:
|
|
mutex_unlock(&ipath_layer_mutex);
|
|
|
|
return ret;
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ipath_layer_open);
|
|
|
|
u16 ipath_layer_get_lid(struct ipath_devdata *dd)
|
|
{
|
|
return dd->ipath_lid;
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ipath_layer_get_lid);
|
|
|
|
/**
|
|
* ipath_layer_get_mac - get the MAC address
|
|
* @dd: the infinipath device
|
|
* @mac: the MAC is put here
|
|
*
|
|
* This is the EUID-64 OUI octets (top 3), then
|
|
* skip the next 2 (which should both be zero or 0xff).
|
|
* The returned MAC is in network order
|
|
* mac points to at least 6 bytes of buffer
|
|
* We assume that by the time the LID is set, that the GUID is as valid
|
|
* as it's ever going to be, rather than adding yet another status bit.
|
|
*/
|
|
|
|
int ipath_layer_get_mac(struct ipath_devdata *dd, u8 * mac)
|
|
{
|
|
u8 *guid;
|
|
|
|
guid = (u8 *) &dd->ipath_guid;
|
|
|
|
mac[0] = guid[0];
|
|
mac[1] = guid[1];
|
|
mac[2] = guid[2];
|
|
mac[3] = guid[5];
|
|
mac[4] = guid[6];
|
|
mac[5] = guid[7];
|
|
if ((guid[3] || guid[4]) && !(guid[3] == 0xff && guid[4] == 0xff))
|
|
ipath_dbg("Warning, guid bytes 3 and 4 not 0 or 0xffff: "
|
|
"%x %x\n", guid[3], guid[4]);
|
|
return 0;
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ipath_layer_get_mac);
|
|
|
|
u16 ipath_layer_get_bcast(struct ipath_devdata *dd)
|
|
{
|
|
return dd->ipath_mlid;
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ipath_layer_get_bcast);
|
|
|
|
int ipath_layer_send_hdr(struct ipath_devdata *dd, struct ether_header *hdr)
|
|
{
|
|
int ret = 0;
|
|
u32 __iomem *piobuf;
|
|
u32 plen, *uhdr;
|
|
size_t count;
|
|
__be16 vlsllnh;
|
|
|
|
if (!(dd->ipath_flags & IPATH_RCVHDRSZ_SET)) {
|
|
ipath_dbg("send while not open\n");
|
|
ret = -EINVAL;
|
|
} else
|
|
if ((dd->ipath_flags & (IPATH_LINKUNK | IPATH_LINKDOWN)) ||
|
|
dd->ipath_lid == 0) {
|
|
/*
|
|
* lid check is for when sma hasn't yet configured
|
|
*/
|
|
ret = -ENETDOWN;
|
|
ipath_cdbg(VERBOSE, "send while not ready, "
|
|
"mylid=%u, flags=0x%x\n",
|
|
dd->ipath_lid, dd->ipath_flags);
|
|
}
|
|
|
|
vlsllnh = *((__be16 *) hdr);
|
|
if (vlsllnh != htons(IPATH_LRH_BTH)) {
|
|
ipath_dbg("Warning: lrh[0] wrong (%x, not %x); "
|
|
"not sending\n", be16_to_cpu(vlsllnh),
|
|
IPATH_LRH_BTH);
|
|
ret = -EINVAL;
|
|
}
|
|
if (ret)
|
|
goto done;
|
|
|
|
/* Get a PIO buffer to use. */
|
|
piobuf = ipath_getpiobuf(dd, NULL);
|
|
if (piobuf == NULL) {
|
|
ret = -EBUSY;
|
|
goto done;
|
|
}
|
|
|
|
plen = (sizeof(*hdr) >> 2); /* actual length */
|
|
ipath_cdbg(EPKT, "0x%x+1w pio %p\n", plen, piobuf);
|
|
|
|
writeq(plen+1, piobuf); /* len (+1 for pad) to pbc, no flags */
|
|
ipath_flush_wc();
|
|
piobuf += 2;
|
|
uhdr = (u32 *)hdr;
|
|
count = plen-1; /* amount we can copy before trigger word */
|
|
__iowrite32_copy(piobuf, uhdr, count);
|
|
ipath_flush_wc();
|
|
__raw_writel(uhdr[count], piobuf + count);
|
|
ipath_flush_wc(); /* ensure it's sent, now */
|
|
|
|
ipath_stats.sps_ether_spkts++; /* ether packet sent */
|
|
|
|
done:
|
|
return ret;
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ipath_layer_send_hdr);
|
|
|
|
int ipath_layer_set_piointbufavail_int(struct ipath_devdata *dd)
|
|
{
|
|
set_bit(IPATH_S_PIOINTBUFAVAIL, &dd->ipath_sendctrl);
|
|
|
|
ipath_write_kreg(dd, dd->ipath_kregs->kr_sendctrl,
|
|
dd->ipath_sendctrl);
|
|
return 0;
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ipath_layer_set_piointbufavail_int);
|