mirror of
https://github.com/xemu-project/xemu.git
synced 2024-11-30 15:00:34 +00:00
2068cabd3f
Stop including cpu.h in files that don't need it. Signed-off-by: Thomas Huth <thuth@redhat.com> Message-Id: <20210416171314.2074665-4-thuth@redhat.com> Signed-off-by: Laurent Vivier <laurent@vivier.eu>
886 lines
27 KiB
C
886 lines
27 KiB
C
/*
|
|
* QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator
|
|
*
|
|
* PAPR Inter-VM Logical Lan, aka ibmveth
|
|
*
|
|
* Copyright (c) 2010,2011 David Gibson, IBM Corporation.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
* of this software and associated documentation files (the "Software"), to deal
|
|
* in the Software without restriction, including without limitation the rights
|
|
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
* copies of the Software, and to permit persons to whom the Software is
|
|
* furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
|
* THE SOFTWARE.
|
|
*
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "qemu/log.h"
|
|
#include "qemu/module.h"
|
|
#include "net/net.h"
|
|
#include "migration/vmstate.h"
|
|
#include "hw/ppc/spapr.h"
|
|
#include "hw/ppc/spapr_vio.h"
|
|
#include "hw/qdev-properties.h"
|
|
#include "sysemu/sysemu.h"
|
|
#include "trace.h"
|
|
|
|
#include <libfdt.h>
|
|
#include "qom/object.h"
|
|
|
|
#define ETH_ALEN 6
|
|
#define MAX_PACKET_SIZE 65536
|
|
|
|
/* Compatibility flags for migration */
|
|
#define SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT 0
|
|
#define SPAPRVLAN_FLAG_RX_BUF_POOLS (1 << SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT)
|
|
|
|
/*
|
|
* Virtual LAN device
|
|
*/
|
|
|
|
typedef uint64_t vlan_bd_t;
|
|
|
|
#define VLAN_BD_VALID 0x8000000000000000ULL
|
|
#define VLAN_BD_TOGGLE 0x4000000000000000ULL
|
|
#define VLAN_BD_NO_CSUM 0x0200000000000000ULL
|
|
#define VLAN_BD_CSUM_GOOD 0x0100000000000000ULL
|
|
#define VLAN_BD_LEN_MASK 0x00ffffff00000000ULL
|
|
#define VLAN_BD_LEN(bd) (((bd) & VLAN_BD_LEN_MASK) >> 32)
|
|
#define VLAN_BD_ADDR_MASK 0x00000000ffffffffULL
|
|
#define VLAN_BD_ADDR(bd) ((bd) & VLAN_BD_ADDR_MASK)
|
|
|
|
#define VLAN_VALID_BD(addr, len) (VLAN_BD_VALID | \
|
|
(((len) << 32) & VLAN_BD_LEN_MASK) | \
|
|
(addr & VLAN_BD_ADDR_MASK))
|
|
|
|
#define VLAN_RXQC_TOGGLE 0x80
|
|
#define VLAN_RXQC_VALID 0x40
|
|
#define VLAN_RXQC_NO_CSUM 0x02
|
|
#define VLAN_RXQC_CSUM_GOOD 0x01
|
|
|
|
#define VLAN_RQ_ALIGNMENT 16
|
|
#define VLAN_RXQ_BD_OFF 0
|
|
#define VLAN_FILTER_BD_OFF 8
|
|
#define VLAN_RX_BDS_OFF 16
|
|
/*
|
|
* The final 8 bytes of the buffer list is a counter of frames dropped
|
|
* because there was not a buffer in the buffer list capable of holding
|
|
* the frame. We must avoid it, or the operating system will report garbage
|
|
* for this statistic.
|
|
*/
|
|
#define VLAN_RX_BDS_LEN (SPAPR_TCE_PAGE_SIZE - VLAN_RX_BDS_OFF - 8)
|
|
#define VLAN_MAX_BUFS (VLAN_RX_BDS_LEN / 8)
|
|
|
|
#define TYPE_VIO_SPAPR_VLAN_DEVICE "spapr-vlan"
|
|
OBJECT_DECLARE_SIMPLE_TYPE(SpaprVioVlan, VIO_SPAPR_VLAN_DEVICE)
|
|
|
|
#define RX_POOL_MAX_BDS 4096
|
|
#define RX_MAX_POOLS 5
|
|
|
|
typedef struct {
|
|
int32_t bufsize;
|
|
int32_t count;
|
|
vlan_bd_t bds[RX_POOL_MAX_BDS];
|
|
} RxBufPool;
|
|
|
|
struct SpaprVioVlan {
|
|
SpaprVioDevice sdev;
|
|
NICConf nicconf;
|
|
NICState *nic;
|
|
MACAddr perm_mac;
|
|
bool isopen;
|
|
hwaddr buf_list;
|
|
uint32_t add_buf_ptr, use_buf_ptr, rx_bufs;
|
|
hwaddr rxq_ptr;
|
|
QEMUTimer *rxp_timer;
|
|
uint32_t compat_flags; /* Compatibility flags for migration */
|
|
RxBufPool *rx_pool[RX_MAX_POOLS]; /* Receive buffer descriptor pools */
|
|
};
|
|
|
|
static bool spapr_vlan_can_receive(NetClientState *nc)
|
|
{
|
|
SpaprVioVlan *dev = qemu_get_nic_opaque(nc);
|
|
|
|
return dev->isopen && dev->rx_bufs > 0;
|
|
}
|
|
|
|
/**
|
|
* The last 8 bytes of the receive buffer list page (that has been
|
|
* supplied by the guest with the H_REGISTER_LOGICAL_LAN call) contain
|
|
* a counter for frames that have been dropped because there was no
|
|
* suitable receive buffer available. This function is used to increase
|
|
* this counter by one.
|
|
*/
|
|
static void spapr_vlan_record_dropped_rx_frame(SpaprVioVlan *dev)
|
|
{
|
|
uint64_t cnt;
|
|
|
|
cnt = vio_ldq(&dev->sdev, dev->buf_list + 4096 - 8);
|
|
vio_stq(&dev->sdev, dev->buf_list + 4096 - 8, cnt + 1);
|
|
}
|
|
|
|
/**
|
|
* Get buffer descriptor from one of our receive buffer pools
|
|
*/
|
|
static vlan_bd_t spapr_vlan_get_rx_bd_from_pool(SpaprVioVlan *dev,
|
|
size_t size)
|
|
{
|
|
vlan_bd_t bd;
|
|
int pool;
|
|
|
|
for (pool = 0; pool < RX_MAX_POOLS; pool++) {
|
|
if (dev->rx_pool[pool]->count > 0 &&
|
|
dev->rx_pool[pool]->bufsize >= size + 8) {
|
|
break;
|
|
}
|
|
}
|
|
if (pool == RX_MAX_POOLS) {
|
|
/* Failed to find a suitable buffer */
|
|
return 0;
|
|
}
|
|
|
|
|
|
trace_spapr_vlan_get_rx_bd_from_pool_found(pool,
|
|
dev->rx_pool[pool]->count,
|
|
dev->rx_bufs);
|
|
|
|
/* Remove the buffer from the pool */
|
|
dev->rx_pool[pool]->count--;
|
|
bd = dev->rx_pool[pool]->bds[dev->rx_pool[pool]->count];
|
|
dev->rx_pool[pool]->bds[dev->rx_pool[pool]->count] = 0;
|
|
|
|
return bd;
|
|
}
|
|
|
|
/**
|
|
* Get buffer descriptor from the receive buffer list page that has been
|
|
* supplied by the guest with the H_REGISTER_LOGICAL_LAN call
|
|
*/
|
|
static vlan_bd_t spapr_vlan_get_rx_bd_from_page(SpaprVioVlan *dev,
|
|
size_t size)
|
|
{
|
|
int buf_ptr = dev->use_buf_ptr;
|
|
vlan_bd_t bd;
|
|
|
|
do {
|
|
buf_ptr += 8;
|
|
if (buf_ptr >= VLAN_RX_BDS_LEN + VLAN_RX_BDS_OFF) {
|
|
buf_ptr = VLAN_RX_BDS_OFF;
|
|
}
|
|
|
|
bd = vio_ldq(&dev->sdev, dev->buf_list + buf_ptr);
|
|
|
|
trace_spapr_vlan_get_rx_bd_from_page(buf_ptr, (uint64_t)bd);
|
|
} while ((!(bd & VLAN_BD_VALID) || VLAN_BD_LEN(bd) < size + 8)
|
|
&& buf_ptr != dev->use_buf_ptr);
|
|
|
|
if (!(bd & VLAN_BD_VALID) || VLAN_BD_LEN(bd) < size + 8) {
|
|
/* Failed to find a suitable buffer */
|
|
return 0;
|
|
}
|
|
|
|
/* Remove the buffer from the pool */
|
|
dev->use_buf_ptr = buf_ptr;
|
|
vio_stq(&dev->sdev, dev->buf_list + dev->use_buf_ptr, 0);
|
|
|
|
trace_spapr_vlan_get_rx_bd_from_page_found(dev->use_buf_ptr, dev->rx_bufs);
|
|
|
|
return bd;
|
|
}
|
|
|
|
static ssize_t spapr_vlan_receive(NetClientState *nc, const uint8_t *buf,
|
|
size_t size)
|
|
{
|
|
SpaprVioVlan *dev = qemu_get_nic_opaque(nc);
|
|
SpaprVioDevice *sdev = VIO_SPAPR_DEVICE(dev);
|
|
vlan_bd_t rxq_bd = vio_ldq(sdev, dev->buf_list + VLAN_RXQ_BD_OFF);
|
|
vlan_bd_t bd;
|
|
uint64_t handle;
|
|
uint8_t control;
|
|
|
|
trace_spapr_vlan_receive(sdev->qdev.id, dev->rx_bufs);
|
|
|
|
if (!dev->isopen) {
|
|
return -1;
|
|
}
|
|
|
|
if (!dev->rx_bufs) {
|
|
spapr_vlan_record_dropped_rx_frame(dev);
|
|
return 0;
|
|
}
|
|
|
|
if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
|
|
bd = spapr_vlan_get_rx_bd_from_pool(dev, size);
|
|
} else {
|
|
bd = spapr_vlan_get_rx_bd_from_page(dev, size);
|
|
}
|
|
if (!bd) {
|
|
spapr_vlan_record_dropped_rx_frame(dev);
|
|
return 0;
|
|
}
|
|
|
|
dev->rx_bufs--;
|
|
|
|
/* Transfer the packet data */
|
|
if (spapr_vio_dma_write(sdev, VLAN_BD_ADDR(bd) + 8, buf, size) < 0) {
|
|
return -1;
|
|
}
|
|
|
|
trace_spapr_vlan_receive_dma_completed();
|
|
|
|
/* Update the receive queue */
|
|
control = VLAN_RXQC_TOGGLE | VLAN_RXQC_VALID;
|
|
if (rxq_bd & VLAN_BD_TOGGLE) {
|
|
control ^= VLAN_RXQC_TOGGLE;
|
|
}
|
|
|
|
handle = vio_ldq(sdev, VLAN_BD_ADDR(bd));
|
|
vio_stq(sdev, VLAN_BD_ADDR(rxq_bd) + dev->rxq_ptr + 8, handle);
|
|
vio_stl(sdev, VLAN_BD_ADDR(rxq_bd) + dev->rxq_ptr + 4, size);
|
|
vio_sth(sdev, VLAN_BD_ADDR(rxq_bd) + dev->rxq_ptr + 2, 8);
|
|
vio_stb(sdev, VLAN_BD_ADDR(rxq_bd) + dev->rxq_ptr, control);
|
|
|
|
trace_spapr_vlan_receive_wrote(dev->rxq_ptr,
|
|
vio_ldq(sdev, VLAN_BD_ADDR(rxq_bd) +
|
|
dev->rxq_ptr),
|
|
vio_ldq(sdev, VLAN_BD_ADDR(rxq_bd) +
|
|
dev->rxq_ptr + 8));
|
|
|
|
dev->rxq_ptr += 16;
|
|
if (dev->rxq_ptr >= VLAN_BD_LEN(rxq_bd)) {
|
|
dev->rxq_ptr = 0;
|
|
vio_stq(sdev, dev->buf_list + VLAN_RXQ_BD_OFF, rxq_bd ^ VLAN_BD_TOGGLE);
|
|
}
|
|
|
|
if (sdev->signal_state & 1) {
|
|
spapr_vio_irq_pulse(sdev);
|
|
}
|
|
|
|
return size;
|
|
}
|
|
|
|
static NetClientInfo net_spapr_vlan_info = {
|
|
.type = NET_CLIENT_DRIVER_NIC,
|
|
.size = sizeof(NICState),
|
|
.can_receive = spapr_vlan_can_receive,
|
|
.receive = spapr_vlan_receive,
|
|
};
|
|
|
|
static void spapr_vlan_flush_rx_queue(void *opaque)
|
|
{
|
|
SpaprVioVlan *dev = opaque;
|
|
|
|
qemu_flush_queued_packets(qemu_get_queue(dev->nic));
|
|
}
|
|
|
|
static void spapr_vlan_reset_rx_pool(RxBufPool *rxp)
|
|
{
|
|
/*
|
|
* Use INT_MAX as bufsize so that unused buffers are moved to the end
|
|
* of the list during the qsort in spapr_vlan_add_rxbuf_to_pool() later.
|
|
*/
|
|
rxp->bufsize = INT_MAX;
|
|
rxp->count = 0;
|
|
memset(rxp->bds, 0, sizeof(rxp->bds));
|
|
}
|
|
|
|
static void spapr_vlan_reset(SpaprVioDevice *sdev)
|
|
{
|
|
SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
|
|
int i;
|
|
|
|
dev->buf_list = 0;
|
|
dev->rx_bufs = 0;
|
|
dev->isopen = 0;
|
|
|
|
if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
|
|
for (i = 0; i < RX_MAX_POOLS; i++) {
|
|
spapr_vlan_reset_rx_pool(dev->rx_pool[i]);
|
|
}
|
|
}
|
|
|
|
memcpy(&dev->nicconf.macaddr.a, &dev->perm_mac.a,
|
|
sizeof(dev->nicconf.macaddr.a));
|
|
qemu_format_nic_info_str(qemu_get_queue(dev->nic), dev->nicconf.macaddr.a);
|
|
}
|
|
|
|
static void spapr_vlan_realize(SpaprVioDevice *sdev, Error **errp)
|
|
{
|
|
SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
|
|
|
|
qemu_macaddr_default_if_unset(&dev->nicconf.macaddr);
|
|
|
|
memcpy(&dev->perm_mac.a, &dev->nicconf.macaddr.a, sizeof(dev->perm_mac.a));
|
|
|
|
dev->nic = qemu_new_nic(&net_spapr_vlan_info, &dev->nicconf,
|
|
object_get_typename(OBJECT(sdev)), sdev->qdev.id, dev);
|
|
qemu_format_nic_info_str(qemu_get_queue(dev->nic), dev->nicconf.macaddr.a);
|
|
|
|
dev->rxp_timer = timer_new_us(QEMU_CLOCK_VIRTUAL, spapr_vlan_flush_rx_queue,
|
|
dev);
|
|
}
|
|
|
|
static void spapr_vlan_instance_init(Object *obj)
|
|
{
|
|
SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(obj);
|
|
int i;
|
|
|
|
device_add_bootindex_property(obj, &dev->nicconf.bootindex,
|
|
"bootindex", "",
|
|
DEVICE(dev));
|
|
|
|
if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
|
|
for (i = 0; i < RX_MAX_POOLS; i++) {
|
|
dev->rx_pool[i] = g_new(RxBufPool, 1);
|
|
spapr_vlan_reset_rx_pool(dev->rx_pool[i]);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void spapr_vlan_instance_finalize(Object *obj)
|
|
{
|
|
SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(obj);
|
|
int i;
|
|
|
|
if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
|
|
for (i = 0; i < RX_MAX_POOLS; i++) {
|
|
g_free(dev->rx_pool[i]);
|
|
dev->rx_pool[i] = NULL;
|
|
}
|
|
}
|
|
|
|
if (dev->rxp_timer) {
|
|
timer_free(dev->rxp_timer);
|
|
}
|
|
}
|
|
|
|
void spapr_vlan_create(SpaprVioBus *bus, NICInfo *nd)
|
|
{
|
|
DeviceState *dev;
|
|
|
|
dev = qdev_new("spapr-vlan");
|
|
|
|
qdev_set_nic_properties(dev, nd);
|
|
|
|
qdev_realize_and_unref(dev, &bus->bus, &error_fatal);
|
|
}
|
|
|
|
static int spapr_vlan_devnode(SpaprVioDevice *dev, void *fdt, int node_off)
|
|
{
|
|
SpaprVioVlan *vdev = VIO_SPAPR_VLAN_DEVICE(dev);
|
|
uint8_t padded_mac[8] = {0, 0};
|
|
int ret;
|
|
|
|
/* Some old phyp versions give the mac address in an 8-byte
|
|
* property. The kernel driver (before 3.10) has an insane workaround;
|
|
* rather than doing the obvious thing and checking the property
|
|
* length, it checks whether the first byte has 0b10 in the low
|
|
* bits. If a correct 6-byte property has a different first byte
|
|
* the kernel will get the wrong mac address, overrunning its
|
|
* buffer in the process (read only, thank goodness).
|
|
*
|
|
* Here we return a 6-byte address unless that would break a pre-3.10
|
|
* driver. In that case we return a padded 8-byte address to allow the old
|
|
* workaround to succeed. */
|
|
if ((vdev->nicconf.macaddr.a[0] & 0x3) == 0x2) {
|
|
ret = fdt_setprop(fdt, node_off, "local-mac-address",
|
|
&vdev->nicconf.macaddr, ETH_ALEN);
|
|
} else {
|
|
memcpy(&padded_mac[2], &vdev->nicconf.macaddr, ETH_ALEN);
|
|
ret = fdt_setprop(fdt, node_off, "local-mac-address",
|
|
padded_mac, sizeof(padded_mac));
|
|
}
|
|
if (ret < 0) {
|
|
return ret;
|
|
}
|
|
|
|
ret = fdt_setprop_cell(fdt, node_off, "ibm,mac-address-filters", 0);
|
|
if (ret < 0) {
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int check_bd(SpaprVioVlan *dev, vlan_bd_t bd,
|
|
target_ulong alignment)
|
|
{
|
|
if ((VLAN_BD_ADDR(bd) % alignment)
|
|
|| (VLAN_BD_LEN(bd) % alignment)) {
|
|
return -1;
|
|
}
|
|
|
|
if (!spapr_vio_dma_valid(&dev->sdev, VLAN_BD_ADDR(bd),
|
|
VLAN_BD_LEN(bd), DMA_DIRECTION_FROM_DEVICE)
|
|
|| !spapr_vio_dma_valid(&dev->sdev, VLAN_BD_ADDR(bd),
|
|
VLAN_BD_LEN(bd), DMA_DIRECTION_TO_DEVICE)) {
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static target_ulong h_register_logical_lan(PowerPCCPU *cpu,
|
|
SpaprMachineState *spapr,
|
|
target_ulong opcode,
|
|
target_ulong *args)
|
|
{
|
|
target_ulong reg = args[0];
|
|
target_ulong buf_list = args[1];
|
|
target_ulong rec_queue = args[2];
|
|
target_ulong filter_list = args[3];
|
|
SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
|
|
SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
|
|
vlan_bd_t filter_list_bd;
|
|
|
|
if (!dev) {
|
|
return H_PARAMETER;
|
|
}
|
|
|
|
if (dev->isopen) {
|
|
hcall_dprintf("H_REGISTER_LOGICAL_LAN called twice without "
|
|
"H_FREE_LOGICAL_LAN\n");
|
|
return H_RESOURCE;
|
|
}
|
|
|
|
if (check_bd(dev, VLAN_VALID_BD(buf_list, SPAPR_TCE_PAGE_SIZE),
|
|
SPAPR_TCE_PAGE_SIZE) < 0) {
|
|
hcall_dprintf("Bad buf_list 0x" TARGET_FMT_lx "\n", buf_list);
|
|
return H_PARAMETER;
|
|
}
|
|
|
|
filter_list_bd = VLAN_VALID_BD(filter_list, SPAPR_TCE_PAGE_SIZE);
|
|
if (check_bd(dev, filter_list_bd, SPAPR_TCE_PAGE_SIZE) < 0) {
|
|
hcall_dprintf("Bad filter_list 0x" TARGET_FMT_lx "\n", filter_list);
|
|
return H_PARAMETER;
|
|
}
|
|
|
|
if (!(rec_queue & VLAN_BD_VALID)
|
|
|| (check_bd(dev, rec_queue, VLAN_RQ_ALIGNMENT) < 0)) {
|
|
hcall_dprintf("Bad receive queue\n");
|
|
return H_PARAMETER;
|
|
}
|
|
|
|
dev->buf_list = buf_list;
|
|
sdev->signal_state = 0;
|
|
|
|
rec_queue &= ~VLAN_BD_TOGGLE;
|
|
|
|
/* Initialize the buffer list */
|
|
vio_stq(sdev, buf_list, rec_queue);
|
|
vio_stq(sdev, buf_list + 8, filter_list_bd);
|
|
spapr_vio_dma_set(sdev, buf_list + VLAN_RX_BDS_OFF, 0,
|
|
SPAPR_TCE_PAGE_SIZE - VLAN_RX_BDS_OFF);
|
|
dev->add_buf_ptr = VLAN_RX_BDS_OFF - 8;
|
|
dev->use_buf_ptr = VLAN_RX_BDS_OFF - 8;
|
|
dev->rx_bufs = 0;
|
|
dev->rxq_ptr = 0;
|
|
|
|
/* Initialize the receive queue */
|
|
spapr_vio_dma_set(sdev, VLAN_BD_ADDR(rec_queue), 0, VLAN_BD_LEN(rec_queue));
|
|
|
|
dev->isopen = 1;
|
|
qemu_flush_queued_packets(qemu_get_queue(dev->nic));
|
|
|
|
return H_SUCCESS;
|
|
}
|
|
|
|
|
|
static target_ulong h_free_logical_lan(PowerPCCPU *cpu,
|
|
SpaprMachineState *spapr,
|
|
target_ulong opcode, target_ulong *args)
|
|
{
|
|
target_ulong reg = args[0];
|
|
SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
|
|
SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
|
|
|
|
if (!dev) {
|
|
return H_PARAMETER;
|
|
}
|
|
|
|
if (!dev->isopen) {
|
|
hcall_dprintf("H_FREE_LOGICAL_LAN called without "
|
|
"H_REGISTER_LOGICAL_LAN\n");
|
|
return H_RESOURCE;
|
|
}
|
|
|
|
spapr_vlan_reset(sdev);
|
|
return H_SUCCESS;
|
|
}
|
|
|
|
/**
|
|
* Used for qsort, this function compares two RxBufPools by size.
|
|
*/
|
|
static int rx_pool_size_compare(const void *p1, const void *p2)
|
|
{
|
|
const RxBufPool *pool1 = *(RxBufPool **)p1;
|
|
const RxBufPool *pool2 = *(RxBufPool **)p2;
|
|
|
|
if (pool1->bufsize < pool2->bufsize) {
|
|
return -1;
|
|
}
|
|
return pool1->bufsize > pool2->bufsize;
|
|
}
|
|
|
|
/**
|
|
* Search for a matching buffer pool with exact matching size,
|
|
* or return -1 if no matching pool has been found.
|
|
*/
|
|
static int spapr_vlan_get_rx_pool_id(SpaprVioVlan *dev, int size)
|
|
{
|
|
int pool;
|
|
|
|
for (pool = 0; pool < RX_MAX_POOLS; pool++) {
|
|
if (dev->rx_pool[pool]->bufsize == size) {
|
|
return pool;
|
|
}
|
|
}
|
|
|
|
return -1;
|
|
}
|
|
|
|
/**
|
|
* Enqueuing receive buffer by adding it to one of our receive buffer pools
|
|
*/
|
|
static target_long spapr_vlan_add_rxbuf_to_pool(SpaprVioVlan *dev,
|
|
target_ulong buf)
|
|
{
|
|
int size = VLAN_BD_LEN(buf);
|
|
int pool;
|
|
|
|
pool = spapr_vlan_get_rx_pool_id(dev, size);
|
|
if (pool < 0) {
|
|
/*
|
|
* No matching pool found? Try to use a new one. If the guest used all
|
|
* pools before, but changed the size of one pool in the meantime, we might
|
|
* need to recycle that pool here (if it's empty already). Thus scan
|
|
* all buffer pools now, starting with the last (likely empty) one.
|
|
*/
|
|
for (pool = RX_MAX_POOLS - 1; pool >= 0 ; pool--) {
|
|
if (dev->rx_pool[pool]->count == 0) {
|
|
dev->rx_pool[pool]->bufsize = size;
|
|
/*
|
|
* Sort pools by size so that spapr_vlan_receive()
|
|
* can later find the smallest buffer pool easily.
|
|
*/
|
|
qsort(dev->rx_pool, RX_MAX_POOLS, sizeof(dev->rx_pool[0]),
|
|
rx_pool_size_compare);
|
|
pool = spapr_vlan_get_rx_pool_id(dev, size);
|
|
trace_spapr_vlan_add_rxbuf_to_pool_create(pool,
|
|
VLAN_BD_LEN(buf));
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
/* Still no usable pool? Give up */
|
|
if (pool < 0 || dev->rx_pool[pool]->count >= RX_POOL_MAX_BDS) {
|
|
return H_RESOURCE;
|
|
}
|
|
|
|
trace_spapr_vlan_add_rxbuf_to_pool(pool, VLAN_BD_LEN(buf),
|
|
dev->rx_pool[pool]->count);
|
|
|
|
dev->rx_pool[pool]->bds[dev->rx_pool[pool]->count++] = buf;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* This is the old way of enqueuing receive buffers: Add it to the rx queue
|
|
* page that has been supplied by the guest (which is quite limited in size).
|
|
*/
|
|
static target_long spapr_vlan_add_rxbuf_to_page(SpaprVioVlan *dev,
|
|
target_ulong buf)
|
|
{
|
|
vlan_bd_t bd;
|
|
|
|
if (dev->rx_bufs >= VLAN_MAX_BUFS) {
|
|
return H_RESOURCE;
|
|
}
|
|
|
|
do {
|
|
dev->add_buf_ptr += 8;
|
|
if (dev->add_buf_ptr >= VLAN_RX_BDS_LEN + VLAN_RX_BDS_OFF) {
|
|
dev->add_buf_ptr = VLAN_RX_BDS_OFF;
|
|
}
|
|
|
|
bd = vio_ldq(&dev->sdev, dev->buf_list + dev->add_buf_ptr);
|
|
} while (bd & VLAN_BD_VALID);
|
|
|
|
vio_stq(&dev->sdev, dev->buf_list + dev->add_buf_ptr, buf);
|
|
|
|
trace_spapr_vlan_add_rxbuf_to_page(dev->add_buf_ptr, dev->rx_bufs, buf);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static target_ulong h_add_logical_lan_buffer(PowerPCCPU *cpu,
|
|
SpaprMachineState *spapr,
|
|
target_ulong opcode,
|
|
target_ulong *args)
|
|
{
|
|
target_ulong reg = args[0];
|
|
target_ulong buf = args[1];
|
|
SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
|
|
SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
|
|
target_long ret;
|
|
|
|
trace_spapr_vlan_h_add_logical_lan_buffer(reg, buf);
|
|
|
|
if (!sdev) {
|
|
hcall_dprintf("Bad device\n");
|
|
return H_PARAMETER;
|
|
}
|
|
|
|
if ((check_bd(dev, buf, 4) < 0)
|
|
|| (VLAN_BD_LEN(buf) < 16)) {
|
|
hcall_dprintf("Bad buffer enqueued\n");
|
|
return H_PARAMETER;
|
|
}
|
|
|
|
if (!dev->isopen) {
|
|
return H_RESOURCE;
|
|
}
|
|
|
|
if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
|
|
ret = spapr_vlan_add_rxbuf_to_pool(dev, buf);
|
|
} else {
|
|
ret = spapr_vlan_add_rxbuf_to_page(dev, buf);
|
|
}
|
|
if (ret) {
|
|
return ret;
|
|
}
|
|
|
|
dev->rx_bufs++;
|
|
|
|
/*
|
|
* Give guest some more time to add additional RX buffers before we
|
|
* flush the receive queue, so that e.g. fragmented IP packets can
|
|
* be passed to the guest in one go later (instead of passing single
|
|
* fragments if there is only one receive buffer available).
|
|
*/
|
|
timer_mod(dev->rxp_timer, qemu_clock_get_us(QEMU_CLOCK_VIRTUAL) + 500);
|
|
|
|
return H_SUCCESS;
|
|
}
|
|
|
|
static target_ulong h_send_logical_lan(PowerPCCPU *cpu,
|
|
SpaprMachineState *spapr,
|
|
target_ulong opcode, target_ulong *args)
|
|
{
|
|
target_ulong reg = args[0];
|
|
target_ulong *bufs = args + 1;
|
|
target_ulong continue_token = args[7];
|
|
SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
|
|
SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
|
|
unsigned total_len;
|
|
uint8_t *p;
|
|
g_autofree uint8_t *lbuf = NULL;
|
|
int i, nbufs;
|
|
int ret;
|
|
|
|
trace_spapr_vlan_h_send_logical_lan(reg, continue_token);
|
|
|
|
if (!sdev) {
|
|
return H_PARAMETER;
|
|
}
|
|
|
|
trace_spapr_vlan_h_send_logical_lan_rxbufs(dev->rx_bufs);
|
|
|
|
if (!dev->isopen) {
|
|
return H_DROPPED;
|
|
}
|
|
|
|
if (continue_token) {
|
|
return H_HARDWARE; /* FIXME actually handle this */
|
|
}
|
|
|
|
total_len = 0;
|
|
for (i = 0; i < 6; i++) {
|
|
trace_spapr_vlan_h_send_logical_lan_buf_desc(bufs[i]);
|
|
if (!(bufs[i] & VLAN_BD_VALID)) {
|
|
break;
|
|
}
|
|
total_len += VLAN_BD_LEN(bufs[i]);
|
|
}
|
|
|
|
nbufs = i;
|
|
trace_spapr_vlan_h_send_logical_lan_total(nbufs, total_len);
|
|
|
|
if (total_len == 0) {
|
|
return H_SUCCESS;
|
|
}
|
|
|
|
if (total_len > MAX_PACKET_SIZE) {
|
|
/* Don't let the guest force too large an allocation */
|
|
return H_RESOURCE;
|
|
}
|
|
|
|
lbuf = g_malloc(total_len);
|
|
p = lbuf;
|
|
for (i = 0; i < nbufs; i++) {
|
|
ret = spapr_vio_dma_read(sdev, VLAN_BD_ADDR(bufs[i]),
|
|
p, VLAN_BD_LEN(bufs[i]));
|
|
if (ret < 0) {
|
|
return ret;
|
|
}
|
|
|
|
p += VLAN_BD_LEN(bufs[i]);
|
|
}
|
|
|
|
qemu_send_packet(qemu_get_queue(dev->nic), lbuf, total_len);
|
|
|
|
return H_SUCCESS;
|
|
}
|
|
|
|
static target_ulong h_multicast_ctrl(PowerPCCPU *cpu, SpaprMachineState *spapr,
|
|
target_ulong opcode, target_ulong *args)
|
|
{
|
|
target_ulong reg = args[0];
|
|
SpaprVioDevice *dev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
|
|
|
|
if (!dev) {
|
|
return H_PARAMETER;
|
|
}
|
|
|
|
return H_SUCCESS;
|
|
}
|
|
|
|
static target_ulong h_change_logical_lan_mac(PowerPCCPU *cpu,
|
|
SpaprMachineState *spapr,
|
|
target_ulong opcode,
|
|
target_ulong *args)
|
|
{
|
|
target_ulong reg = args[0];
|
|
target_ulong macaddr = args[1];
|
|
SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
|
|
SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
|
|
int i;
|
|
|
|
for (i = 0; i < ETH_ALEN; i++) {
|
|
dev->nicconf.macaddr.a[ETH_ALEN - i - 1] = macaddr & 0xff;
|
|
macaddr >>= 8;
|
|
}
|
|
|
|
qemu_format_nic_info_str(qemu_get_queue(dev->nic), dev->nicconf.macaddr.a);
|
|
|
|
return H_SUCCESS;
|
|
}
|
|
|
|
static Property spapr_vlan_properties[] = {
|
|
DEFINE_SPAPR_PROPERTIES(SpaprVioVlan, sdev),
|
|
DEFINE_NIC_PROPERTIES(SpaprVioVlan, nicconf),
|
|
DEFINE_PROP_BIT("use-rx-buffer-pools", SpaprVioVlan,
|
|
compat_flags, SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT, true),
|
|
DEFINE_PROP_END_OF_LIST(),
|
|
};
|
|
|
|
static bool spapr_vlan_rx_buffer_pools_needed(void *opaque)
|
|
{
|
|
SpaprVioVlan *dev = opaque;
|
|
|
|
return (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) != 0;
|
|
}
|
|
|
|
static const VMStateDescription vmstate_rx_buffer_pool = {
|
|
.name = "spapr_llan/rx_buffer_pool",
|
|
.version_id = 1,
|
|
.minimum_version_id = 1,
|
|
.needed = spapr_vlan_rx_buffer_pools_needed,
|
|
.fields = (VMStateField[]) {
|
|
VMSTATE_INT32(bufsize, RxBufPool),
|
|
VMSTATE_INT32(count, RxBufPool),
|
|
VMSTATE_UINT64_ARRAY(bds, RxBufPool, RX_POOL_MAX_BDS),
|
|
VMSTATE_END_OF_LIST()
|
|
}
|
|
};
|
|
|
|
static const VMStateDescription vmstate_rx_pools = {
|
|
.name = "spapr_llan/rx_pools",
|
|
.version_id = 1,
|
|
.minimum_version_id = 1,
|
|
.needed = spapr_vlan_rx_buffer_pools_needed,
|
|
.fields = (VMStateField[]) {
|
|
VMSTATE_ARRAY_OF_POINTER_TO_STRUCT(rx_pool, SpaprVioVlan,
|
|
RX_MAX_POOLS, 1,
|
|
vmstate_rx_buffer_pool, RxBufPool),
|
|
VMSTATE_END_OF_LIST()
|
|
}
|
|
};
|
|
|
|
static const VMStateDescription vmstate_spapr_llan = {
|
|
.name = "spapr_llan",
|
|
.version_id = 1,
|
|
.minimum_version_id = 1,
|
|
.fields = (VMStateField[]) {
|
|
VMSTATE_SPAPR_VIO(sdev, SpaprVioVlan),
|
|
/* LLAN state */
|
|
VMSTATE_BOOL(isopen, SpaprVioVlan),
|
|
VMSTATE_UINT64(buf_list, SpaprVioVlan),
|
|
VMSTATE_UINT32(add_buf_ptr, SpaprVioVlan),
|
|
VMSTATE_UINT32(use_buf_ptr, SpaprVioVlan),
|
|
VMSTATE_UINT32(rx_bufs, SpaprVioVlan),
|
|
VMSTATE_UINT64(rxq_ptr, SpaprVioVlan),
|
|
|
|
VMSTATE_END_OF_LIST()
|
|
},
|
|
.subsections = (const VMStateDescription * []) {
|
|
&vmstate_rx_pools,
|
|
NULL
|
|
}
|
|
};
|
|
|
|
static void spapr_vlan_class_init(ObjectClass *klass, void *data)
|
|
{
|
|
DeviceClass *dc = DEVICE_CLASS(klass);
|
|
SpaprVioDeviceClass *k = VIO_SPAPR_DEVICE_CLASS(klass);
|
|
|
|
k->realize = spapr_vlan_realize;
|
|
k->reset = spapr_vlan_reset;
|
|
k->devnode = spapr_vlan_devnode;
|
|
k->dt_name = "l-lan";
|
|
k->dt_type = "network";
|
|
k->dt_compatible = "IBM,l-lan";
|
|
k->signal_mask = 0x1;
|
|
set_bit(DEVICE_CATEGORY_NETWORK, dc->categories);
|
|
device_class_set_props(dc, spapr_vlan_properties);
|
|
k->rtce_window_size = 0x10000000;
|
|
dc->vmsd = &vmstate_spapr_llan;
|
|
}
|
|
|
|
static const TypeInfo spapr_vlan_info = {
|
|
.name = TYPE_VIO_SPAPR_VLAN_DEVICE,
|
|
.parent = TYPE_VIO_SPAPR_DEVICE,
|
|
.instance_size = sizeof(SpaprVioVlan),
|
|
.class_init = spapr_vlan_class_init,
|
|
.instance_init = spapr_vlan_instance_init,
|
|
.instance_finalize = spapr_vlan_instance_finalize,
|
|
};
|
|
|
|
static void spapr_vlan_register_types(void)
|
|
{
|
|
spapr_register_hypercall(H_REGISTER_LOGICAL_LAN, h_register_logical_lan);
|
|
spapr_register_hypercall(H_FREE_LOGICAL_LAN, h_free_logical_lan);
|
|
spapr_register_hypercall(H_SEND_LOGICAL_LAN, h_send_logical_lan);
|
|
spapr_register_hypercall(H_ADD_LOGICAL_LAN_BUFFER,
|
|
h_add_logical_lan_buffer);
|
|
spapr_register_hypercall(H_MULTICAST_CTRL, h_multicast_ctrl);
|
|
spapr_register_hypercall(H_CHANGE_LOGICAL_LAN_MAC,
|
|
h_change_logical_lan_mac);
|
|
type_register_static(&spapr_vlan_info);
|
|
}
|
|
|
|
type_init(spapr_vlan_register_types)
|