mirror of
https://github.com/FEX-Emu/linux.git
synced 2024-12-27 11:55:53 +00:00
0c401df37e
Push the cdev lock_kernel() call into cris drivers. Signed-off-by: Jonathan Corbet <corbet@lwn.net>
1544 lines
44 KiB
C
1544 lines
44 KiB
C
/*
|
|
* Simple synchronous serial port driver for ETRAX FS and Artpec-3.
|
|
*
|
|
* Copyright (c) 2005 Axis Communications AB
|
|
*
|
|
* Author: Mikael Starvik
|
|
*
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/types.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/major.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/smp_lock.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/poll.h>
|
|
#include <linux/init.h>
|
|
#include <linux/timer.h>
|
|
#include <linux/spinlock.h>
|
|
|
|
#include <asm/io.h>
|
|
#include <dma.h>
|
|
#include <pinmux.h>
|
|
#include <hwregs/reg_rdwr.h>
|
|
#include <hwregs/sser_defs.h>
|
|
#include <hwregs/dma_defs.h>
|
|
#include <hwregs/dma.h>
|
|
#include <hwregs/intr_vect_defs.h>
|
|
#include <hwregs/intr_vect.h>
|
|
#include <hwregs/reg_map.h>
|
|
#include <asm/sync_serial.h>
|
|
|
|
|
|
/* The receiver is a bit tricky beacuse of the continuous stream of data.*/
|
|
/* */
|
|
/* Three DMA descriptors are linked together. Each DMA descriptor is */
|
|
/* responsible for port->bufchunk of a common buffer. */
|
|
/* */
|
|
/* +---------------------------------------------+ */
|
|
/* | +----------+ +----------+ +----------+ | */
|
|
/* +-> | Descr[0] |-->| Descr[1] |-->| Descr[2] |-+ */
|
|
/* +----------+ +----------+ +----------+ */
|
|
/* | | | */
|
|
/* v v v */
|
|
/* +-------------------------------------+ */
|
|
/* | BUFFER | */
|
|
/* +-------------------------------------+ */
|
|
/* |<- data_avail ->| */
|
|
/* readp writep */
|
|
/* */
|
|
/* If the application keeps up the pace readp will be right after writep.*/
|
|
/* If the application can't keep the pace we have to throw away data. */
|
|
/* The idea is that readp should be ready with the data pointed out by */
|
|
/* Descr[i] when the DMA has filled in Descr[i+1]. */
|
|
/* Otherwise we will discard */
|
|
/* the rest of the data pointed out by Descr1 and set readp to the start */
|
|
/* of Descr2 */
|
|
|
|
#define SYNC_SERIAL_MAJOR 125
|
|
|
|
/* IN_BUFFER_SIZE should be a multiple of 6 to make sure that 24 bit */
|
|
/* words can be handled */
|
|
#define IN_BUFFER_SIZE 12288
|
|
#define IN_DESCR_SIZE 256
|
|
#define NBR_IN_DESCR (IN_BUFFER_SIZE/IN_DESCR_SIZE)
|
|
|
|
#define OUT_BUFFER_SIZE 1024*8
|
|
#define NBR_OUT_DESCR 8
|
|
|
|
#define DEFAULT_FRAME_RATE 0
|
|
#define DEFAULT_WORD_RATE 7
|
|
|
|
/* NOTE: Enabling some debug will likely cause overrun or underrun,
|
|
* especially if manual mode is use.
|
|
*/
|
|
#define DEBUG(x)
|
|
#define DEBUGREAD(x)
|
|
#define DEBUGWRITE(x)
|
|
#define DEBUGPOLL(x)
|
|
#define DEBUGRXINT(x)
|
|
#define DEBUGTXINT(x)
|
|
#define DEBUGTRDMA(x)
|
|
#define DEBUGOUTBUF(x)
|
|
|
|
typedef struct sync_port
|
|
{
|
|
reg_scope_instances regi_sser;
|
|
reg_scope_instances regi_dmain;
|
|
reg_scope_instances regi_dmaout;
|
|
|
|
char started; /* 1 if port has been started */
|
|
char port_nbr; /* Port 0 or 1 */
|
|
char busy; /* 1 if port is busy */
|
|
|
|
char enabled; /* 1 if port is enabled */
|
|
char use_dma; /* 1 if port uses dma */
|
|
char tr_running;
|
|
|
|
char init_irqs;
|
|
int output;
|
|
int input;
|
|
|
|
/* Next byte to be read by application */
|
|
volatile unsigned char *volatile readp;
|
|
/* Next byte to be written by etrax */
|
|
volatile unsigned char *volatile writep;
|
|
|
|
unsigned int in_buffer_size;
|
|
unsigned int inbufchunk;
|
|
unsigned char out_buffer[OUT_BUFFER_SIZE] __attribute__ ((aligned(32)));
|
|
unsigned char in_buffer[IN_BUFFER_SIZE]__attribute__ ((aligned(32)));
|
|
unsigned char flip[IN_BUFFER_SIZE] __attribute__ ((aligned(32)));
|
|
struct dma_descr_data* next_rx_desc;
|
|
struct dma_descr_data* prev_rx_desc;
|
|
|
|
/* Pointer to the first available descriptor in the ring,
|
|
* unless active_tr_descr == catch_tr_descr and a dma
|
|
* transfer is active */
|
|
struct dma_descr_data *active_tr_descr;
|
|
|
|
/* Pointer to the first allocated descriptor in the ring */
|
|
struct dma_descr_data *catch_tr_descr;
|
|
|
|
/* Pointer to the descriptor with the current end-of-list */
|
|
struct dma_descr_data *prev_tr_descr;
|
|
int full;
|
|
|
|
/* Pointer to the first byte being read by DMA
|
|
* or current position in out_buffer if not using DMA. */
|
|
unsigned char *out_rd_ptr;
|
|
|
|
/* Number of bytes currently locked for being read by DMA */
|
|
int out_buf_count;
|
|
|
|
dma_descr_data in_descr[NBR_IN_DESCR] __attribute__ ((__aligned__(16)));
|
|
dma_descr_context in_context __attribute__ ((__aligned__(32)));
|
|
dma_descr_data out_descr[NBR_OUT_DESCR]
|
|
__attribute__ ((__aligned__(16)));
|
|
dma_descr_context out_context __attribute__ ((__aligned__(32)));
|
|
wait_queue_head_t out_wait_q;
|
|
wait_queue_head_t in_wait_q;
|
|
|
|
spinlock_t lock;
|
|
} sync_port;
|
|
|
|
static int etrax_sync_serial_init(void);
|
|
static void initialize_port(int portnbr);
|
|
static inline int sync_data_avail(struct sync_port *port);
|
|
|
|
static int sync_serial_open(struct inode *, struct file*);
|
|
static int sync_serial_release(struct inode*, struct file*);
|
|
static unsigned int sync_serial_poll(struct file *filp, poll_table *wait);
|
|
|
|
static int sync_serial_ioctl(struct inode*, struct file*,
|
|
unsigned int cmd, unsigned long arg);
|
|
static ssize_t sync_serial_write(struct file * file, const char * buf,
|
|
size_t count, loff_t *ppos);
|
|
static ssize_t sync_serial_read(struct file *file, char *buf,
|
|
size_t count, loff_t *ppos);
|
|
|
|
#if (defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL_PORT0) && \
|
|
defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL0_DMA)) || \
|
|
(defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL_PORT1) && \
|
|
defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL1_DMA))
|
|
#define SYNC_SER_DMA
|
|
#endif
|
|
|
|
static void send_word(sync_port* port);
|
|
static void start_dma_out(struct sync_port *port, const char *data, int count);
|
|
static void start_dma_in(sync_port* port);
|
|
#ifdef SYNC_SER_DMA
|
|
static irqreturn_t tr_interrupt(int irq, void *dev_id);
|
|
static irqreturn_t rx_interrupt(int irq, void *dev_id);
|
|
#endif
|
|
|
|
#if (defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL_PORT0) && \
|
|
!defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL0_DMA)) || \
|
|
(defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL_PORT1) && \
|
|
!defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL1_DMA))
|
|
#define SYNC_SER_MANUAL
|
|
#endif
|
|
#ifdef SYNC_SER_MANUAL
|
|
static irqreturn_t manual_interrupt(int irq, void *dev_id);
|
|
#endif
|
|
|
|
#ifdef CONFIG_ETRAXFS /* ETRAX FS */
|
|
#define OUT_DMA_NBR 4
|
|
#define IN_DMA_NBR 5
|
|
#define PINMUX_SSER pinmux_sser0
|
|
#define SYNCSER_INST regi_sser0
|
|
#define SYNCSER_INTR_VECT SSER0_INTR_VECT
|
|
#define OUT_DMA_INST regi_dma4
|
|
#define IN_DMA_INST regi_dma5
|
|
#define DMA_OUT_INTR_VECT DMA4_INTR_VECT
|
|
#define DMA_IN_INTR_VECT DMA5_INTR_VECT
|
|
#define REQ_DMA_SYNCSER dma_sser0
|
|
#else /* Artpec-3 */
|
|
#define OUT_DMA_NBR 6
|
|
#define IN_DMA_NBR 7
|
|
#define PINMUX_SSER pinmux_sser
|
|
#define SYNCSER_INST regi_sser
|
|
#define SYNCSER_INTR_VECT SSER_INTR_VECT
|
|
#define OUT_DMA_INST regi_dma6
|
|
#define IN_DMA_INST regi_dma7
|
|
#define DMA_OUT_INTR_VECT DMA6_INTR_VECT
|
|
#define DMA_IN_INTR_VECT DMA7_INTR_VECT
|
|
#define REQ_DMA_SYNCSER dma_sser
|
|
#endif
|
|
|
|
/* The ports */
|
|
static struct sync_port ports[]=
|
|
{
|
|
{
|
|
.regi_sser = SYNCSER_INST,
|
|
.regi_dmaout = OUT_DMA_INST,
|
|
.regi_dmain = IN_DMA_INST,
|
|
#if defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL0_DMA)
|
|
.use_dma = 1,
|
|
#else
|
|
.use_dma = 0,
|
|
#endif
|
|
}
|
|
#ifdef CONFIG_ETRAXFS
|
|
,
|
|
|
|
{
|
|
.regi_sser = regi_sser1,
|
|
.regi_dmaout = regi_dma6,
|
|
.regi_dmain = regi_dma7,
|
|
#if defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL1_DMA)
|
|
.use_dma = 1,
|
|
#else
|
|
.use_dma = 0,
|
|
#endif
|
|
}
|
|
#endif
|
|
};
|
|
|
|
#define NBR_PORTS ARRAY_SIZE(ports)
|
|
|
|
static const struct file_operations sync_serial_fops = {
|
|
.owner = THIS_MODULE,
|
|
.write = sync_serial_write,
|
|
.read = sync_serial_read,
|
|
.poll = sync_serial_poll,
|
|
.ioctl = sync_serial_ioctl,
|
|
.open = sync_serial_open,
|
|
.release = sync_serial_release
|
|
};
|
|
|
|
static int __init etrax_sync_serial_init(void)
|
|
{
|
|
ports[0].enabled = 0;
|
|
#ifdef CONFIG_ETRAXFS
|
|
ports[1].enabled = 0;
|
|
#endif
|
|
if (register_chrdev(SYNC_SERIAL_MAJOR, "sync serial",
|
|
&sync_serial_fops) < 0) {
|
|
printk(KERN_WARNING
|
|
"Unable to get major for synchronous serial port\n");
|
|
return -EBUSY;
|
|
}
|
|
|
|
/* Initialize Ports */
|
|
#if defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL_PORT0)
|
|
if (crisv32_pinmux_alloc_fixed(PINMUX_SSER)) {
|
|
printk(KERN_WARNING
|
|
"Unable to alloc pins for synchronous serial port 0\n");
|
|
return -EIO;
|
|
}
|
|
ports[0].enabled = 1;
|
|
initialize_port(0);
|
|
#endif
|
|
|
|
#if defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL_PORT1)
|
|
if (crisv32_pinmux_alloc_fixed(pinmux_sser1)) {
|
|
printk(KERN_WARNING
|
|
"Unable to alloc pins for synchronous serial port 0\n");
|
|
return -EIO;
|
|
}
|
|
ports[1].enabled = 1;
|
|
initialize_port(1);
|
|
#endif
|
|
|
|
#ifdef CONFIG_ETRAXFS
|
|
printk(KERN_INFO "ETRAX FS synchronous serial port driver\n");
|
|
#else
|
|
printk(KERN_INFO "Artpec-3 synchronous serial port driver\n");
|
|
#endif
|
|
return 0;
|
|
}
|
|
|
|
static void __init initialize_port(int portnbr)
|
|
{
|
|
int __attribute__((unused)) i;
|
|
struct sync_port *port = &ports[portnbr];
|
|
reg_sser_rw_cfg cfg = {0};
|
|
reg_sser_rw_frm_cfg frm_cfg = {0};
|
|
reg_sser_rw_tr_cfg tr_cfg = {0};
|
|
reg_sser_rw_rec_cfg rec_cfg = {0};
|
|
|
|
DEBUG(printk(KERN_DEBUG "Init sync serial port %d\n", portnbr));
|
|
|
|
port->port_nbr = portnbr;
|
|
port->init_irqs = 1;
|
|
|
|
port->out_rd_ptr = port->out_buffer;
|
|
port->out_buf_count = 0;
|
|
|
|
port->output = 1;
|
|
port->input = 0;
|
|
|
|
port->readp = port->flip;
|
|
port->writep = port->flip;
|
|
port->in_buffer_size = IN_BUFFER_SIZE;
|
|
port->inbufchunk = IN_DESCR_SIZE;
|
|
port->next_rx_desc = &port->in_descr[0];
|
|
port->prev_rx_desc = &port->in_descr[NBR_IN_DESCR-1];
|
|
port->prev_rx_desc->eol = 1;
|
|
|
|
init_waitqueue_head(&port->out_wait_q);
|
|
init_waitqueue_head(&port->in_wait_q);
|
|
|
|
spin_lock_init(&port->lock);
|
|
|
|
cfg.out_clk_src = regk_sser_intern_clk;
|
|
cfg.out_clk_pol = regk_sser_pos;
|
|
cfg.clk_od_mode = regk_sser_no;
|
|
cfg.clk_dir = regk_sser_out;
|
|
cfg.gate_clk = regk_sser_no;
|
|
cfg.base_freq = regk_sser_f29_493;
|
|
cfg.clk_div = 256;
|
|
REG_WR(sser, port->regi_sser, rw_cfg, cfg);
|
|
|
|
frm_cfg.wordrate = DEFAULT_WORD_RATE;
|
|
frm_cfg.type = regk_sser_edge;
|
|
frm_cfg.frame_pin_dir = regk_sser_out;
|
|
frm_cfg.frame_pin_use = regk_sser_frm;
|
|
frm_cfg.status_pin_dir = regk_sser_in;
|
|
frm_cfg.status_pin_use = regk_sser_hold;
|
|
frm_cfg.out_on = regk_sser_tr;
|
|
frm_cfg.tr_delay = 1;
|
|
REG_WR(sser, port->regi_sser, rw_frm_cfg, frm_cfg);
|
|
|
|
tr_cfg.urun_stop = regk_sser_no;
|
|
tr_cfg.sample_size = 7;
|
|
tr_cfg.sh_dir = regk_sser_msbfirst;
|
|
tr_cfg.use_dma = port->use_dma ? regk_sser_yes : regk_sser_no;
|
|
#if 0
|
|
tr_cfg.rate_ctrl = regk_sser_bulk;
|
|
tr_cfg.data_pin_use = regk_sser_dout;
|
|
#else
|
|
tr_cfg.rate_ctrl = regk_sser_iso;
|
|
tr_cfg.data_pin_use = regk_sser_dout;
|
|
#endif
|
|
tr_cfg.bulk_wspace = 1;
|
|
REG_WR(sser, port->regi_sser, rw_tr_cfg, tr_cfg);
|
|
|
|
rec_cfg.sample_size = 7;
|
|
rec_cfg.sh_dir = regk_sser_msbfirst;
|
|
rec_cfg.use_dma = port->use_dma ? regk_sser_yes : regk_sser_no;
|
|
rec_cfg.fifo_thr = regk_sser_inf;
|
|
REG_WR(sser, port->regi_sser, rw_rec_cfg, rec_cfg);
|
|
|
|
#ifdef SYNC_SER_DMA
|
|
/* Setup the descriptor ring for dma out/transmit. */
|
|
for (i = 0; i < NBR_OUT_DESCR; i++) {
|
|
port->out_descr[i].wait = 0;
|
|
port->out_descr[i].intr = 1;
|
|
port->out_descr[i].eol = 0;
|
|
port->out_descr[i].out_eop = 0;
|
|
port->out_descr[i].next =
|
|
(dma_descr_data *)virt_to_phys(&port->out_descr[i+1]);
|
|
}
|
|
|
|
/* Create a ring from the list. */
|
|
port->out_descr[NBR_OUT_DESCR-1].next =
|
|
(dma_descr_data *)virt_to_phys(&port->out_descr[0]);
|
|
|
|
/* Setup context for traversing the ring. */
|
|
port->active_tr_descr = &port->out_descr[0];
|
|
port->prev_tr_descr = &port->out_descr[NBR_OUT_DESCR-1];
|
|
port->catch_tr_descr = &port->out_descr[0];
|
|
#endif
|
|
}
|
|
|
|
static inline int sync_data_avail(struct sync_port *port)
|
|
{
|
|
int avail;
|
|
unsigned char *start;
|
|
unsigned char *end;
|
|
|
|
start = (unsigned char*)port->readp; /* cast away volatile */
|
|
end = (unsigned char*)port->writep; /* cast away volatile */
|
|
/* 0123456789 0123456789
|
|
* ----- - -----
|
|
* ^rp ^wp ^wp ^rp
|
|
*/
|
|
|
|
if (end >= start)
|
|
avail = end - start;
|
|
else
|
|
avail = port->in_buffer_size - (start - end);
|
|
return avail;
|
|
}
|
|
|
|
static inline int sync_data_avail_to_end(struct sync_port *port)
|
|
{
|
|
int avail;
|
|
unsigned char *start;
|
|
unsigned char *end;
|
|
|
|
start = (unsigned char*)port->readp; /* cast away volatile */
|
|
end = (unsigned char*)port->writep; /* cast away volatile */
|
|
/* 0123456789 0123456789
|
|
* ----- -----
|
|
* ^rp ^wp ^wp ^rp
|
|
*/
|
|
|
|
if (end >= start)
|
|
avail = end - start;
|
|
else
|
|
avail = port->flip + port->in_buffer_size - start;
|
|
return avail;
|
|
}
|
|
|
|
static int sync_serial_open(struct inode *inode, struct file *file)
|
|
{
|
|
int dev = iminor(inode);
|
|
int ret = -EBUSY;
|
|
sync_port *port;
|
|
reg_dma_rw_cfg cfg = {.en = regk_dma_yes};
|
|
reg_dma_rw_intr_mask intr_mask = {.data = regk_dma_yes};
|
|
|
|
lock_kernel();
|
|
DEBUG(printk(KERN_DEBUG "Open sync serial port %d\n", dev));
|
|
|
|
if (dev < 0 || dev >= NBR_PORTS || !ports[dev].enabled)
|
|
{
|
|
DEBUG(printk(KERN_DEBUG "Invalid minor %d\n", dev));
|
|
ret = -ENODEV;
|
|
goto out;
|
|
}
|
|
port = &ports[dev];
|
|
/* Allow open this device twice (assuming one reader and one writer) */
|
|
if (port->busy == 2)
|
|
{
|
|
DEBUG(printk(KERN_DEBUG "Device is busy.. \n"));
|
|
goto out;
|
|
}
|
|
|
|
|
|
if (port->init_irqs) {
|
|
if (port->use_dma) {
|
|
if (port == &ports[0]) {
|
|
#ifdef SYNC_SER_DMA
|
|
if (request_irq(DMA_OUT_INTR_VECT,
|
|
tr_interrupt,
|
|
0,
|
|
"synchronous serial 0 dma tr",
|
|
&ports[0])) {
|
|
printk(KERN_CRIT "Can't allocate sync serial port 0 IRQ");
|
|
goto out;
|
|
} else if (request_irq(DMA_IN_INTR_VECT,
|
|
rx_interrupt,
|
|
0,
|
|
"synchronous serial 1 dma rx",
|
|
&ports[0])) {
|
|
free_irq(DMA_OUT_INTR_VECT, &port[0]);
|
|
printk(KERN_CRIT "Can't allocate sync serial port 0 IRQ");
|
|
goto out;
|
|
} else if (crisv32_request_dma(OUT_DMA_NBR,
|
|
"synchronous serial 0 dma tr",
|
|
DMA_VERBOSE_ON_ERROR,
|
|
0,
|
|
REQ_DMA_SYNCSER)) {
|
|
free_irq(DMA_OUT_INTR_VECT, &port[0]);
|
|
free_irq(DMA_IN_INTR_VECT, &port[0]);
|
|
printk(KERN_CRIT "Can't allocate sync serial port 0 TX DMA channel");
|
|
goto out;
|
|
} else if (crisv32_request_dma(IN_DMA_NBR,
|
|
"synchronous serial 0 dma rec",
|
|
DMA_VERBOSE_ON_ERROR,
|
|
0,
|
|
REQ_DMA_SYNCSER)) {
|
|
crisv32_free_dma(OUT_DMA_NBR);
|
|
free_irq(DMA_OUT_INTR_VECT, &port[0]);
|
|
free_irq(DMA_IN_INTR_VECT, &port[0]);
|
|
printk(KERN_CRIT "Can't allocate sync serial port 1 RX DMA channel");
|
|
goto out;
|
|
}
|
|
#endif
|
|
}
|
|
#ifdef CONFIG_ETRAXFS
|
|
else if (port == &ports[1]) {
|
|
#ifdef SYNC_SER_DMA
|
|
if (request_irq(DMA6_INTR_VECT,
|
|
tr_interrupt,
|
|
0,
|
|
"synchronous serial 1 dma tr",
|
|
&ports[1])) {
|
|
printk(KERN_CRIT "Can't allocate sync serial port 1 IRQ");
|
|
goto out;
|
|
} else if (request_irq(DMA7_INTR_VECT,
|
|
rx_interrupt,
|
|
0,
|
|
"synchronous serial 1 dma rx",
|
|
&ports[1])) {
|
|
free_irq(DMA6_INTR_VECT, &ports[1]);
|
|
printk(KERN_CRIT "Can't allocate sync serial port 3 IRQ");
|
|
goto out;
|
|
} else if (crisv32_request_dma(
|
|
SYNC_SER1_TX_DMA_NBR,
|
|
"synchronous serial 1 dma tr",
|
|
DMA_VERBOSE_ON_ERROR,
|
|
0,
|
|
dma_sser1)) {
|
|
free_irq(DMA6_INTR_VECT, &ports[1]);
|
|
free_irq(DMA7_INTR_VECT, &ports[1]);
|
|
printk(KERN_CRIT "Can't allocate sync serial port 3 TX DMA channel");
|
|
goto out;
|
|
} else if (crisv32_request_dma(
|
|
SYNC_SER1_RX_DMA_NBR,
|
|
"synchronous serial 3 dma rec",
|
|
DMA_VERBOSE_ON_ERROR,
|
|
0,
|
|
dma_sser1)) {
|
|
crisv32_free_dma(SYNC_SER1_TX_DMA_NBR);
|
|
free_irq(DMA6_INTR_VECT, &ports[1]);
|
|
free_irq(DMA7_INTR_VECT, &ports[1]);
|
|
printk(KERN_CRIT "Can't allocate sync serial port 3 RX DMA channel");
|
|
goto out;
|
|
}
|
|
#endif
|
|
}
|
|
#endif
|
|
/* Enable DMAs */
|
|
REG_WR(dma, port->regi_dmain, rw_cfg, cfg);
|
|
REG_WR(dma, port->regi_dmaout, rw_cfg, cfg);
|
|
/* Enable DMA IRQs */
|
|
REG_WR(dma, port->regi_dmain, rw_intr_mask, intr_mask);
|
|
REG_WR(dma, port->regi_dmaout, rw_intr_mask, intr_mask);
|
|
/* Set up wordsize = 1 for DMAs. */
|
|
DMA_WR_CMD (port->regi_dmain, regk_dma_set_w_size1);
|
|
DMA_WR_CMD (port->regi_dmaout, regk_dma_set_w_size1);
|
|
|
|
start_dma_in(port);
|
|
port->init_irqs = 0;
|
|
} else { /* !port->use_dma */
|
|
#ifdef SYNC_SER_MANUAL
|
|
if (port == &ports[0]) {
|
|
if (request_irq(SYNCSER_INTR_VECT,
|
|
manual_interrupt,
|
|
0,
|
|
"synchronous serial manual irq",
|
|
&ports[0])) {
|
|
printk("Can't allocate sync serial manual irq");
|
|
goto out;
|
|
}
|
|
}
|
|
#ifdef CONFIG_ETRAXFS
|
|
else if (port == &ports[1]) {
|
|
if (request_irq(SSER1_INTR_VECT,
|
|
manual_interrupt,
|
|
0,
|
|
"synchronous serial manual irq",
|
|
&ports[1])) {
|
|
printk(KERN_CRIT "Can't allocate sync serial manual irq");
|
|
goto out;
|
|
}
|
|
}
|
|
#endif
|
|
port->init_irqs = 0;
|
|
#else
|
|
panic("sync_serial: Manual mode not supported.\n");
|
|
#endif /* SYNC_SER_MANUAL */
|
|
}
|
|
|
|
} /* port->init_irqs */
|
|
|
|
port->busy++;
|
|
ret = 0;
|
|
out:
|
|
unlock_kernel();
|
|
return ret;
|
|
}
|
|
|
|
static int sync_serial_release(struct inode *inode, struct file *file)
|
|
{
|
|
int dev = iminor(inode);
|
|
sync_port *port;
|
|
|
|
if (dev < 0 || dev >= NBR_PORTS || !ports[dev].enabled)
|
|
{
|
|
DEBUG(printk("Invalid minor %d\n", dev));
|
|
return -ENODEV;
|
|
}
|
|
port = &ports[dev];
|
|
if (port->busy)
|
|
port->busy--;
|
|
if (!port->busy)
|
|
/* XXX */ ;
|
|
return 0;
|
|
}
|
|
|
|
static unsigned int sync_serial_poll(struct file *file, poll_table *wait)
|
|
{
|
|
int dev = iminor(file->f_path.dentry->d_inode);
|
|
unsigned int mask = 0;
|
|
sync_port *port;
|
|
DEBUGPOLL( static unsigned int prev_mask = 0; );
|
|
|
|
port = &ports[dev];
|
|
|
|
if (!port->started) {
|
|
reg_sser_rw_cfg cfg = REG_RD(sser, port->regi_sser, rw_cfg);
|
|
reg_sser_rw_rec_cfg rec_cfg =
|
|
REG_RD(sser, port->regi_sser, rw_rec_cfg);
|
|
cfg.en = regk_sser_yes;
|
|
rec_cfg.rec_en = port->input;
|
|
REG_WR(sser, port->regi_sser, rw_cfg, cfg);
|
|
REG_WR(sser, port->regi_sser, rw_rec_cfg, rec_cfg);
|
|
port->started = 1;
|
|
}
|
|
|
|
poll_wait(file, &port->out_wait_q, wait);
|
|
poll_wait(file, &port->in_wait_q, wait);
|
|
|
|
/* No active transfer, descriptors are available */
|
|
if (port->output && !port->tr_running)
|
|
mask |= POLLOUT | POLLWRNORM;
|
|
|
|
/* Descriptor and buffer space available. */
|
|
if (port->output &&
|
|
port->active_tr_descr != port->catch_tr_descr &&
|
|
port->out_buf_count < OUT_BUFFER_SIZE)
|
|
mask |= POLLOUT | POLLWRNORM;
|
|
|
|
/* At least an inbufchunk of data */
|
|
if (port->input && sync_data_avail(port) >= port->inbufchunk)
|
|
mask |= POLLIN | POLLRDNORM;
|
|
|
|
DEBUGPOLL(if (mask != prev_mask)
|
|
printk("sync_serial_poll: mask 0x%08X %s %s\n", mask,
|
|
mask&POLLOUT?"POLLOUT":"", mask&POLLIN?"POLLIN":"");
|
|
prev_mask = mask;
|
|
);
|
|
return mask;
|
|
}
|
|
|
|
static int sync_serial_ioctl(struct inode *inode, struct file *file,
|
|
unsigned int cmd, unsigned long arg)
|
|
{
|
|
int return_val = 0;
|
|
int dma_w_size = regk_dma_set_w_size1;
|
|
int dev = iminor(file->f_path.dentry->d_inode);
|
|
sync_port *port;
|
|
reg_sser_rw_tr_cfg tr_cfg;
|
|
reg_sser_rw_rec_cfg rec_cfg;
|
|
reg_sser_rw_frm_cfg frm_cfg;
|
|
reg_sser_rw_cfg gen_cfg;
|
|
reg_sser_rw_intr_mask intr_mask;
|
|
|
|
if (dev < 0 || dev >= NBR_PORTS || !ports[dev].enabled)
|
|
{
|
|
DEBUG(printk("Invalid minor %d\n", dev));
|
|
return -1;
|
|
}
|
|
port = &ports[dev];
|
|
spin_lock_irq(&port->lock);
|
|
|
|
tr_cfg = REG_RD(sser, port->regi_sser, rw_tr_cfg);
|
|
rec_cfg = REG_RD(sser, port->regi_sser, rw_rec_cfg);
|
|
frm_cfg = REG_RD(sser, port->regi_sser, rw_frm_cfg);
|
|
gen_cfg = REG_RD(sser, port->regi_sser, rw_cfg);
|
|
intr_mask = REG_RD(sser, port->regi_sser, rw_intr_mask);
|
|
|
|
switch(cmd)
|
|
{
|
|
case SSP_SPEED:
|
|
if (GET_SPEED(arg) == CODEC)
|
|
{
|
|
unsigned int freq;
|
|
|
|
gen_cfg.base_freq = regk_sser_f32;
|
|
|
|
/* Clock divider will internally be
|
|
* gen_cfg.clk_div + 1.
|
|
*/
|
|
|
|
freq = GET_FREQ(arg);
|
|
switch (freq) {
|
|
case FREQ_32kHz:
|
|
case FREQ_64kHz:
|
|
case FREQ_128kHz:
|
|
case FREQ_256kHz:
|
|
gen_cfg.clk_div = 125 *
|
|
(1 << (freq - FREQ_256kHz)) - 1;
|
|
break;
|
|
case FREQ_512kHz:
|
|
gen_cfg.clk_div = 62;
|
|
break;
|
|
case FREQ_1MHz:
|
|
case FREQ_2MHz:
|
|
case FREQ_4MHz:
|
|
gen_cfg.clk_div = 8 * (1 << freq) - 1;
|
|
break;
|
|
}
|
|
} else {
|
|
gen_cfg.base_freq = regk_sser_f29_493;
|
|
switch (GET_SPEED(arg)) {
|
|
case SSP150:
|
|
gen_cfg.clk_div = 29493000 / (150 * 8) - 1;
|
|
break;
|
|
case SSP300:
|
|
gen_cfg.clk_div = 29493000 / (300 * 8) - 1;
|
|
break;
|
|
case SSP600:
|
|
gen_cfg.clk_div = 29493000 / (600 * 8) - 1;
|
|
break;
|
|
case SSP1200:
|
|
gen_cfg.clk_div = 29493000 / (1200 * 8) - 1;
|
|
break;
|
|
case SSP2400:
|
|
gen_cfg.clk_div = 29493000 / (2400 * 8) - 1;
|
|
break;
|
|
case SSP4800:
|
|
gen_cfg.clk_div = 29493000 / (4800 * 8) - 1;
|
|
break;
|
|
case SSP9600:
|
|
gen_cfg.clk_div = 29493000 / (9600 * 8) - 1;
|
|
break;
|
|
case SSP19200:
|
|
gen_cfg.clk_div = 29493000 / (19200 * 8) - 1;
|
|
break;
|
|
case SSP28800:
|
|
gen_cfg.clk_div = 29493000 / (28800 * 8) - 1;
|
|
break;
|
|
case SSP57600:
|
|
gen_cfg.clk_div = 29493000 / (57600 * 8) - 1;
|
|
break;
|
|
case SSP115200:
|
|
gen_cfg.clk_div = 29493000 / (115200 * 8) - 1;
|
|
break;
|
|
case SSP230400:
|
|
gen_cfg.clk_div = 29493000 / (230400 * 8) - 1;
|
|
break;
|
|
case SSP460800:
|
|
gen_cfg.clk_div = 29493000 / (460800 * 8) - 1;
|
|
break;
|
|
case SSP921600:
|
|
gen_cfg.clk_div = 29493000 / (921600 * 8) - 1;
|
|
break;
|
|
case SSP3125000:
|
|
gen_cfg.base_freq = regk_sser_f100;
|
|
gen_cfg.clk_div = 100000000 / (3125000 * 8) - 1;
|
|
break;
|
|
|
|
}
|
|
}
|
|
frm_cfg.wordrate = GET_WORD_RATE(arg);
|
|
|
|
break;
|
|
case SSP_MODE:
|
|
switch(arg)
|
|
{
|
|
case MASTER_OUTPUT:
|
|
port->output = 1;
|
|
port->input = 0;
|
|
frm_cfg.out_on = regk_sser_tr;
|
|
frm_cfg.frame_pin_dir = regk_sser_out;
|
|
gen_cfg.clk_dir = regk_sser_out;
|
|
break;
|
|
case SLAVE_OUTPUT:
|
|
port->output = 1;
|
|
port->input = 0;
|
|
frm_cfg.frame_pin_dir = regk_sser_in;
|
|
gen_cfg.clk_dir = regk_sser_in;
|
|
break;
|
|
case MASTER_INPUT:
|
|
port->output = 0;
|
|
port->input = 1;
|
|
frm_cfg.frame_pin_dir = regk_sser_out;
|
|
frm_cfg.out_on = regk_sser_intern_tb;
|
|
gen_cfg.clk_dir = regk_sser_out;
|
|
break;
|
|
case SLAVE_INPUT:
|
|
port->output = 0;
|
|
port->input = 1;
|
|
frm_cfg.frame_pin_dir = regk_sser_in;
|
|
gen_cfg.clk_dir = regk_sser_in;
|
|
break;
|
|
case MASTER_BIDIR:
|
|
port->output = 1;
|
|
port->input = 1;
|
|
frm_cfg.frame_pin_dir = regk_sser_out;
|
|
frm_cfg.out_on = regk_sser_intern_tb;
|
|
gen_cfg.clk_dir = regk_sser_out;
|
|
break;
|
|
case SLAVE_BIDIR:
|
|
port->output = 1;
|
|
port->input = 1;
|
|
frm_cfg.frame_pin_dir = regk_sser_in;
|
|
gen_cfg.clk_dir = regk_sser_in;
|
|
break;
|
|
default:
|
|
spin_unlock_irq(&port->lock);
|
|
return -EINVAL;
|
|
}
|
|
if (!port->use_dma || (arg == MASTER_OUTPUT || arg == SLAVE_OUTPUT))
|
|
intr_mask.rdav = regk_sser_yes;
|
|
break;
|
|
case SSP_FRAME_SYNC:
|
|
if (arg & NORMAL_SYNC) {
|
|
frm_cfg.rec_delay = 1;
|
|
frm_cfg.tr_delay = 1;
|
|
}
|
|
else if (arg & EARLY_SYNC)
|
|
frm_cfg.rec_delay = frm_cfg.tr_delay = 0;
|
|
else if (arg & SECOND_WORD_SYNC) {
|
|
frm_cfg.rec_delay = 7;
|
|
frm_cfg.tr_delay = 1;
|
|
}
|
|
|
|
tr_cfg.bulk_wspace = frm_cfg.tr_delay;
|
|
frm_cfg.early_wend = regk_sser_yes;
|
|
if (arg & BIT_SYNC)
|
|
frm_cfg.type = regk_sser_edge;
|
|
else if (arg & WORD_SYNC)
|
|
frm_cfg.type = regk_sser_level;
|
|
else if (arg & EXTENDED_SYNC)
|
|
frm_cfg.early_wend = regk_sser_no;
|
|
|
|
if (arg & SYNC_ON)
|
|
frm_cfg.frame_pin_use = regk_sser_frm;
|
|
else if (arg & SYNC_OFF)
|
|
frm_cfg.frame_pin_use = regk_sser_gio0;
|
|
|
|
dma_w_size = regk_dma_set_w_size2;
|
|
if (arg & WORD_SIZE_8) {
|
|
rec_cfg.sample_size = tr_cfg.sample_size = 7;
|
|
dma_w_size = regk_dma_set_w_size1;
|
|
} else if (arg & WORD_SIZE_12)
|
|
rec_cfg.sample_size = tr_cfg.sample_size = 11;
|
|
else if (arg & WORD_SIZE_16)
|
|
rec_cfg.sample_size = tr_cfg.sample_size = 15;
|
|
else if (arg & WORD_SIZE_24)
|
|
rec_cfg.sample_size = tr_cfg.sample_size = 23;
|
|
else if (arg & WORD_SIZE_32)
|
|
rec_cfg.sample_size = tr_cfg.sample_size = 31;
|
|
|
|
if (arg & BIT_ORDER_MSB)
|
|
rec_cfg.sh_dir = tr_cfg.sh_dir = regk_sser_msbfirst;
|
|
else if (arg & BIT_ORDER_LSB)
|
|
rec_cfg.sh_dir = tr_cfg.sh_dir = regk_sser_lsbfirst;
|
|
|
|
if (arg & FLOW_CONTROL_ENABLE) {
|
|
frm_cfg.status_pin_use = regk_sser_frm;
|
|
rec_cfg.fifo_thr = regk_sser_thr16;
|
|
} else if (arg & FLOW_CONTROL_DISABLE) {
|
|
frm_cfg.status_pin_use = regk_sser_gio0;
|
|
rec_cfg.fifo_thr = regk_sser_inf;
|
|
}
|
|
|
|
if (arg & CLOCK_NOT_GATED)
|
|
gen_cfg.gate_clk = regk_sser_no;
|
|
else if (arg & CLOCK_GATED)
|
|
gen_cfg.gate_clk = regk_sser_yes;
|
|
|
|
break;
|
|
case SSP_IPOLARITY:
|
|
/* NOTE!! negedge is considered NORMAL */
|
|
if (arg & CLOCK_NORMAL)
|
|
rec_cfg.clk_pol = regk_sser_neg;
|
|
else if (arg & CLOCK_INVERT)
|
|
rec_cfg.clk_pol = regk_sser_pos;
|
|
|
|
if (arg & FRAME_NORMAL)
|
|
frm_cfg.level = regk_sser_pos_hi;
|
|
else if (arg & FRAME_INVERT)
|
|
frm_cfg.level = regk_sser_neg_lo;
|
|
|
|
if (arg & STATUS_NORMAL)
|
|
gen_cfg.hold_pol = regk_sser_pos;
|
|
else if (arg & STATUS_INVERT)
|
|
gen_cfg.hold_pol = regk_sser_neg;
|
|
break;
|
|
case SSP_OPOLARITY:
|
|
if (arg & CLOCK_NORMAL)
|
|
gen_cfg.out_clk_pol = regk_sser_pos;
|
|
else if (arg & CLOCK_INVERT)
|
|
gen_cfg.out_clk_pol = regk_sser_neg;
|
|
|
|
if (arg & FRAME_NORMAL)
|
|
frm_cfg.level = regk_sser_pos_hi;
|
|
else if (arg & FRAME_INVERT)
|
|
frm_cfg.level = regk_sser_neg_lo;
|
|
|
|
if (arg & STATUS_NORMAL)
|
|
gen_cfg.hold_pol = regk_sser_pos;
|
|
else if (arg & STATUS_INVERT)
|
|
gen_cfg.hold_pol = regk_sser_neg;
|
|
break;
|
|
case SSP_SPI:
|
|
rec_cfg.fifo_thr = regk_sser_inf;
|
|
rec_cfg.sh_dir = tr_cfg.sh_dir = regk_sser_msbfirst;
|
|
rec_cfg.sample_size = tr_cfg.sample_size = 7;
|
|
frm_cfg.frame_pin_use = regk_sser_frm;
|
|
frm_cfg.type = regk_sser_level;
|
|
frm_cfg.tr_delay = 1;
|
|
frm_cfg.level = regk_sser_neg_lo;
|
|
if (arg & SPI_SLAVE)
|
|
{
|
|
rec_cfg.clk_pol = regk_sser_neg;
|
|
gen_cfg.clk_dir = regk_sser_in;
|
|
port->input = 1;
|
|
port->output = 0;
|
|
}
|
|
else
|
|
{
|
|
gen_cfg.out_clk_pol = regk_sser_pos;
|
|
port->input = 0;
|
|
port->output = 1;
|
|
gen_cfg.clk_dir = regk_sser_out;
|
|
}
|
|
break;
|
|
case SSP_INBUFCHUNK:
|
|
break;
|
|
default:
|
|
return_val = -1;
|
|
}
|
|
|
|
|
|
if (port->started) {
|
|
rec_cfg.rec_en = port->input;
|
|
gen_cfg.en = (port->output | port->input);
|
|
}
|
|
|
|
REG_WR(sser, port->regi_sser, rw_tr_cfg, tr_cfg);
|
|
REG_WR(sser, port->regi_sser, rw_rec_cfg, rec_cfg);
|
|
REG_WR(sser, port->regi_sser, rw_frm_cfg, frm_cfg);
|
|
REG_WR(sser, port->regi_sser, rw_intr_mask, intr_mask);
|
|
REG_WR(sser, port->regi_sser, rw_cfg, gen_cfg);
|
|
|
|
|
|
if (cmd == SSP_FRAME_SYNC && (arg & (WORD_SIZE_8 | WORD_SIZE_12 |
|
|
WORD_SIZE_16 | WORD_SIZE_24 | WORD_SIZE_32))) {
|
|
int en = gen_cfg.en;
|
|
gen_cfg.en = 0;
|
|
REG_WR(sser, port->regi_sser, rw_cfg, gen_cfg);
|
|
/* ##### Should DMA be stoped before we change dma size? */
|
|
DMA_WR_CMD(port->regi_dmain, dma_w_size);
|
|
DMA_WR_CMD(port->regi_dmaout, dma_w_size);
|
|
gen_cfg.en = en;
|
|
REG_WR(sser, port->regi_sser, rw_cfg, gen_cfg);
|
|
}
|
|
|
|
spin_unlock_irq(&port->lock);
|
|
return return_val;
|
|
}
|
|
|
|
/* NOTE: sync_serial_write does not support concurrency */
|
|
static ssize_t sync_serial_write(struct file *file, const char *buf,
|
|
size_t count, loff_t *ppos)
|
|
{
|
|
int dev = iminor(file->f_path.dentry->d_inode);
|
|
DECLARE_WAITQUEUE(wait, current);
|
|
struct sync_port *port;
|
|
int trunc_count;
|
|
unsigned long flags;
|
|
int bytes_free;
|
|
int out_buf_count;
|
|
|
|
unsigned char *rd_ptr; /* First allocated byte in the buffer */
|
|
unsigned char *wr_ptr; /* First free byte in the buffer */
|
|
unsigned char *buf_stop_ptr; /* Last byte + 1 */
|
|
|
|
if (dev < 0 || dev >= NBR_PORTS || !ports[dev].enabled) {
|
|
DEBUG(printk("Invalid minor %d\n", dev));
|
|
return -ENODEV;
|
|
}
|
|
port = &ports[dev];
|
|
|
|
/* |<- OUT_BUFFER_SIZE ->|
|
|
* |<- out_buf_count ->|
|
|
* |<- trunc_count ->| ...->|
|
|
* ______________________________________________________
|
|
* | free | data | free |
|
|
* |_________|___________________|________________________|
|
|
* ^ rd_ptr ^ wr_ptr
|
|
*/
|
|
DEBUGWRITE(printk(KERN_DEBUG "W d%d c %lu a: %p c: %p\n",
|
|
port->port_nbr, count, port->active_tr_descr,
|
|
port->catch_tr_descr));
|
|
|
|
/* Read variables that may be updated by interrupts */
|
|
spin_lock_irqsave(&port->lock, flags);
|
|
rd_ptr = port->out_rd_ptr;
|
|
out_buf_count = port->out_buf_count;
|
|
spin_unlock_irqrestore(&port->lock, flags);
|
|
|
|
/* Check if resources are available */
|
|
if (port->tr_running &&
|
|
((port->use_dma && port->active_tr_descr == port->catch_tr_descr) ||
|
|
out_buf_count >= OUT_BUFFER_SIZE)) {
|
|
DEBUGWRITE(printk(KERN_DEBUG "sser%d full\n", dev));
|
|
return -EAGAIN;
|
|
}
|
|
|
|
buf_stop_ptr = port->out_buffer + OUT_BUFFER_SIZE;
|
|
|
|
/* Determine pointer to the first free byte, before copying. */
|
|
wr_ptr = rd_ptr + out_buf_count;
|
|
if (wr_ptr >= buf_stop_ptr)
|
|
wr_ptr -= OUT_BUFFER_SIZE;
|
|
|
|
/* If we wrap the ring buffer, let the user space program handle it by
|
|
* truncating the data. This could be more elegant, small buffer
|
|
* fragments may occur.
|
|
*/
|
|
bytes_free = OUT_BUFFER_SIZE - out_buf_count;
|
|
if (wr_ptr + bytes_free > buf_stop_ptr)
|
|
bytes_free = buf_stop_ptr - wr_ptr;
|
|
trunc_count = (count < bytes_free) ? count : bytes_free;
|
|
|
|
if (copy_from_user(wr_ptr, buf, trunc_count))
|
|
return -EFAULT;
|
|
|
|
DEBUGOUTBUF(printk(KERN_DEBUG "%-4d + %-4d = %-4d %p %p %p\n",
|
|
out_buf_count, trunc_count,
|
|
port->out_buf_count, port->out_buffer,
|
|
wr_ptr, buf_stop_ptr));
|
|
|
|
/* Make sure transmitter/receiver is running */
|
|
if (!port->started) {
|
|
reg_sser_rw_cfg cfg = REG_RD(sser, port->regi_sser, rw_cfg);
|
|
reg_sser_rw_rec_cfg rec_cfg = REG_RD(sser, port->regi_sser, rw_rec_cfg);
|
|
cfg.en = regk_sser_yes;
|
|
rec_cfg.rec_en = port->input;
|
|
REG_WR(sser, port->regi_sser, rw_cfg, cfg);
|
|
REG_WR(sser, port->regi_sser, rw_rec_cfg, rec_cfg);
|
|
port->started = 1;
|
|
}
|
|
|
|
/* Setup wait if blocking */
|
|
if (!(file->f_flags & O_NONBLOCK)) {
|
|
add_wait_queue(&port->out_wait_q, &wait);
|
|
set_current_state(TASK_INTERRUPTIBLE);
|
|
}
|
|
|
|
spin_lock_irqsave(&port->lock, flags);
|
|
port->out_buf_count += trunc_count;
|
|
if (port->use_dma) {
|
|
start_dma_out(port, wr_ptr, trunc_count);
|
|
} else if (!port->tr_running) {
|
|
reg_sser_rw_intr_mask intr_mask;
|
|
intr_mask = REG_RD(sser, port->regi_sser, rw_intr_mask);
|
|
/* Start sender by writing data */
|
|
send_word(port);
|
|
/* and enable transmitter ready IRQ */
|
|
intr_mask.trdy = 1;
|
|
REG_WR(sser, port->regi_sser, rw_intr_mask, intr_mask);
|
|
}
|
|
spin_unlock_irqrestore(&port->lock, flags);
|
|
|
|
/* Exit if non blocking */
|
|
if (file->f_flags & O_NONBLOCK) {
|
|
DEBUGWRITE(printk(KERN_DEBUG "w d%d c %lu %08x\n",
|
|
port->port_nbr, trunc_count,
|
|
REG_RD_INT(dma, port->regi_dmaout, r_intr)));
|
|
return trunc_count;
|
|
}
|
|
|
|
schedule();
|
|
set_current_state(TASK_RUNNING);
|
|
remove_wait_queue(&port->out_wait_q, &wait);
|
|
|
|
if (signal_pending(current))
|
|
return -EINTR;
|
|
|
|
DEBUGWRITE(printk(KERN_DEBUG "w d%d c %lu\n",
|
|
port->port_nbr, trunc_count));
|
|
return trunc_count;
|
|
}
|
|
|
|
static ssize_t sync_serial_read(struct file * file, char * buf,
|
|
size_t count, loff_t *ppos)
|
|
{
|
|
int dev = iminor(file->f_path.dentry->d_inode);
|
|
int avail;
|
|
sync_port *port;
|
|
unsigned char* start;
|
|
unsigned char* end;
|
|
unsigned long flags;
|
|
|
|
if (dev < 0 || dev >= NBR_PORTS || !ports[dev].enabled)
|
|
{
|
|
DEBUG(printk("Invalid minor %d\n", dev));
|
|
return -ENODEV;
|
|
}
|
|
port = &ports[dev];
|
|
|
|
DEBUGREAD(printk("R%d c %d ri %lu wi %lu /%lu\n", dev, count, port->readp - port->flip, port->writep - port->flip, port->in_buffer_size));
|
|
|
|
if (!port->started)
|
|
{
|
|
reg_sser_rw_cfg cfg = REG_RD(sser, port->regi_sser, rw_cfg);
|
|
reg_sser_rw_tr_cfg tr_cfg = REG_RD(sser, port->regi_sser, rw_tr_cfg);
|
|
reg_sser_rw_rec_cfg rec_cfg = REG_RD(sser, port->regi_sser, rw_rec_cfg);
|
|
cfg.en = regk_sser_yes;
|
|
tr_cfg.tr_en = regk_sser_yes;
|
|
rec_cfg.rec_en = regk_sser_yes;
|
|
REG_WR(sser, port->regi_sser, rw_cfg, cfg);
|
|
REG_WR(sser, port->regi_sser, rw_tr_cfg, tr_cfg);
|
|
REG_WR(sser, port->regi_sser, rw_rec_cfg, rec_cfg);
|
|
port->started = 1;
|
|
}
|
|
|
|
/* Calculate number of available bytes */
|
|
/* Save pointers to avoid that they are modified by interrupt */
|
|
spin_lock_irqsave(&port->lock, flags);
|
|
start = (unsigned char*)port->readp; /* cast away volatile */
|
|
end = (unsigned char*)port->writep; /* cast away volatile */
|
|
spin_unlock_irqrestore(&port->lock, flags);
|
|
while ((start == end) && !port->full) /* No data */
|
|
{
|
|
DEBUGREAD(printk(KERN_DEBUG "&"));
|
|
if (file->f_flags & O_NONBLOCK)
|
|
return -EAGAIN;
|
|
|
|
interruptible_sleep_on(&port->in_wait_q);
|
|
if (signal_pending(current))
|
|
return -EINTR;
|
|
|
|
spin_lock_irqsave(&port->lock, flags);
|
|
start = (unsigned char*)port->readp; /* cast away volatile */
|
|
end = (unsigned char*)port->writep; /* cast away volatile */
|
|
spin_unlock_irqrestore(&port->lock, flags);
|
|
}
|
|
|
|
/* Lazy read, never return wrapped data. */
|
|
if (port->full)
|
|
avail = port->in_buffer_size;
|
|
else if (end > start)
|
|
avail = end - start;
|
|
else
|
|
avail = port->flip + port->in_buffer_size - start;
|
|
|
|
count = count > avail ? avail : count;
|
|
if (copy_to_user(buf, start, count))
|
|
return -EFAULT;
|
|
/* Disable interrupts while updating readp */
|
|
spin_lock_irqsave(&port->lock, flags);
|
|
port->readp += count;
|
|
if (port->readp >= port->flip + port->in_buffer_size) /* Wrap? */
|
|
port->readp = port->flip;
|
|
port->full = 0;
|
|
spin_unlock_irqrestore(&port->lock, flags);
|
|
DEBUGREAD(printk("r %d\n", count));
|
|
return count;
|
|
}
|
|
|
|
static void send_word(sync_port* port)
|
|
{
|
|
reg_sser_rw_tr_cfg tr_cfg = REG_RD(sser, port->regi_sser, rw_tr_cfg);
|
|
reg_sser_rw_tr_data tr_data = {0};
|
|
|
|
switch(tr_cfg.sample_size)
|
|
{
|
|
case 8:
|
|
port->out_buf_count--;
|
|
tr_data.data = *port->out_rd_ptr++;
|
|
REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
|
|
if (port->out_rd_ptr >= port->out_buffer + OUT_BUFFER_SIZE)
|
|
port->out_rd_ptr = port->out_buffer;
|
|
break;
|
|
case 12:
|
|
{
|
|
int data = (*port->out_rd_ptr++) << 8;
|
|
data |= *port->out_rd_ptr++;
|
|
port->out_buf_count -= 2;
|
|
tr_data.data = data;
|
|
REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
|
|
if (port->out_rd_ptr >= port->out_buffer + OUT_BUFFER_SIZE)
|
|
port->out_rd_ptr = port->out_buffer;
|
|
}
|
|
break;
|
|
case 16:
|
|
port->out_buf_count -= 2;
|
|
tr_data.data = *(unsigned short *)port->out_rd_ptr;
|
|
REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
|
|
port->out_rd_ptr += 2;
|
|
if (port->out_rd_ptr >= port->out_buffer + OUT_BUFFER_SIZE)
|
|
port->out_rd_ptr = port->out_buffer;
|
|
break;
|
|
case 24:
|
|
port->out_buf_count -= 3;
|
|
tr_data.data = *(unsigned short *)port->out_rd_ptr;
|
|
REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
|
|
port->out_rd_ptr += 2;
|
|
tr_data.data = *port->out_rd_ptr++;
|
|
REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
|
|
if (port->out_rd_ptr >= port->out_buffer + OUT_BUFFER_SIZE)
|
|
port->out_rd_ptr = port->out_buffer;
|
|
break;
|
|
case 32:
|
|
port->out_buf_count -= 4;
|
|
tr_data.data = *(unsigned short *)port->out_rd_ptr;
|
|
REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
|
|
port->out_rd_ptr += 2;
|
|
tr_data.data = *(unsigned short *)port->out_rd_ptr;
|
|
REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
|
|
port->out_rd_ptr += 2;
|
|
if (port->out_rd_ptr >= port->out_buffer + OUT_BUFFER_SIZE)
|
|
port->out_rd_ptr = port->out_buffer;
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void start_dma_out(struct sync_port *port,
|
|
const char *data, int count)
|
|
{
|
|
port->active_tr_descr->buf = (char *) virt_to_phys((char *) data);
|
|
port->active_tr_descr->after = port->active_tr_descr->buf + count;
|
|
port->active_tr_descr->intr = 1;
|
|
|
|
port->active_tr_descr->eol = 1;
|
|
port->prev_tr_descr->eol = 0;
|
|
|
|
DEBUGTRDMA(printk(KERN_DEBUG "Inserting eolr:%p eol@:%p\n",
|
|
port->prev_tr_descr, port->active_tr_descr));
|
|
port->prev_tr_descr = port->active_tr_descr;
|
|
port->active_tr_descr = phys_to_virt((int) port->active_tr_descr->next);
|
|
|
|
if (!port->tr_running) {
|
|
reg_sser_rw_tr_cfg tr_cfg = REG_RD(sser, port->regi_sser,
|
|
rw_tr_cfg);
|
|
|
|
port->out_context.next = 0;
|
|
port->out_context.saved_data =
|
|
(dma_descr_data *)virt_to_phys(port->prev_tr_descr);
|
|
port->out_context.saved_data_buf = port->prev_tr_descr->buf;
|
|
|
|
DMA_START_CONTEXT(port->regi_dmaout,
|
|
virt_to_phys((char *)&port->out_context));
|
|
|
|
tr_cfg.tr_en = regk_sser_yes;
|
|
REG_WR(sser, port->regi_sser, rw_tr_cfg, tr_cfg);
|
|
DEBUGTRDMA(printk(KERN_DEBUG "dma s\n"););
|
|
} else {
|
|
DMA_CONTINUE_DATA(port->regi_dmaout);
|
|
DEBUGTRDMA(printk(KERN_DEBUG "dma c\n"););
|
|
}
|
|
|
|
port->tr_running = 1;
|
|
}
|
|
|
|
static void start_dma_in(sync_port *port)
|
|
{
|
|
int i;
|
|
char *buf;
|
|
port->writep = port->flip;
|
|
|
|
if (port->writep > port->flip + port->in_buffer_size) {
|
|
panic("Offset too large in sync serial driver\n");
|
|
return;
|
|
}
|
|
buf = (char*)virt_to_phys(port->in_buffer);
|
|
for (i = 0; i < NBR_IN_DESCR; i++) {
|
|
port->in_descr[i].buf = buf;
|
|
port->in_descr[i].after = buf + port->inbufchunk;
|
|
port->in_descr[i].intr = 1;
|
|
port->in_descr[i].next = (dma_descr_data*)virt_to_phys(&port->in_descr[i+1]);
|
|
port->in_descr[i].buf = buf;
|
|
buf += port->inbufchunk;
|
|
}
|
|
/* Link the last descriptor to the first */
|
|
port->in_descr[i-1].next = (dma_descr_data*)virt_to_phys(&port->in_descr[0]);
|
|
port->in_descr[i-1].eol = regk_sser_yes;
|
|
port->next_rx_desc = &port->in_descr[0];
|
|
port->prev_rx_desc = &port->in_descr[NBR_IN_DESCR - 1];
|
|
port->in_context.saved_data = (dma_descr_data*)virt_to_phys(&port->in_descr[0]);
|
|
port->in_context.saved_data_buf = port->in_descr[0].buf;
|
|
DMA_START_CONTEXT(port->regi_dmain, virt_to_phys(&port->in_context));
|
|
}
|
|
|
|
#ifdef SYNC_SER_DMA
|
|
static irqreturn_t tr_interrupt(int irq, void *dev_id)
|
|
{
|
|
reg_dma_r_masked_intr masked;
|
|
reg_dma_rw_ack_intr ack_intr = {.data = regk_dma_yes};
|
|
reg_dma_rw_stat stat;
|
|
int i;
|
|
int found = 0;
|
|
int stop_sser = 0;
|
|
|
|
for (i = 0; i < NBR_PORTS; i++) {
|
|
sync_port *port = &ports[i];
|
|
if (!port->enabled || !port->use_dma)
|
|
continue;
|
|
|
|
/* IRQ active for the port? */
|
|
masked = REG_RD(dma, port->regi_dmaout, r_masked_intr);
|
|
if (!masked.data)
|
|
continue;
|
|
|
|
found = 1;
|
|
|
|
/* Check if we should stop the DMA transfer */
|
|
stat = REG_RD(dma, port->regi_dmaout, rw_stat);
|
|
if (stat.list_state == regk_dma_data_at_eol)
|
|
stop_sser = 1;
|
|
|
|
/* Clear IRQ */
|
|
REG_WR(dma, port->regi_dmaout, rw_ack_intr, ack_intr);
|
|
|
|
if (!stop_sser) {
|
|
/* The DMA has completed a descriptor, EOL was not
|
|
* encountered, so step relevant descriptor and
|
|
* datapointers forward. */
|
|
int sent;
|
|
sent = port->catch_tr_descr->after -
|
|
port->catch_tr_descr->buf;
|
|
DEBUGTXINT(printk(KERN_DEBUG "%-4d - %-4d = %-4d\t"
|
|
"in descr %p (ac: %p)\n",
|
|
port->out_buf_count, sent,
|
|
port->out_buf_count - sent,
|
|
port->catch_tr_descr,
|
|
port->active_tr_descr););
|
|
port->out_buf_count -= sent;
|
|
port->catch_tr_descr =
|
|
phys_to_virt((int) port->catch_tr_descr->next);
|
|
port->out_rd_ptr =
|
|
phys_to_virt((int) port->catch_tr_descr->buf);
|
|
} else {
|
|
int i, sent;
|
|
/* EOL handler.
|
|
* Note that if an EOL was encountered during the irq
|
|
* locked section of sync_ser_write the DMA will be
|
|
* restarted and the eol flag will be cleared.
|
|
* The remaining descriptors will be traversed by
|
|
* the descriptor interrupts as usual.
|
|
*/
|
|
i = 0;
|
|
while (!port->catch_tr_descr->eol) {
|
|
sent = port->catch_tr_descr->after -
|
|
port->catch_tr_descr->buf;
|
|
DEBUGOUTBUF(printk(KERN_DEBUG
|
|
"traversing descr %p -%d (%d)\n",
|
|
port->catch_tr_descr,
|
|
sent,
|
|
port->out_buf_count));
|
|
port->out_buf_count -= sent;
|
|
port->catch_tr_descr = phys_to_virt(
|
|
(int)port->catch_tr_descr->next);
|
|
i++;
|
|
if (i >= NBR_OUT_DESCR) {
|
|
/* TODO: Reset and recover */
|
|
panic("sync_serial: missing eol");
|
|
}
|
|
}
|
|
sent = port->catch_tr_descr->after -
|
|
port->catch_tr_descr->buf;
|
|
DEBUGOUTBUF(printk(KERN_DEBUG
|
|
"eol at descr %p -%d (%d)\n",
|
|
port->catch_tr_descr,
|
|
sent,
|
|
port->out_buf_count));
|
|
|
|
port->out_buf_count -= sent;
|
|
|
|
/* Update read pointer to first free byte, we
|
|
* may already be writing data there. */
|
|
port->out_rd_ptr =
|
|
phys_to_virt((int) port->catch_tr_descr->after);
|
|
if (port->out_rd_ptr > port->out_buffer +
|
|
OUT_BUFFER_SIZE)
|
|
port->out_rd_ptr = port->out_buffer;
|
|
|
|
reg_sser_rw_tr_cfg tr_cfg =
|
|
REG_RD(sser, port->regi_sser, rw_tr_cfg);
|
|
DEBUGTXINT(printk(KERN_DEBUG
|
|
"tr_int DMA stop %d, set catch @ %p\n",
|
|
port->out_buf_count,
|
|
port->active_tr_descr));
|
|
if (port->out_buf_count != 0)
|
|
printk(KERN_CRIT "sync_ser: buffer not "
|
|
"empty after eol.\n");
|
|
port->catch_tr_descr = port->active_tr_descr;
|
|
port->tr_running = 0;
|
|
tr_cfg.tr_en = regk_sser_no;
|
|
REG_WR(sser, port->regi_sser, rw_tr_cfg, tr_cfg);
|
|
}
|
|
/* wake up the waiting process */
|
|
wake_up_interruptible(&port->out_wait_q);
|
|
}
|
|
return IRQ_RETVAL(found);
|
|
} /* tr_interrupt */
|
|
|
|
static irqreturn_t rx_interrupt(int irq, void *dev_id)
|
|
{
|
|
reg_dma_r_masked_intr masked;
|
|
reg_dma_rw_ack_intr ack_intr = {.data = regk_dma_yes};
|
|
|
|
int i;
|
|
int found = 0;
|
|
|
|
for (i = 0; i < NBR_PORTS; i++)
|
|
{
|
|
sync_port *port = &ports[i];
|
|
|
|
if (!port->enabled || !port->use_dma )
|
|
continue;
|
|
|
|
masked = REG_RD(dma, port->regi_dmain, r_masked_intr);
|
|
|
|
if (masked.data) /* Descriptor interrupt */
|
|
{
|
|
found = 1;
|
|
while (REG_RD(dma, port->regi_dmain, rw_data) !=
|
|
virt_to_phys(port->next_rx_desc)) {
|
|
DEBUGRXINT(printk(KERN_DEBUG "!"));
|
|
if (port->writep + port->inbufchunk > port->flip + port->in_buffer_size) {
|
|
int first_size = port->flip + port->in_buffer_size - port->writep;
|
|
memcpy((char*)port->writep, phys_to_virt((unsigned)port->next_rx_desc->buf), first_size);
|
|
memcpy(port->flip, phys_to_virt((unsigned)port->next_rx_desc->buf+first_size), port->inbufchunk - first_size);
|
|
port->writep = port->flip + port->inbufchunk - first_size;
|
|
} else {
|
|
memcpy((char*)port->writep,
|
|
phys_to_virt((unsigned)port->next_rx_desc->buf),
|
|
port->inbufchunk);
|
|
port->writep += port->inbufchunk;
|
|
if (port->writep >= port->flip + port->in_buffer_size)
|
|
port->writep = port->flip;
|
|
}
|
|
if (port->writep == port->readp)
|
|
{
|
|
port->full = 1;
|
|
}
|
|
|
|
port->next_rx_desc->eol = 1;
|
|
port->prev_rx_desc->eol = 0;
|
|
/* Cache bug workaround */
|
|
flush_dma_descr(port->prev_rx_desc, 0);
|
|
port->prev_rx_desc = port->next_rx_desc;
|
|
port->next_rx_desc = phys_to_virt((unsigned)port->next_rx_desc->next);
|
|
/* Cache bug workaround */
|
|
flush_dma_descr(port->prev_rx_desc, 1);
|
|
/* wake up the waiting process */
|
|
wake_up_interruptible(&port->in_wait_q);
|
|
DMA_CONTINUE(port->regi_dmain);
|
|
REG_WR(dma, port->regi_dmain, rw_ack_intr, ack_intr);
|
|
|
|
}
|
|
}
|
|
}
|
|
return IRQ_RETVAL(found);
|
|
} /* rx_interrupt */
|
|
#endif /* SYNC_SER_DMA */
|
|
|
|
#ifdef SYNC_SER_MANUAL
|
|
static irqreturn_t manual_interrupt(int irq, void *dev_id)
|
|
{
|
|
int i;
|
|
int found = 0;
|
|
reg_sser_r_masked_intr masked;
|
|
|
|
for (i = 0; i < NBR_PORTS; i++)
|
|
{
|
|
sync_port *port = &ports[i];
|
|
|
|
if (!port->enabled || port->use_dma)
|
|
{
|
|
continue;
|
|
}
|
|
|
|
masked = REG_RD(sser, port->regi_sser, r_masked_intr);
|
|
if (masked.rdav) /* Data received? */
|
|
{
|
|
reg_sser_rw_rec_cfg rec_cfg = REG_RD(sser, port->regi_sser, rw_rec_cfg);
|
|
reg_sser_r_rec_data data = REG_RD(sser, port->regi_sser, r_rec_data);
|
|
found = 1;
|
|
/* Read data */
|
|
switch(rec_cfg.sample_size)
|
|
{
|
|
case 8:
|
|
*port->writep++ = data.data & 0xff;
|
|
break;
|
|
case 12:
|
|
*port->writep = (data.data & 0x0ff0) >> 4;
|
|
*(port->writep + 1) = data.data & 0x0f;
|
|
port->writep+=2;
|
|
break;
|
|
case 16:
|
|
*(unsigned short*)port->writep = data.data;
|
|
port->writep+=2;
|
|
break;
|
|
case 24:
|
|
*(unsigned int*)port->writep = data.data;
|
|
port->writep+=3;
|
|
break;
|
|
case 32:
|
|
*(unsigned int*)port->writep = data.data;
|
|
port->writep+=4;
|
|
break;
|
|
}
|
|
|
|
if (port->writep >= port->flip + port->in_buffer_size) /* Wrap? */
|
|
port->writep = port->flip;
|
|
if (port->writep == port->readp) {
|
|
/* receive buffer overrun, discard oldest data
|
|
*/
|
|
port->readp++;
|
|
if (port->readp >= port->flip + port->in_buffer_size) /* Wrap? */
|
|
port->readp = port->flip;
|
|
}
|
|
if (sync_data_avail(port) >= port->inbufchunk)
|
|
wake_up_interruptible(&port->in_wait_q); /* Wake up application */
|
|
}
|
|
|
|
if (masked.trdy) /* Transmitter ready? */
|
|
{
|
|
found = 1;
|
|
if (port->out_buf_count > 0) /* More data to send */
|
|
send_word(port);
|
|
else /* transmission finished */
|
|
{
|
|
reg_sser_rw_intr_mask intr_mask;
|
|
intr_mask = REG_RD(sser, port->regi_sser, rw_intr_mask);
|
|
intr_mask.trdy = 0;
|
|
REG_WR(sser, port->regi_sser, rw_intr_mask, intr_mask);
|
|
wake_up_interruptible(&port->out_wait_q); /* Wake up application */
|
|
}
|
|
}
|
|
}
|
|
return IRQ_RETVAL(found);
|
|
}
|
|
#endif
|
|
|
|
module_init(etrax_sync_serial_init);
|