zephyr/drivers/dma/dma_nios2_msgdma.c

235 lines
6.2 KiB
C

/*
* Copyright (c) 2018 Intel Corporation
*
* SPDX-License-Identifier: Apache-2.0
*/
#define DT_DRV_COMPAT altr_msgdma
#include <zephyr/device.h>
#include <errno.h>
#include <zephyr/init.h>
#include <string.h>
#include <soc.h>
#include <zephyr/drivers/dma.h>
#include <altera_common.h>
#include "altera_msgdma_csr_regs.h"
#include "altera_msgdma_descriptor_regs.h"
#include "altera_msgdma.h"
#include <zephyr/logging/log.h>
#include <zephyr/irq.h>
LOG_MODULE_REGISTER(dma_nios2, CONFIG_DMA_LOG_LEVEL);
/* Device configuration parameters */
struct nios2_msgdma_dev_data {
const struct device *dev;
alt_msgdma_dev *msgdma_dev;
alt_msgdma_standard_descriptor desc;
uint32_t direction;
struct k_sem sem_lock;
void *user_data;
dma_callback_t dma_callback;
};
static void nios2_msgdma_isr(void *arg)
{
const struct device *dev = (const struct device *)arg;
struct nios2_msgdma_dev_data *dev_data = (struct nios2_msgdma_dev_data *)dev->data;
/* Call Altera HAL driver ISR */
alt_handle_irq(dev_data->msgdma_dev, DT_INST_IRQN(0));
}
static void nios2_msgdma_callback(void *context)
{
struct nios2_msgdma_dev_data *dev_data =
(struct nios2_msgdma_dev_data *)context;
int dma_status;
uint32_t status;
status = IORD_ALTERA_MSGDMA_CSR_STATUS(dev_data->msgdma_dev->csr_base);
if (status & ALTERA_MSGDMA_CSR_STOPPED_ON_ERROR_MASK) {
dma_status = -EIO;
} else if (status & ALTERA_MSGDMA_CSR_BUSY_MASK) {
dma_status = -EBUSY;
} else {
dma_status = DMA_STATUS_COMPLETE;
}
LOG_DBG("msgdma csr status Reg: 0x%x", status);
dev_data->dma_callback(dev_data->dev, dev_data->user_data, 0, dma_status);
}
static int nios2_msgdma_config(const struct device *dev, uint32_t channel,
struct dma_config *cfg)
{
struct nios2_msgdma_dev_data *dev_data = (struct nios2_msgdma_dev_data *)dev->data;
struct dma_block_config *dma_block;
int status;
uint32_t control;
/* Nios-II MSGDMA supports only one channel per DMA core */
if (channel != 0U) {
LOG_ERR("invalid channel number");
return -EINVAL;
}
#if MSGDMA_0_CSR_PREFETCHER_ENABLE
if (cfg->block_count > 1) {
LOG_ERR("driver yet add support multiple descriptors");
return -EINVAL;
}
#else
if (cfg->block_count != 1U) {
LOG_ERR("invalid block count!!");
return -EINVAL;
}
#endif
if (cfg->head_block == NULL) {
LOG_ERR("head_block ptr NULL!!");
return -EINVAL;
}
if (cfg->head_block->block_size > MSGDMA_0_DESCRIPTOR_SLAVE_MAX_BYTE) {
LOG_ERR("DMA error: Data size too big: %d",
cfg->head_block->block_size);
return -EINVAL;
}
k_sem_take(&dev_data->sem_lock, K_FOREVER);
dev_data->dma_callback = cfg->dma_callback;
dev_data->user_data = cfg->user_data;
dev_data->direction = cfg->channel_direction;
dma_block = cfg->head_block;
control = ALTERA_MSGDMA_DESCRIPTOR_CONTROL_TRANSFER_COMPLETE_IRQ_MASK |
ALTERA_MSGDMA_DESCRIPTOR_CONTROL_EARLY_TERMINATION_IRQ_MASK;
if (dev_data->direction == MEMORY_TO_MEMORY) {
status = alt_msgdma_construct_standard_mm_to_mm_descriptor(
dev_data->msgdma_dev, &dev_data->desc,
(alt_u32 *)dma_block->source_address,
(alt_u32 *)dma_block->dest_address,
dma_block->block_size,
control);
} else if (dev_data->direction == MEMORY_TO_PERIPHERAL) {
status = alt_msgdma_construct_standard_mm_to_st_descriptor(
dev_data->msgdma_dev, &dev_data->desc,
(alt_u32 *)dma_block->source_address,
dma_block->block_size,
control);
} else if (dev_data->direction == PERIPHERAL_TO_MEMORY) {
status = alt_msgdma_construct_standard_st_to_mm_descriptor(
dev_data->msgdma_dev, &dev_data->desc,
(alt_u32 *)dma_block->dest_address,
dma_block->block_size,
control);
} else {
LOG_ERR("invalid channel direction");
status = -EINVAL;
}
/* Register msgdma callback */
alt_msgdma_register_callback(dev_data->msgdma_dev,
nios2_msgdma_callback,
ALTERA_MSGDMA_CSR_GLOBAL_INTERRUPT_MASK |
ALTERA_MSGDMA_CSR_STOP_ON_ERROR_MASK |
ALTERA_MSGDMA_CSR_STOP_ON_EARLY_TERMINATION_MASK,
dev_data);
/* Clear the IRQ status */
IOWR_ALTERA_MSGDMA_CSR_STATUS(dev_data->msgdma_dev->csr_base,
ALTERA_MSGDMA_CSR_IRQ_SET_MASK);
k_sem_give(&dev_data->sem_lock);
return status;
}
static int nios2_msgdma_transfer_start(const struct device *dev,
uint32_t channel)
{
struct nios2_msgdma_dev_data *cfg = (struct nios2_msgdma_dev_data *)dev->data;
int status;
/* Nios-II MSGDMA supports only one channel per DMA core */
if (channel != 0U) {
LOG_ERR("Invalid channel number");
return -EINVAL;
}
k_sem_take(&cfg->sem_lock, K_FOREVER);
status = alt_msgdma_standard_descriptor_async_transfer(cfg->msgdma_dev,
&cfg->desc);
k_sem_give(&cfg->sem_lock);
if (status < 0) {
LOG_ERR("DMA transfer error (%d)", status);
}
return status;
}
static int nios2_msgdma_transfer_stop(const struct device *dev,
uint32_t channel)
{
struct nios2_msgdma_dev_data *cfg = (struct nios2_msgdma_dev_data *)dev->data;
int ret = -EIO;
uint32_t status;
k_sem_take(&cfg->sem_lock, K_FOREVER);
/* Stop the DMA Dispatcher */
IOWR_ALTERA_MSGDMA_CSR_CONTROL(cfg->msgdma_dev->csr_base,
ALTERA_MSGDMA_CSR_STOP_MASK);
status = IORD_ALTERA_MSGDMA_CSR_STATUS(cfg->msgdma_dev->csr_base);
k_sem_give(&cfg->sem_lock);
if (status & ALTERA_MSGDMA_CSR_STOP_STATE_MASK) {
LOG_DBG("DMA Dispatcher stopped");
ret = 0;
}
LOG_DBG("msgdma csr status Reg: 0x%x", status);
return status;
}
static const struct dma_driver_api nios2_msgdma_driver_api = {
.config = nios2_msgdma_config,
.start = nios2_msgdma_transfer_start,
.stop = nios2_msgdma_transfer_stop,
};
static int nios2_msgdma0_initialize(const struct device *dev)
{
struct nios2_msgdma_dev_data *dev_data = (struct nios2_msgdma_dev_data *)dev->data;
dev_data->dev = dev;
/* Initialize semaphore */
k_sem_init(&dev_data->sem_lock, 1, 1);
alt_msgdma_init(dev_data->msgdma_dev, 0, DT_INST_IRQN(0));
IRQ_CONNECT(DT_INST_IRQN(0), DT_INST_IRQ(0, priority),
nios2_msgdma_isr, DEVICE_DT_INST_GET(0), 0);
irq_enable(DT_INST_IRQN(0));
return 0;
}
ALTERA_MSGDMA_CSR_DESCRIPTOR_SLAVE_INSTANCE(MSGDMA_0, MSGDMA_0_CSR,
MSGDMA_0_DESCRIPTOR_SLAVE, msgdma_dev0)
static struct nios2_msgdma_dev_data dma0_nios2_data = {
.msgdma_dev = &msgdma_dev0,
};
DEVICE_DT_INST_DEFINE(0, &nios2_msgdma0_initialize,
NULL, &dma0_nios2_data, NULL, POST_KERNEL,
CONFIG_DMA_INIT_PRIORITY, &nios2_msgdma_driver_api);