576 lines
16 KiB
C
576 lines
16 KiB
C
/*
|
|
* Copyright (c) 2020 NXP
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*/
|
|
|
|
/**
|
|
* @brief Common part of DMA drivers for some NXP SoC.
|
|
*/
|
|
#include <zephyr/kernel.h>
|
|
#include <zephyr/device.h>
|
|
#include <soc.h>
|
|
#include <zephyr/drivers/dma.h>
|
|
#include <fsl_dma.h>
|
|
#include <fsl_inputmux.h>
|
|
#include <zephyr/logging/log.h>
|
|
|
|
#define DT_DRV_COMPAT nxp_lpc_dma
|
|
|
|
LOG_MODULE_REGISTER(dma_mcux_lpc, CONFIG_DMA_LOG_LEVEL);
|
|
|
|
struct dma_mcux_lpc_config {
|
|
DMA_Type *base;
|
|
uint32_t num_of_channels;
|
|
void (*irq_config_func)(const struct device *dev);
|
|
};
|
|
|
|
struct call_back {
|
|
dma_descriptor_t *dma_descriptor_table;
|
|
dma_handle_t dma_handle;
|
|
const struct device *dev;
|
|
void *user_data;
|
|
dma_callback_t dma_callback;
|
|
enum dma_channel_direction dir;
|
|
uint32_t descriptor_index;
|
|
uint32_t descriptor_used;
|
|
dma_descriptor_t *curr_transfer;
|
|
uint32_t width;
|
|
bool busy;
|
|
};
|
|
|
|
struct dma_mcux_lpc_dma_data {
|
|
struct call_back *data_cb;
|
|
uint32_t *channel_index;
|
|
uint32_t num_channels_used;
|
|
};
|
|
|
|
#define DEV_BASE(dev) \
|
|
((DMA_Type *)((const struct dma_mcux_lpc_config *const)(dev)->config)->base)
|
|
|
|
#define DEV_CHANNEL_DATA(dev, ch) \
|
|
((struct call_back *)(&(((struct dma_mcux_lpc_dma_data *)dev->data)->data_cb[ch])))
|
|
|
|
#define DEV_DMA_HANDLE(dev, ch) \
|
|
((dma_handle_t *)(&(DEV_CHANNEL_DATA(dev, ch)->dma_handle)))
|
|
|
|
static void nxp_lpc_dma_callback(dma_handle_t *handle, void *param,
|
|
bool transferDone, uint32_t intmode)
|
|
{
|
|
int ret = 1;
|
|
struct call_back *data = (struct call_back *)param;
|
|
uint32_t channel = handle->channel;
|
|
|
|
if (transferDone) {
|
|
ret = 0;
|
|
}
|
|
|
|
if (intmode == kDMA_IntError) {
|
|
DMA_AbortTransfer(handle);
|
|
}
|
|
|
|
data->dma_callback(data->dev, data->user_data, channel, ret);
|
|
}
|
|
|
|
/* Handles DMA interrupts and dispatches to the individual channel */
|
|
static void dma_mcux_lpc_irq_handler(const struct device *dev)
|
|
{
|
|
DMA_IRQHandle(DEV_BASE(dev));
|
|
/*
|
|
* Add for ARM errata 838869, affects Cortex-M4, Cortex-M4F Store
|
|
* immediate overlapping exception return operation might vector
|
|
* to incorrect interrupt
|
|
*/
|
|
#if defined __CORTEX_M && (__CORTEX_M == 4U)
|
|
__DSB();
|
|
#endif
|
|
}
|
|
|
|
/* Configure a channel */
|
|
static int dma_mcux_lpc_configure(const struct device *dev, uint32_t channel,
|
|
struct dma_config *config)
|
|
{
|
|
const struct dma_mcux_lpc_config *dev_config = dev->config;
|
|
dma_handle_t *p_handle;
|
|
uint32_t xferConfig = 0U;
|
|
struct call_back *data;
|
|
struct dma_mcux_lpc_dma_data *dma_data = dev->data;
|
|
struct dma_block_config *block_config = config->head_block;
|
|
uint32_t virtual_channel;
|
|
uint32_t total_dma_channels;
|
|
uint8_t src_inc, dst_inc;
|
|
bool is_periph = true;
|
|
|
|
if (NULL == dev || NULL == config) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Check if have a free slot to store DMA channel data */
|
|
if (dma_data->num_channels_used > dev_config->num_of_channels) {
|
|
LOG_ERR("out of DMA channel %d", channel);
|
|
return -EINVAL;
|
|
}
|
|
|
|
#if defined FSL_FEATURE_DMA_NUMBER_OF_CHANNELS
|
|
total_dma_channels = FSL_FEATURE_DMA_NUMBER_OF_CHANNELS;
|
|
#else
|
|
total_dma_channels = FSL_FEATURE_DMA_NUMBER_OF_CHANNELSn(DEV_BASE(dev));
|
|
#endif
|
|
|
|
/* Check if the dma channel number is valid */
|
|
if (channel >= total_dma_channels) {
|
|
LOG_ERR("invalid DMA channel number %d", channel);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (config->source_data_size != 4U &&
|
|
config->source_data_size != 2U &&
|
|
config->source_data_size != 1U) {
|
|
LOG_ERR("Source unit size error, %d", config->source_data_size);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (config->dest_data_size != 4U &&
|
|
config->dest_data_size != 2U &&
|
|
config->dest_data_size != 1U) {
|
|
LOG_ERR("Dest unit size error, %d", config->dest_data_size);
|
|
return -EINVAL;
|
|
}
|
|
|
|
switch (config->channel_direction) {
|
|
case MEMORY_TO_MEMORY:
|
|
is_periph = false;
|
|
src_inc = 1;
|
|
dst_inc = 1;
|
|
break;
|
|
case MEMORY_TO_PERIPHERAL:
|
|
src_inc = 1;
|
|
dst_inc = 0;
|
|
break;
|
|
case PERIPHERAL_TO_MEMORY:
|
|
src_inc = 0;
|
|
dst_inc = 1;
|
|
break;
|
|
default:
|
|
LOG_ERR("not support transfer direction");
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* If needed, allocate a slot to store dma channel data */
|
|
if (dma_data->channel_index[channel] == -1) {
|
|
dma_data->channel_index[channel] = dma_data->num_channels_used;
|
|
dma_data->num_channels_used++;
|
|
}
|
|
|
|
/* Get the slot number that has the dma channel data */
|
|
virtual_channel = dma_data->channel_index[channel];
|
|
/* dma channel data */
|
|
p_handle = DEV_DMA_HANDLE(dev, virtual_channel);
|
|
data = DEV_CHANNEL_DATA(dev, virtual_channel);
|
|
|
|
data->dir = config->channel_direction;
|
|
|
|
if (data->busy) {
|
|
DMA_AbortTransfer(p_handle);
|
|
}
|
|
DMA_CreateHandle(p_handle, DEV_BASE(dev), channel);
|
|
DMA_SetCallback(p_handle, nxp_lpc_dma_callback, (void *)data);
|
|
|
|
LOG_DBG("channel is %d", p_handle->channel);
|
|
|
|
if (config->source_chaining_en && config->dest_chaining_en) {
|
|
LOG_DBG("link dma out 0 to channel %d", config->linked_channel);
|
|
/* Link DMA_OTRIG 0 to channel */
|
|
INPUTMUX_AttachSignal(INPUTMUX, 0, config->linked_channel);
|
|
}
|
|
|
|
/* In case of SPI Transmit where no data is transmitted, we queue
|
|
* dummy data to the buffer that does not require the source or
|
|
* destination address to change
|
|
*/
|
|
if ((block_config->source_addr_adj == DMA_ADDR_ADJ_NO_CHANGE) &&
|
|
(block_config->dest_addr_adj == DMA_ADDR_ADJ_NO_CHANGE)) {
|
|
src_inc = 0;
|
|
dst_inc = 0;
|
|
}
|
|
|
|
data->descriptor_used = 0;
|
|
|
|
data->curr_transfer = NULL;
|
|
|
|
if (block_config->source_gather_en || block_config->dest_scatter_en) {
|
|
if (config->block_count > CONFIG_DMA_LINK_QUEUE_SIZE) {
|
|
LOG_ERR("please config DMA_LINK_QUEUE_SIZE as %d",
|
|
config->block_count);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Allocate the descriptor table structures if needed */
|
|
if (data->dma_descriptor_table == NULL) {
|
|
data->dma_descriptor_table = k_malloc(CONFIG_DMA_LINK_QUEUE_SIZE *
|
|
(sizeof(dma_descriptor_t) +
|
|
FSL_FEATURE_DMA_LINK_DESCRIPTOR_ALIGN_SIZE));
|
|
|
|
if (!data->dma_descriptor_table) {
|
|
LOG_ERR("HEAP_MEM_POOL_SIZE is too small");
|
|
return -ENOMEM;
|
|
}
|
|
}
|
|
|
|
dma_descriptor_t *next_transfer;
|
|
uint32_t dest_width = config->dest_data_size;
|
|
|
|
if (block_config->next_block == NULL) {
|
|
/* Single block transfer, no additional descriptors required */
|
|
data->curr_transfer = NULL;
|
|
} else {
|
|
/* Ensure queued descriptor is aligned */
|
|
data->curr_transfer = (dma_descriptor_t *)ROUND_UP(
|
|
data->dma_descriptor_table,
|
|
FSL_FEATURE_DMA_LINK_DESCRIPTOR_ALIGN_SIZE);
|
|
}
|
|
|
|
/* Enable interrupt */
|
|
xferConfig = DMA_CHANNEL_XFER(1UL, 0UL, 1UL, 0UL,
|
|
dest_width,
|
|
src_inc,
|
|
dst_inc,
|
|
block_config->block_size);
|
|
|
|
DMA_SubmitChannelTransferParameter(p_handle,
|
|
xferConfig,
|
|
(void *)block_config->source_address,
|
|
(void *)block_config->dest_address,
|
|
(void *)data->curr_transfer);
|
|
|
|
/* Get the next block and start queuing descriptors */
|
|
block_config = block_config->next_block;
|
|
|
|
while (block_config != NULL) {
|
|
next_transfer = data->curr_transfer + sizeof(dma_descriptor_t);
|
|
|
|
/* Ensure descriptor is aligned */
|
|
next_transfer = (dma_descriptor_t *)ROUND_UP(
|
|
next_transfer,
|
|
FSL_FEATURE_DMA_LINK_DESCRIPTOR_ALIGN_SIZE);
|
|
|
|
/* SPI TX transfers need to queue a DMA descriptor to
|
|
* indicate an end of transfer. Source or destination
|
|
* address does not need to be change for these
|
|
* transactions and the transfer width is 4 bytes
|
|
*/
|
|
if ((block_config->source_addr_adj == DMA_ADDR_ADJ_NO_CHANGE) &&
|
|
(block_config->dest_addr_adj == DMA_ADDR_ADJ_NO_CHANGE)) {
|
|
src_inc = 0;
|
|
dst_inc = 0;
|
|
dest_width = sizeof(uint32_t);
|
|
next_transfer = NULL;
|
|
}
|
|
|
|
/* Set interrupt to be true for the descriptor */
|
|
xferConfig = DMA_CHANNEL_XFER(1UL, 0UL, 1U, 0U,
|
|
dest_width,
|
|
src_inc,
|
|
dst_inc,
|
|
block_config->block_size);
|
|
|
|
DMA_SetupDescriptor(data->curr_transfer,
|
|
xferConfig,
|
|
(void *)block_config->source_address,
|
|
(void *)block_config->dest_address,
|
|
(void *)next_transfer);
|
|
|
|
block_config = block_config->next_block;
|
|
data->curr_transfer = next_transfer;
|
|
data->descriptor_used++;
|
|
}
|
|
|
|
if (data->curr_transfer != NULL) {
|
|
/* Set a descriptor pointing to the start of the chain */
|
|
block_config = config->head_block;
|
|
next_transfer = data->dma_descriptor_table;
|
|
|
|
/* Ensure descriptor is aligned */
|
|
next_transfer = (dma_descriptor_t *)ROUND_UP(
|
|
next_transfer,
|
|
FSL_FEATURE_DMA_LINK_DESCRIPTOR_ALIGN_SIZE);
|
|
|
|
xferConfig = DMA_CHANNEL_XFER(1UL, 0UL, 1U, 0U,
|
|
dest_width,
|
|
src_inc,
|
|
dst_inc,
|
|
block_config->block_size);
|
|
|
|
DMA_SetupDescriptor(data->curr_transfer,
|
|
xferConfig,
|
|
(void *)block_config->source_address,
|
|
(void *)block_config->dest_address,
|
|
(void *)next_transfer);
|
|
|
|
data->descriptor_used++;
|
|
/* Set chain index to last descriptor entry that was added */
|
|
data->descriptor_index = data->descriptor_used;
|
|
}
|
|
} else {
|
|
/* block_count shall be 1 */
|
|
/* Only one buffer, enable interrupt */
|
|
xferConfig = DMA_CHANNEL_XFER(0UL, 0UL, 1UL, 0UL,
|
|
config->dest_data_size,
|
|
src_inc,
|
|
dst_inc,
|
|
block_config->block_size);
|
|
DMA_SubmitChannelTransferParameter(p_handle,
|
|
xferConfig,
|
|
(void *)block_config->source_address,
|
|
(void *)block_config->dest_address,
|
|
NULL);
|
|
}
|
|
|
|
if (is_periph) {
|
|
DMA_EnableChannelPeriphRq(p_handle->base, p_handle->channel);
|
|
} else {
|
|
DMA_DisableChannelPeriphRq(p_handle->base, p_handle->channel);
|
|
}
|
|
|
|
data->width = config->dest_data_size;
|
|
data->busy = false;
|
|
if (config->dma_callback) {
|
|
LOG_DBG("INSTALL call back on channel %d", channel);
|
|
data->user_data = config->user_data;
|
|
data->dma_callback = config->dma_callback;
|
|
data->dev = dev;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int dma_mcux_lpc_start(const struct device *dev, uint32_t channel)
|
|
{
|
|
struct dma_mcux_lpc_dma_data *dev_data = dev->data;
|
|
uint32_t virtual_channel = dev_data->channel_index[channel];
|
|
struct call_back *data = DEV_CHANNEL_DATA(dev, virtual_channel);
|
|
|
|
LOG_DBG("START TRANSFER");
|
|
LOG_DBG("DMA CTRL 0x%x", DEV_BASE(dev)->CTRL);
|
|
data->busy = true;
|
|
DMA_StartTransfer(DEV_DMA_HANDLE(dev, virtual_channel));
|
|
return 0;
|
|
}
|
|
|
|
static int dma_mcux_lpc_stop(const struct device *dev, uint32_t channel)
|
|
{
|
|
struct dma_mcux_lpc_dma_data *dev_data = dev->data;
|
|
uint32_t virtual_channel = dev_data->channel_index[channel];
|
|
struct call_back *data = DEV_CHANNEL_DATA(dev, virtual_channel);
|
|
|
|
if (!data->busy) {
|
|
return 0;
|
|
}
|
|
DMA_AbortTransfer(DEV_DMA_HANDLE(dev, virtual_channel));
|
|
/* Free any memory allocated for DMA descriptors */
|
|
if (data->dma_descriptor_table != NULL) {
|
|
k_free(data->dma_descriptor_table);
|
|
data->dma_descriptor_table = NULL;
|
|
}
|
|
data->busy = false;
|
|
return 0;
|
|
}
|
|
|
|
static int dma_mcux_lpc_reload(const struct device *dev, uint32_t channel,
|
|
uint32_t src, uint32_t dst, size_t size)
|
|
{
|
|
struct dma_mcux_lpc_dma_data *dev_data = dev->data;
|
|
uint32_t virtual_channel = dev_data->channel_index[channel];
|
|
struct call_back *data = DEV_CHANNEL_DATA(dev, virtual_channel);
|
|
uint8_t src_inc, dst_inc;
|
|
uint32_t xferConfig = 0U;
|
|
dma_descriptor_t *next_transfer;
|
|
|
|
switch (data->dir) {
|
|
case MEMORY_TO_MEMORY:
|
|
src_inc = 1;
|
|
dst_inc = 1;
|
|
break;
|
|
case MEMORY_TO_PERIPHERAL:
|
|
src_inc = 1;
|
|
dst_inc = 0;
|
|
break;
|
|
case PERIPHERAL_TO_MEMORY:
|
|
src_inc = 0;
|
|
dst_inc = 1;
|
|
break;
|
|
default:
|
|
LOG_ERR("not support transfer direction");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (data->descriptor_used == 0) {
|
|
dma_handle_t *p_handle;
|
|
|
|
p_handle = DEV_DMA_HANDLE(dev, virtual_channel);
|
|
|
|
/* Only one buffer, enable interrupt */
|
|
xferConfig = DMA_CHANNEL_XFER(0UL, 0UL, 1UL, 0UL,
|
|
data->width,
|
|
src_inc,
|
|
dst_inc,
|
|
size);
|
|
DMA_SubmitChannelTransferParameter(p_handle,
|
|
xferConfig,
|
|
(void *)src,
|
|
(void *)dst,
|
|
NULL);
|
|
} else {
|
|
if (data->descriptor_index == data->descriptor_used) {
|
|
/* Reset to start of the descriptor table chain */
|
|
next_transfer = data->dma_descriptor_table;
|
|
data->descriptor_index = 1;
|
|
} else {
|
|
next_transfer = data->curr_transfer + sizeof(dma_descriptor_t);
|
|
data->descriptor_index++;
|
|
}
|
|
/* Ensure descriptor is aligned */
|
|
next_transfer = (dma_descriptor_t *)ROUND_UP(
|
|
next_transfer,
|
|
FSL_FEATURE_DMA_LINK_DESCRIPTOR_ALIGN_SIZE);
|
|
|
|
xferConfig = DMA_CHANNEL_XFER(1UL, 0UL, 1UL, 0UL,
|
|
data->width,
|
|
src_inc,
|
|
dst_inc,
|
|
size);
|
|
|
|
DMA_SetupDescriptor(data->curr_transfer,
|
|
xferConfig,
|
|
(void *)src,
|
|
(void *)dst,
|
|
(void *)next_transfer);
|
|
|
|
data->curr_transfer = next_transfer;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int dma_mcux_lpc_get_status(const struct device *dev, uint32_t channel,
|
|
struct dma_status *status)
|
|
{
|
|
struct dma_mcux_lpc_dma_data *dev_data = dev->data;
|
|
uint32_t virtual_channel = dev_data->channel_index[channel];
|
|
struct call_back *data = DEV_CHANNEL_DATA(dev, virtual_channel);
|
|
|
|
if (data->busy) {
|
|
status->busy = true;
|
|
status->pending_length = DMA_GetRemainingBytes(DEV_BASE(dev), channel);
|
|
} else {
|
|
status->busy = false;
|
|
status->pending_length = 0;
|
|
}
|
|
status->dir = data->dir;
|
|
LOG_DBG("DMA CR 0x%x", DEV_BASE(dev)->CTRL);
|
|
LOG_DBG("DMA INT 0x%x", DEV_BASE(dev)->INTSTAT);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int dma_mcux_lpc_init(const struct device *dev)
|
|
{
|
|
const struct dma_mcux_lpc_config *config = dev->config;
|
|
struct dma_mcux_lpc_dma_data *data = dev->data;
|
|
int size_channel_data;
|
|
int total_dma_channels;
|
|
|
|
/* Array to store DMA channel data */
|
|
size_channel_data =
|
|
sizeof(struct call_back) * config->num_of_channels;
|
|
data->data_cb = k_malloc(size_channel_data);
|
|
if (!data->data_cb) {
|
|
LOG_ERR("HEAP_MEM_POOL_SIZE is too small");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
memset(data->data_cb, 0, size_channel_data);
|
|
|
|
#if defined FSL_FEATURE_DMA_NUMBER_OF_CHANNELS
|
|
total_dma_channels = FSL_FEATURE_DMA_NUMBER_OF_CHANNELS;
|
|
#else
|
|
total_dma_channels = FSL_FEATURE_DMA_NUMBER_OF_CHANNELSn(DEV_BASE(dev));
|
|
#endif
|
|
/*
|
|
* This array is used to hold the index associated with the array
|
|
* holding channel data
|
|
*/
|
|
data->channel_index = k_malloc(sizeof(uint32_t) * total_dma_channels);
|
|
if (!data->channel_index) {
|
|
LOG_ERR("HEAP_MEM_POOL_SIZE is too small");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/*
|
|
* Initialize to -1 to indicate dma channel does not have a slot
|
|
* assigned to store dma channel data
|
|
*/
|
|
for (int i = 0; i < total_dma_channels; i++) {
|
|
data->channel_index[i] = -1;
|
|
}
|
|
|
|
data->num_channels_used = 0;
|
|
|
|
|
|
DMA_Init(DEV_BASE(dev));
|
|
INPUTMUX_Init(INPUTMUX);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct dma_driver_api dma_mcux_lpc_api = {
|
|
.config = dma_mcux_lpc_configure,
|
|
.start = dma_mcux_lpc_start,
|
|
.stop = dma_mcux_lpc_stop,
|
|
.reload = dma_mcux_lpc_reload,
|
|
.get_status = dma_mcux_lpc_get_status,
|
|
};
|
|
|
|
#define DMA_MCUX_LPC_CONFIG_FUNC(n) \
|
|
static void dma_mcux_lpc_config_func_##n(const struct device *dev) \
|
|
{ \
|
|
IRQ_CONNECT(DT_INST_IRQN(n), \
|
|
DT_INST_IRQ(n, priority), \
|
|
dma_mcux_lpc_irq_handler, DEVICE_DT_INST_GET(n), 0);\
|
|
\
|
|
irq_enable(DT_INST_IRQN(n)); \
|
|
}
|
|
#define DMA_MCUX_LPC_IRQ_CFG_FUNC_INIT(n) \
|
|
.irq_config_func = dma_mcux_lpc_config_func_##n
|
|
#define DMA_MCUX_LPC_INIT_CFG(n) \
|
|
DMA_MCUX_LPC_DECLARE_CFG(n, \
|
|
DMA_MCUX_LPC_IRQ_CFG_FUNC_INIT(n))
|
|
|
|
#define DMA_MCUX_LPC_DECLARE_CFG(n, IRQ_FUNC_INIT) \
|
|
static const struct dma_mcux_lpc_config dma_##n##_config = { \
|
|
.base = (DMA_Type *)DT_INST_REG_ADDR(n), \
|
|
.num_of_channels = DT_INST_PROP(n, dma_channels), \
|
|
IRQ_FUNC_INIT \
|
|
}
|
|
|
|
#define DMA_INIT(n) \
|
|
\
|
|
static const struct dma_mcux_lpc_config dma_##n##_config;\
|
|
\
|
|
static struct dma_mcux_lpc_dma_data dma_data_##n = { \
|
|
.data_cb = NULL, \
|
|
}; \
|
|
\
|
|
DEVICE_DT_INST_DEFINE(n, \
|
|
&dma_mcux_lpc_init, \
|
|
NULL, \
|
|
&dma_data_##n, &dma_##n##_config,\
|
|
POST_KERNEL, CONFIG_DMA_INIT_PRIORITY, \
|
|
&dma_mcux_lpc_api); \
|
|
\
|
|
DMA_MCUX_LPC_CONFIG_FUNC(n) \
|
|
\
|
|
DMA_MCUX_LPC_INIT_CFG(n);
|
|
|
|
DT_INST_FOREACH_STATUS_OKAY(DMA_INIT)
|