325 lines
8.9 KiB
C
325 lines
8.9 KiB
C
/*
|
|
* Copyright (c) 2016, Freescale Semiconductor, Inc.
|
|
* Copyright (c) 2017, NXP
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*/
|
|
|
|
#define DT_DRV_COMPAT nxp_kinetis_dspi
|
|
|
|
#include <errno.h>
|
|
#include <drivers/spi.h>
|
|
#include <drivers/clock_control.h>
|
|
#include <fsl_dspi.h>
|
|
|
|
#define LOG_LEVEL CONFIG_SPI_LOG_LEVEL
|
|
#include <logging/log.h>
|
|
LOG_MODULE_REGISTER(spi_mcux_dspi);
|
|
|
|
#include "spi_context.h"
|
|
|
|
struct spi_mcux_config {
|
|
SPI_Type *base;
|
|
const struct device *clock_dev;
|
|
clock_control_subsys_t clock_subsys;
|
|
void (*irq_config_func)(const struct device *dev);
|
|
uint32_t pcs_sck_delay;
|
|
uint32_t sck_pcs_delay;
|
|
uint32_t transfer_delay;
|
|
};
|
|
|
|
struct spi_mcux_data {
|
|
const struct device *dev;
|
|
dspi_master_handle_t handle;
|
|
struct spi_context ctx;
|
|
size_t transfer_len;
|
|
};
|
|
|
|
static int spi_mcux_transfer_next_packet(const struct device *dev)
|
|
{
|
|
const struct spi_mcux_config *config = dev->config;
|
|
struct spi_mcux_data *data = dev->data;
|
|
SPI_Type *base = config->base;
|
|
struct spi_context *ctx = &data->ctx;
|
|
dspi_transfer_t transfer;
|
|
status_t status;
|
|
|
|
if ((ctx->tx_len == 0) && (ctx->rx_len == 0)) {
|
|
/* nothing left to rx or tx, we're done! */
|
|
spi_context_cs_control(&data->ctx, false);
|
|
spi_context_complete(&data->ctx, 0);
|
|
return 0;
|
|
}
|
|
|
|
transfer.configFlags = kDSPI_MasterCtar0 | kDSPI_MasterPcsContinuous |
|
|
(ctx->config->slave << DSPI_MASTER_PCS_SHIFT);
|
|
|
|
if (ctx->tx_len == 0) {
|
|
/* rx only, nothing to tx */
|
|
transfer.txData = NULL;
|
|
transfer.rxData = ctx->rx_buf;
|
|
transfer.dataSize = ctx->rx_len;
|
|
} else if (ctx->rx_len == 0) {
|
|
/* tx only, nothing to rx */
|
|
transfer.txData = (uint8_t *) ctx->tx_buf;
|
|
transfer.rxData = NULL;
|
|
transfer.dataSize = ctx->tx_len;
|
|
} else if (ctx->tx_len == ctx->rx_len) {
|
|
/* rx and tx are the same length */
|
|
transfer.txData = (uint8_t *) ctx->tx_buf;
|
|
transfer.rxData = ctx->rx_buf;
|
|
transfer.dataSize = ctx->tx_len;
|
|
} else if (ctx->tx_len > ctx->rx_len) {
|
|
/* Break up the tx into multiple transfers so we don't have to
|
|
* rx into a longer intermediate buffer. Leave chip select
|
|
* active between transfers.
|
|
*/
|
|
transfer.txData = (uint8_t *) ctx->tx_buf;
|
|
transfer.rxData = ctx->rx_buf;
|
|
transfer.dataSize = ctx->rx_len;
|
|
transfer.configFlags |= kDSPI_MasterActiveAfterTransfer;
|
|
} else {
|
|
/* Break up the rx into multiple transfers so we don't have to
|
|
* tx from a longer intermediate buffer. Leave chip select
|
|
* active between transfers.
|
|
*/
|
|
transfer.txData = (uint8_t *) ctx->tx_buf;
|
|
transfer.rxData = ctx->rx_buf;
|
|
transfer.dataSize = ctx->tx_len;
|
|
transfer.configFlags |= kDSPI_MasterActiveAfterTransfer;
|
|
}
|
|
|
|
if (!(ctx->tx_count <= 1 && ctx->rx_count <= 1)) {
|
|
transfer.configFlags |= kDSPI_MasterActiveAfterTransfer;
|
|
}
|
|
|
|
data->transfer_len = transfer.dataSize;
|
|
|
|
status = DSPI_MasterTransferNonBlocking(base, &data->handle, &transfer);
|
|
if (status != kStatus_Success) {
|
|
LOG_ERR("Transfer could not start");
|
|
}
|
|
|
|
return status == kStatus_Success ? 0 :
|
|
status == kDSPI_Busy ? -EBUSY : -EINVAL;
|
|
}
|
|
|
|
static void spi_mcux_isr(const struct device *dev)
|
|
{
|
|
const struct spi_mcux_config *config = dev->config;
|
|
struct spi_mcux_data *data = dev->data;
|
|
SPI_Type *base = config->base;
|
|
|
|
DSPI_MasterTransferHandleIRQ(base, &data->handle);
|
|
}
|
|
|
|
static void spi_mcux_master_transfer_callback(SPI_Type *base,
|
|
dspi_master_handle_t *handle, status_t status, void *userData)
|
|
{
|
|
struct spi_mcux_data *data = userData;
|
|
|
|
spi_context_update_tx(&data->ctx, 1, data->transfer_len);
|
|
spi_context_update_rx(&data->ctx, 1, data->transfer_len);
|
|
|
|
spi_mcux_transfer_next_packet(data->dev);
|
|
}
|
|
|
|
static int spi_mcux_configure(const struct device *dev,
|
|
const struct spi_config *spi_cfg)
|
|
{
|
|
const struct spi_mcux_config *config = dev->config;
|
|
struct spi_mcux_data *data = dev->data;
|
|
SPI_Type *base = config->base;
|
|
dspi_master_config_t master_config;
|
|
uint32_t clock_freq;
|
|
uint32_t word_size;
|
|
|
|
dspi_master_ctar_config_t *ctar_config = &master_config.ctarConfig;
|
|
|
|
if (spi_context_configured(&data->ctx, spi_cfg)) {
|
|
/* This configuration is already in use */
|
|
return 0;
|
|
}
|
|
|
|
DSPI_MasterGetDefaultConfig(&master_config);
|
|
|
|
if (spi_cfg->slave > FSL_FEATURE_DSPI_CHIP_SELECT_COUNT) {
|
|
LOG_ERR("Slave %d is greater than %d",
|
|
spi_cfg->slave, FSL_FEATURE_DSPI_CHIP_SELECT_COUNT);
|
|
return -EINVAL;
|
|
}
|
|
|
|
word_size = SPI_WORD_SIZE_GET(spi_cfg->operation);
|
|
if (word_size > FSL_FEATURE_DSPI_MAX_DATA_WIDTH) {
|
|
LOG_ERR("Word size %d is greater than %d",
|
|
word_size, FSL_FEATURE_DSPI_MAX_DATA_WIDTH);
|
|
return -EINVAL;
|
|
}
|
|
|
|
ctar_config->bitsPerFrame = word_size;
|
|
|
|
ctar_config->cpol =
|
|
(SPI_MODE_GET(spi_cfg->operation) & SPI_MODE_CPOL)
|
|
? kDSPI_ClockPolarityActiveLow
|
|
: kDSPI_ClockPolarityActiveHigh;
|
|
|
|
ctar_config->cpha =
|
|
(SPI_MODE_GET(spi_cfg->operation) & SPI_MODE_CPHA)
|
|
? kDSPI_ClockPhaseSecondEdge
|
|
: kDSPI_ClockPhaseFirstEdge;
|
|
|
|
ctar_config->direction =
|
|
(spi_cfg->operation & SPI_TRANSFER_LSB)
|
|
? kDSPI_LsbFirst
|
|
: kDSPI_MsbFirst;
|
|
|
|
ctar_config->baudRate = spi_cfg->frequency;
|
|
|
|
ctar_config->pcsToSckDelayInNanoSec = config->pcs_sck_delay;
|
|
ctar_config->lastSckToPcsDelayInNanoSec = config->sck_pcs_delay;
|
|
ctar_config->betweenTransferDelayInNanoSec = config->transfer_delay;
|
|
|
|
if (clock_control_get_rate(config->clock_dev, config->clock_subsys,
|
|
&clock_freq)) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
DSPI_MasterInit(base, &master_config, clock_freq);
|
|
|
|
DSPI_MasterTransferCreateHandle(base, &data->handle,
|
|
spi_mcux_master_transfer_callback,
|
|
data);
|
|
|
|
DSPI_SetDummyData(base, 0);
|
|
|
|
data->ctx.config = spi_cfg;
|
|
spi_context_cs_configure(&data->ctx);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int transceive(const struct device *dev,
|
|
const struct spi_config *spi_cfg,
|
|
const struct spi_buf_set *tx_bufs,
|
|
const struct spi_buf_set *rx_bufs,
|
|
bool asynchronous,
|
|
struct k_poll_signal *signal)
|
|
{
|
|
struct spi_mcux_data *data = dev->data;
|
|
int ret;
|
|
|
|
spi_context_lock(&data->ctx, asynchronous, signal, spi_cfg);
|
|
|
|
ret = spi_mcux_configure(dev, spi_cfg);
|
|
if (ret) {
|
|
goto out;
|
|
}
|
|
|
|
spi_context_buffers_setup(&data->ctx, tx_bufs, rx_bufs, 1);
|
|
|
|
spi_context_cs_control(&data->ctx, true);
|
|
|
|
ret = spi_mcux_transfer_next_packet(dev);
|
|
if (ret) {
|
|
goto out;
|
|
}
|
|
|
|
ret = spi_context_wait_for_completion(&data->ctx);
|
|
out:
|
|
spi_context_release(&data->ctx, ret);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int spi_mcux_transceive(const struct device *dev,
|
|
const struct spi_config *spi_cfg,
|
|
const struct spi_buf_set *tx_bufs,
|
|
const struct spi_buf_set *rx_bufs)
|
|
{
|
|
return transceive(dev, spi_cfg, tx_bufs, rx_bufs, false, NULL);
|
|
}
|
|
|
|
#ifdef CONFIG_SPI_ASYNC
|
|
static int spi_mcux_transceive_async(const struct device *dev,
|
|
const struct spi_config *spi_cfg,
|
|
const struct spi_buf_set *tx_bufs,
|
|
const struct spi_buf_set *rx_bufs,
|
|
struct k_poll_signal *async)
|
|
{
|
|
return transceive(dev, spi_cfg, tx_bufs, rx_bufs, true, async);
|
|
}
|
|
#endif /* CONFIG_SPI_ASYNC */
|
|
|
|
static int spi_mcux_release(const struct device *dev,
|
|
const struct spi_config *spi_cfg)
|
|
{
|
|
struct spi_mcux_data *data = dev->data;
|
|
|
|
spi_context_unlock_unconditionally(&data->ctx);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int spi_mcux_init(const struct device *dev)
|
|
{
|
|
const struct spi_mcux_config *config = dev->config;
|
|
struct spi_mcux_data *data = dev->data;
|
|
|
|
data->dev = dev;
|
|
|
|
config->irq_config_func(dev);
|
|
|
|
spi_context_unlock_unconditionally(&data->ctx);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct spi_driver_api spi_mcux_driver_api = {
|
|
.transceive = spi_mcux_transceive,
|
|
#ifdef CONFIG_SPI_ASYNC
|
|
.transceive_async = spi_mcux_transceive_async,
|
|
#endif
|
|
.release = spi_mcux_release,
|
|
};
|
|
|
|
#define SPI_MCUX_DSPI_DEVICE(id) \
|
|
static void spi_mcux_config_func_##id(const struct device *dev); \
|
|
static const struct spi_mcux_config spi_mcux_config_##id = { \
|
|
.base = (SPI_Type *)DT_INST_REG_ADDR(id), \
|
|
.clock_dev = DEVICE_DT_GET(DT_INST_CLOCKS_CTLR(id)), \
|
|
.clock_subsys = \
|
|
(clock_control_subsys_t)DT_INST_CLOCKS_CELL(id, name), \
|
|
.irq_config_func = spi_mcux_config_func_##id, \
|
|
.pcs_sck_delay = UTIL_AND( \
|
|
DT_INST_NODE_HAS_PROP(id, pcs_sck_delay), \
|
|
DT_INST_PROP(id, pcs_sck_delay)), \
|
|
.sck_pcs_delay = UTIL_AND( \
|
|
DT_INST_NODE_HAS_PROP(id, sck_pcs_delay), \
|
|
DT_INST_PROP(id, sck_pcs_delay)), \
|
|
.transfer_delay = UTIL_AND( \
|
|
DT_INST_NODE_HAS_PROP(id, transfer_delay), \
|
|
DT_INST_PROP(id, transfer_delay)), \
|
|
}; \
|
|
static struct spi_mcux_data spi_mcux_data_##id = { \
|
|
SPI_CONTEXT_INIT_LOCK(spi_mcux_data_##id, ctx), \
|
|
SPI_CONTEXT_INIT_SYNC(spi_mcux_data_##id, ctx), \
|
|
}; \
|
|
DEVICE_DT_INST_DEFINE(id, \
|
|
&spi_mcux_init, \
|
|
device_pm_control_nop, \
|
|
&spi_mcux_data_##id, \
|
|
&spi_mcux_config_##id, \
|
|
POST_KERNEL, \
|
|
CONFIG_KERNEL_INIT_PRIORITY_DEVICE, \
|
|
&spi_mcux_driver_api); \
|
|
static void spi_mcux_config_func_##id(const struct device *dev) \
|
|
{ \
|
|
IRQ_CONNECT(DT_INST_IRQN(id), \
|
|
DT_INST_IRQ(id, priority), \
|
|
spi_mcux_isr, DEVICE_DT_INST_GET(id), \
|
|
0); \
|
|
irq_enable(DT_INST_IRQN(id)); \
|
|
}
|
|
|
|
DT_INST_FOREACH_STATUS_OKAY(SPI_MCUX_DSPI_DEVICE)
|