/* * Copyright (c) 2016, Freescale Semiconductor, Inc. * Copyright (c) 2017, NXP * * SPDX-License-Identifier: Apache-2.0 */ #define DT_DRV_COMPAT nxp_kinetis_dspi #include #include #include #include #define LOG_LEVEL CONFIG_SPI_LOG_LEVEL #include LOG_MODULE_REGISTER(spi_mcux_dspi); #include "spi_context.h" struct spi_mcux_config { SPI_Type *base; const struct device *clock_dev; clock_control_subsys_t clock_subsys; void (*irq_config_func)(const struct device *dev); uint32_t pcs_sck_delay; uint32_t sck_pcs_delay; uint32_t transfer_delay; }; struct spi_mcux_data { const struct device *dev; dspi_master_handle_t handle; struct spi_context ctx; size_t transfer_len; }; static int spi_mcux_transfer_next_packet(const struct device *dev) { const struct spi_mcux_config *config = dev->config; struct spi_mcux_data *data = dev->data; SPI_Type *base = config->base; struct spi_context *ctx = &data->ctx; dspi_transfer_t transfer; status_t status; if ((ctx->tx_len == 0) && (ctx->rx_len == 0)) { /* nothing left to rx or tx, we're done! */ spi_context_cs_control(&data->ctx, false); spi_context_complete(&data->ctx, 0); return 0; } transfer.configFlags = kDSPI_MasterCtar0 | kDSPI_MasterPcsContinuous | (ctx->config->slave << DSPI_MASTER_PCS_SHIFT); if (ctx->tx_len == 0) { /* rx only, nothing to tx */ transfer.txData = NULL; transfer.rxData = ctx->rx_buf; transfer.dataSize = ctx->rx_len; } else if (ctx->rx_len == 0) { /* tx only, nothing to rx */ transfer.txData = (uint8_t *) ctx->tx_buf; transfer.rxData = NULL; transfer.dataSize = ctx->tx_len; } else if (ctx->tx_len == ctx->rx_len) { /* rx and tx are the same length */ transfer.txData = (uint8_t *) ctx->tx_buf; transfer.rxData = ctx->rx_buf; transfer.dataSize = ctx->tx_len; } else if (ctx->tx_len > ctx->rx_len) { /* Break up the tx into multiple transfers so we don't have to * rx into a longer intermediate buffer. Leave chip select * active between transfers. */ transfer.txData = (uint8_t *) ctx->tx_buf; transfer.rxData = ctx->rx_buf; transfer.dataSize = ctx->rx_len; transfer.configFlags |= kDSPI_MasterActiveAfterTransfer; } else { /* Break up the rx into multiple transfers so we don't have to * tx from a longer intermediate buffer. Leave chip select * active between transfers. */ transfer.txData = (uint8_t *) ctx->tx_buf; transfer.rxData = ctx->rx_buf; transfer.dataSize = ctx->tx_len; transfer.configFlags |= kDSPI_MasterActiveAfterTransfer; } if (!(ctx->tx_count <= 1 && ctx->rx_count <= 1)) { transfer.configFlags |= kDSPI_MasterActiveAfterTransfer; } data->transfer_len = transfer.dataSize; status = DSPI_MasterTransferNonBlocking(base, &data->handle, &transfer); if (status != kStatus_Success) { LOG_ERR("Transfer could not start"); } return status == kStatus_Success ? 0 : status == kDSPI_Busy ? -EBUSY : -EINVAL; } static void spi_mcux_isr(const struct device *dev) { const struct spi_mcux_config *config = dev->config; struct spi_mcux_data *data = dev->data; SPI_Type *base = config->base; DSPI_MasterTransferHandleIRQ(base, &data->handle); } static void spi_mcux_master_transfer_callback(SPI_Type *base, dspi_master_handle_t *handle, status_t status, void *userData) { struct spi_mcux_data *data = userData; spi_context_update_tx(&data->ctx, 1, data->transfer_len); spi_context_update_rx(&data->ctx, 1, data->transfer_len); spi_mcux_transfer_next_packet(data->dev); } static int spi_mcux_configure(const struct device *dev, const struct spi_config *spi_cfg) { const struct spi_mcux_config *config = dev->config; struct spi_mcux_data *data = dev->data; SPI_Type *base = config->base; dspi_master_config_t master_config; uint32_t clock_freq; uint32_t word_size; dspi_master_ctar_config_t *ctar_config = &master_config.ctarConfig; if (spi_context_configured(&data->ctx, spi_cfg)) { /* This configuration is already in use */ return 0; } DSPI_MasterGetDefaultConfig(&master_config); if (spi_cfg->slave > FSL_FEATURE_DSPI_CHIP_SELECT_COUNT) { LOG_ERR("Slave %d is greater than %d", spi_cfg->slave, FSL_FEATURE_DSPI_CHIP_SELECT_COUNT); return -EINVAL; } word_size = SPI_WORD_SIZE_GET(spi_cfg->operation); if (word_size > FSL_FEATURE_DSPI_MAX_DATA_WIDTH) { LOG_ERR("Word size %d is greater than %d", word_size, FSL_FEATURE_DSPI_MAX_DATA_WIDTH); return -EINVAL; } ctar_config->bitsPerFrame = word_size; ctar_config->cpol = (SPI_MODE_GET(spi_cfg->operation) & SPI_MODE_CPOL) ? kDSPI_ClockPolarityActiveLow : kDSPI_ClockPolarityActiveHigh; ctar_config->cpha = (SPI_MODE_GET(spi_cfg->operation) & SPI_MODE_CPHA) ? kDSPI_ClockPhaseSecondEdge : kDSPI_ClockPhaseFirstEdge; ctar_config->direction = (spi_cfg->operation & SPI_TRANSFER_LSB) ? kDSPI_LsbFirst : kDSPI_MsbFirst; ctar_config->baudRate = spi_cfg->frequency; ctar_config->pcsToSckDelayInNanoSec = config->pcs_sck_delay; ctar_config->lastSckToPcsDelayInNanoSec = config->sck_pcs_delay; ctar_config->betweenTransferDelayInNanoSec = config->transfer_delay; if (clock_control_get_rate(config->clock_dev, config->clock_subsys, &clock_freq)) { return -EINVAL; } DSPI_MasterInit(base, &master_config, clock_freq); DSPI_MasterTransferCreateHandle(base, &data->handle, spi_mcux_master_transfer_callback, data); DSPI_SetDummyData(base, 0); data->ctx.config = spi_cfg; spi_context_cs_configure(&data->ctx); return 0; } static int transceive(const struct device *dev, const struct spi_config *spi_cfg, const struct spi_buf_set *tx_bufs, const struct spi_buf_set *rx_bufs, bool asynchronous, struct k_poll_signal *signal) { struct spi_mcux_data *data = dev->data; int ret; spi_context_lock(&data->ctx, asynchronous, signal, spi_cfg); ret = spi_mcux_configure(dev, spi_cfg); if (ret) { goto out; } spi_context_buffers_setup(&data->ctx, tx_bufs, rx_bufs, 1); spi_context_cs_control(&data->ctx, true); ret = spi_mcux_transfer_next_packet(dev); if (ret) { goto out; } ret = spi_context_wait_for_completion(&data->ctx); out: spi_context_release(&data->ctx, ret); return ret; } static int spi_mcux_transceive(const struct device *dev, const struct spi_config *spi_cfg, const struct spi_buf_set *tx_bufs, const struct spi_buf_set *rx_bufs) { return transceive(dev, spi_cfg, tx_bufs, rx_bufs, false, NULL); } #ifdef CONFIG_SPI_ASYNC static int spi_mcux_transceive_async(const struct device *dev, const struct spi_config *spi_cfg, const struct spi_buf_set *tx_bufs, const struct spi_buf_set *rx_bufs, struct k_poll_signal *async) { return transceive(dev, spi_cfg, tx_bufs, rx_bufs, true, async); } #endif /* CONFIG_SPI_ASYNC */ static int spi_mcux_release(const struct device *dev, const struct spi_config *spi_cfg) { struct spi_mcux_data *data = dev->data; spi_context_unlock_unconditionally(&data->ctx); return 0; } static int spi_mcux_init(const struct device *dev) { const struct spi_mcux_config *config = dev->config; struct spi_mcux_data *data = dev->data; data->dev = dev; config->irq_config_func(dev); spi_context_unlock_unconditionally(&data->ctx); return 0; } static const struct spi_driver_api spi_mcux_driver_api = { .transceive = spi_mcux_transceive, #ifdef CONFIG_SPI_ASYNC .transceive_async = spi_mcux_transceive_async, #endif .release = spi_mcux_release, }; #define SPI_MCUX_DSPI_DEVICE(id) \ static void spi_mcux_config_func_##id(const struct device *dev); \ static const struct spi_mcux_config spi_mcux_config_##id = { \ .base = (SPI_Type *)DT_INST_REG_ADDR(id), \ .clock_dev = DEVICE_DT_GET(DT_INST_CLOCKS_CTLR(id)), \ .clock_subsys = \ (clock_control_subsys_t)DT_INST_CLOCKS_CELL(id, name), \ .irq_config_func = spi_mcux_config_func_##id, \ .pcs_sck_delay = UTIL_AND( \ DT_INST_NODE_HAS_PROP(id, pcs_sck_delay), \ DT_INST_PROP(id, pcs_sck_delay)), \ .sck_pcs_delay = UTIL_AND( \ DT_INST_NODE_HAS_PROP(id, sck_pcs_delay), \ DT_INST_PROP(id, sck_pcs_delay)), \ .transfer_delay = UTIL_AND( \ DT_INST_NODE_HAS_PROP(id, transfer_delay), \ DT_INST_PROP(id, transfer_delay)), \ }; \ static struct spi_mcux_data spi_mcux_data_##id = { \ SPI_CONTEXT_INIT_LOCK(spi_mcux_data_##id, ctx), \ SPI_CONTEXT_INIT_SYNC(spi_mcux_data_##id, ctx), \ }; \ DEVICE_DT_INST_DEFINE(id, \ &spi_mcux_init, \ device_pm_control_nop, \ &spi_mcux_data_##id, \ &spi_mcux_config_##id, \ POST_KERNEL, \ CONFIG_KERNEL_INIT_PRIORITY_DEVICE, \ &spi_mcux_driver_api); \ static void spi_mcux_config_func_##id(const struct device *dev) \ { \ IRQ_CONNECT(DT_INST_IRQN(id), \ DT_INST_IRQ(id, priority), \ spi_mcux_isr, DEVICE_DT_INST_GET(id), \ 0); \ irq_enable(DT_INST_IRQN(id)); \ } DT_INST_FOREACH_STATUS_OKAY(SPI_MCUX_DSPI_DEVICE)