384 lines
9.6 KiB
C
384 lines
9.6 KiB
C
/*
|
|
* Copyright (c) 2021 BrainCo Inc.
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*/
|
|
|
|
#define DT_DRV_COMPAT gd_gd32_spi
|
|
|
|
#include <errno.h>
|
|
#include <zephyr/kernel.h>
|
|
#include <zephyr/drivers/clock_control.h>
|
|
#include <zephyr/drivers/clock_control/gd32.h>
|
|
#include <zephyr/drivers/pinctrl.h>
|
|
#include <zephyr/drivers/reset.h>
|
|
#include <zephyr/drivers/spi.h>
|
|
|
|
#include <gd32_spi.h>
|
|
|
|
#include <zephyr/logging/log.h>
|
|
LOG_MODULE_REGISTER(spi_gd32);
|
|
|
|
#include "spi_context.h"
|
|
|
|
/* SPI error status mask. */
|
|
#define SPI_GD32_ERR_MASK (SPI_STAT_RXORERR | SPI_STAT_CONFERR | SPI_STAT_CRCERR)
|
|
|
|
#define GD32_SPI_PSC_MAX 0x7U
|
|
|
|
#if defined(CONFIG_SOC_SERIES_GD32F4XX) || \
|
|
defined(CONFIG_SOC_SERIES_GD32F403) || \
|
|
defined(CONFIG_SOC_SERIES_GD32VF103) || \
|
|
defined(CONFIG_SOC_SERIES_GD32E10X)
|
|
#define RCU_APB1EN_OFFSET APB1EN_REG_OFFSET
|
|
#elif defined(CONFIG_SOC_SERIES_GD32F3X0)
|
|
#define RCU_APB1EN_OFFSET IDX_APB1EN
|
|
#else
|
|
#error Unknown GD32 soc series
|
|
#endif
|
|
struct spi_gd32_config {
|
|
uint32_t reg;
|
|
uint16_t clkid;
|
|
struct reset_dt_spec reset;
|
|
const struct pinctrl_dev_config *pcfg;
|
|
#ifdef CONFIG_SPI_GD32_INTERRUPT
|
|
void (*irq_configure)();
|
|
#endif
|
|
};
|
|
|
|
struct spi_gd32_data {
|
|
struct spi_context ctx;
|
|
};
|
|
|
|
static int spi_gd32_get_err(const struct spi_gd32_config *cfg)
|
|
{
|
|
uint32_t stat = SPI_STAT(cfg->reg);
|
|
|
|
if (stat & SPI_GD32_ERR_MASK) {
|
|
LOG_ERR("spi%u error status detected, err = %u",
|
|
cfg->reg, stat & (uint32_t)SPI_GD32_ERR_MASK);
|
|
|
|
return -EIO;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static bool spi_gd32_transfer_ongoing(struct spi_gd32_data *data)
|
|
{
|
|
return spi_context_tx_on(&data->ctx) ||
|
|
spi_context_rx_on(&data->ctx);
|
|
}
|
|
|
|
static int spi_gd32_configure(const struct device *dev,
|
|
const struct spi_config *config)
|
|
{
|
|
struct spi_gd32_data *data = dev->data;
|
|
const struct spi_gd32_config *cfg = dev->config;
|
|
uint32_t bus_freq;
|
|
|
|
if (spi_context_configured(&data->ctx, config)) {
|
|
return 0;
|
|
}
|
|
|
|
if (SPI_OP_MODE_GET(config->operation) == SPI_OP_MODE_SLAVE) {
|
|
LOG_ERR("Slave mode not supported");
|
|
return -ENOTSUP;
|
|
}
|
|
|
|
SPI_CTL0(cfg->reg) &= ~SPI_CTL0_SPIEN;
|
|
|
|
SPI_CTL0(cfg->reg) |= SPI_MASTER;
|
|
SPI_CTL0(cfg->reg) &= ~SPI_TRANSMODE_BDTRANSMIT;
|
|
|
|
if (SPI_WORD_SIZE_GET(config->operation) == 8) {
|
|
SPI_CTL0(cfg->reg) |= SPI_FRAMESIZE_8BIT;
|
|
} else {
|
|
SPI_CTL0(cfg->reg) |= SPI_FRAMESIZE_16BIT;
|
|
}
|
|
|
|
/* Reset to hardware NSS mode. */
|
|
SPI_CTL0(cfg->reg) &= ~SPI_CTL0_SWNSSEN;
|
|
if (config->cs != NULL) {
|
|
SPI_CTL0(cfg->reg) |= SPI_CTL0_SWNSSEN;
|
|
} else {
|
|
/*
|
|
* For single master env,
|
|
* hardware NSS mode also need to set the NSSDRV bit.
|
|
*/
|
|
SPI_CTL1(cfg->reg) |= SPI_CTL1_NSSDRV;
|
|
}
|
|
|
|
SPI_CTL0(cfg->reg) &= ~SPI_CTL0_LF;
|
|
if (config->operation & SPI_TRANSFER_LSB) {
|
|
SPI_CTL0(cfg->reg) |= SPI_CTL0_LF;
|
|
}
|
|
|
|
SPI_CTL0(cfg->reg) &= ~SPI_CTL0_CKPL;
|
|
if (config->operation & SPI_MODE_CPOL) {
|
|
SPI_CTL0(cfg->reg) |= SPI_CTL0_CKPL;
|
|
}
|
|
|
|
SPI_CTL0(cfg->reg) &= ~SPI_CTL0_CKPH;
|
|
if (config->operation & SPI_MODE_CPHA) {
|
|
SPI_CTL0(cfg->reg) |= SPI_CTL0_CKPH;
|
|
}
|
|
|
|
(void)clock_control_get_rate(GD32_CLOCK_CONTROLLER,
|
|
(clock_control_subsys_t *)&cfg->clkid,
|
|
&bus_freq);
|
|
|
|
for (uint8_t i = 0U; i <= GD32_SPI_PSC_MAX; i++) {
|
|
bus_freq = bus_freq >> 1U;
|
|
if (bus_freq <= config->frequency) {
|
|
SPI_CTL0(cfg->reg) &= ~SPI_CTL0_PSC;
|
|
SPI_CTL0(cfg->reg) |= CTL0_PSC(i);
|
|
break;
|
|
}
|
|
}
|
|
|
|
data->ctx.config = config;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int spi_gd32_frame_exchange(const struct device *dev)
|
|
{
|
|
struct spi_gd32_data *data = dev->data;
|
|
const struct spi_gd32_config *cfg = dev->config;
|
|
struct spi_context *ctx = &data->ctx;
|
|
uint16_t tx_frame = 0U, rx_frame = 0U;
|
|
|
|
while ((SPI_STAT(cfg->reg) & SPI_STAT_TBE) == 0) {
|
|
/* NOP */
|
|
}
|
|
|
|
if (SPI_WORD_SIZE_GET(ctx->config->operation) == 8) {
|
|
if (spi_context_tx_buf_on(ctx)) {
|
|
tx_frame = UNALIGNED_GET((uint8_t *)(data->ctx.tx_buf));
|
|
}
|
|
/* For 8 bits mode, spi will forced SPI_DATA[15:8] to 0. */
|
|
SPI_DATA(cfg->reg) = tx_frame;
|
|
|
|
spi_context_update_tx(ctx, 1, 1);
|
|
} else {
|
|
if (spi_context_tx_buf_on(ctx)) {
|
|
tx_frame = UNALIGNED_GET((uint8_t *)(data->ctx.tx_buf));
|
|
}
|
|
SPI_DATA(cfg->reg) = tx_frame;
|
|
|
|
spi_context_update_tx(ctx, 2, 1);
|
|
}
|
|
|
|
while ((SPI_STAT(cfg->reg) & SPI_STAT_RBNE) == 0) {
|
|
/* NOP */
|
|
}
|
|
|
|
if (SPI_WORD_SIZE_GET(data->ctx.config->operation) == 8) {
|
|
/* For 8 bits mode, spi will forced SPI_DATA[15:8] to 0. */
|
|
rx_frame = SPI_DATA(cfg->reg);
|
|
if (spi_context_rx_buf_on(ctx)) {
|
|
UNALIGNED_PUT(rx_frame, (uint8_t *)data->ctx.rx_buf);
|
|
}
|
|
|
|
spi_context_update_rx(ctx, 1, 1);
|
|
} else {
|
|
rx_frame = SPI_DATA(cfg->reg);
|
|
if (spi_context_rx_buf_on(ctx)) {
|
|
UNALIGNED_PUT(rx_frame, (uint16_t *)data->ctx.rx_buf);
|
|
}
|
|
|
|
spi_context_update_rx(ctx, 2, 1);
|
|
}
|
|
|
|
return spi_gd32_get_err(cfg);
|
|
}
|
|
|
|
static int spi_gd32_transceive_impl(const struct device *dev,
|
|
const struct spi_config *config,
|
|
const struct spi_buf_set *tx_bufs,
|
|
const struct spi_buf_set *rx_bufs,
|
|
spi_callback_t cb,
|
|
void *userdata)
|
|
{
|
|
struct spi_gd32_data *data = dev->data;
|
|
const struct spi_gd32_config *cfg = dev->config;
|
|
int ret;
|
|
|
|
spi_context_lock(&data->ctx, (cb != NULL), cb, userdata, config);
|
|
|
|
ret = spi_gd32_configure(dev, config);
|
|
if (ret < 0) {
|
|
goto error;
|
|
}
|
|
|
|
SPI_CTL0(cfg->reg) |= SPI_CTL0_SPIEN;
|
|
|
|
spi_context_buffers_setup(&data->ctx, tx_bufs, rx_bufs, 1);
|
|
|
|
spi_context_cs_control(&data->ctx, true);
|
|
|
|
#ifdef CONFIG_SPI_GD32_INTERRUPT
|
|
SPI_STAT(cfg->reg) &= ~(SPI_STAT_RBNE | SPI_STAT_TBE | SPI_GD32_ERR_MASK);
|
|
SPI_CTL1(cfg->reg) |= (SPI_CTL1_RBNEIE | SPI_CTL1_TBEIE | SPI_CTL1_ERRIE);
|
|
ret = spi_context_wait_for_completion(&data->ctx);
|
|
#else
|
|
do {
|
|
ret = spi_gd32_frame_exchange(dev);
|
|
if (ret < 0) {
|
|
break;
|
|
}
|
|
} while (spi_gd32_transfer_ongoing(data));
|
|
|
|
#ifdef CONFIG_SPI_ASYNC
|
|
spi_context_complete(&data->ctx, dev, ret);
|
|
#endif
|
|
#endif
|
|
|
|
while (!(SPI_STAT(cfg->reg) & SPI_STAT_TBE) ||
|
|
(SPI_STAT(cfg->reg) & SPI_STAT_TRANS)) {
|
|
/* Wait until last frame transfer complete. */
|
|
}
|
|
|
|
spi_context_cs_control(&data->ctx, false);
|
|
|
|
SPI_CTL0(cfg->reg) &= ~SPI_CTL0_SPIEN;
|
|
|
|
error:
|
|
spi_context_release(&data->ctx, ret);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int spi_gd32_transceive(const struct device *dev,
|
|
const struct spi_config *config,
|
|
const struct spi_buf_set *tx_bufs,
|
|
const struct spi_buf_set *rx_bufs)
|
|
{
|
|
return spi_gd32_transceive_impl(dev, config, tx_bufs, rx_bufs, NULL, NULL);
|
|
}
|
|
|
|
#ifdef CONFIG_SPI_ASYNC
|
|
static int spi_gd32_transceive_async(const struct device *dev,
|
|
const struct spi_config *config,
|
|
const struct spi_buf_set *tx_bufs,
|
|
const struct spi_buf_set *rx_bufs,
|
|
spi_callback_t cb,
|
|
void *userdata)
|
|
{
|
|
return spi_gd32_transceive_impl(dev, config, tx_bufs, rx_bufs, cb, userdata);
|
|
}
|
|
#endif
|
|
|
|
#ifdef CONFIG_SPI_GD32_INTERRUPT
|
|
|
|
static void spi_gd32_complete(const struct device *dev, int status)
|
|
{
|
|
struct spi_gd32_data *data = dev->data;
|
|
const struct spi_gd32_config *cfg = dev->config;
|
|
|
|
SPI_CTL1(cfg->reg) &= ~(SPI_CTL1_RBNEIE | SPI_CTL1_TBEIE | SPI_CTL1_ERRIE);
|
|
|
|
spi_context_complete(&data->ctx, dev, status);
|
|
}
|
|
|
|
static void spi_gd32_isr(struct device *dev)
|
|
{
|
|
const struct spi_gd32_config *cfg = dev->config;
|
|
struct spi_gd32_data *data = dev->data;
|
|
int err = 0;
|
|
|
|
if ((SPI_STAT(cfg->reg) & SPI_GD32_ERR_MASK) != 0) {
|
|
err = spi_gd32_get_err(cfg);
|
|
} else {
|
|
err = spi_gd32_frame_exchange(dev);
|
|
}
|
|
|
|
if (err || !spi_gd32_transfer_ongoing(data)) {
|
|
spi_gd32_complete(dev, err);
|
|
}
|
|
|
|
SPI_STAT(cfg->reg) = 0;
|
|
}
|
|
|
|
#endif /* INTERRUPT */
|
|
|
|
static int spi_gd32_release(const struct device *dev,
|
|
const struct spi_config *config)
|
|
{
|
|
struct spi_gd32_data *data = dev->data;
|
|
|
|
spi_context_unlock_unconditionally(&data->ctx);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct spi_driver_api spi_gd32_driver_api = {
|
|
.transceive = spi_gd32_transceive,
|
|
#ifdef CONFIG_SPI_ASYNC
|
|
.transceive_async = spi_gd32_transceive_async,
|
|
#endif
|
|
.release = spi_gd32_release
|
|
};
|
|
|
|
int spi_gd32_init(const struct device *dev)
|
|
{
|
|
struct spi_gd32_data *data = dev->data;
|
|
const struct spi_gd32_config *cfg = dev->config;
|
|
int ret;
|
|
|
|
(void)clock_control_on(GD32_CLOCK_CONTROLLER,
|
|
(clock_control_subsys_t *)&cfg->clkid);
|
|
|
|
(void)reset_line_toggle_dt(&cfg->reset);
|
|
|
|
ret = pinctrl_apply_state(cfg->pcfg, PINCTRL_STATE_DEFAULT);
|
|
if (ret) {
|
|
LOG_ERR("Failed to apply pinctrl state");
|
|
return ret;
|
|
}
|
|
|
|
ret = spi_context_cs_configure_all(&data->ctx);
|
|
if (ret < 0) {
|
|
return ret;
|
|
}
|
|
|
|
#ifdef CONFIG_SPI_GD32_INTERRUPT
|
|
cfg->irq_configure(dev);
|
|
#endif
|
|
|
|
spi_context_unlock_unconditionally(&data->ctx);
|
|
|
|
return 0;
|
|
}
|
|
|
|
#define GD32_IRQ_CONFIGURE(idx) \
|
|
static void spi_gd32_irq_configure_##idx(void) \
|
|
{ \
|
|
IRQ_CONNECT(DT_INST_IRQN(idx), DT_INST_IRQ(idx, priority), \
|
|
spi_gd32_isr, \
|
|
DEVICE_DT_INST_GET(idx), 0); \
|
|
irq_enable(DT_INST_IRQN(idx)); \
|
|
}
|
|
|
|
#define GD32_SPI_INIT(idx) \
|
|
PINCTRL_DT_INST_DEFINE(idx); \
|
|
IF_ENABLED(CONFIG_SPI_GD32_INTERRUPT, (GD32_IRQ_CONFIGURE(idx))); \
|
|
static struct spi_gd32_data spi_gd32_data_##idx = { \
|
|
SPI_CONTEXT_INIT_LOCK(spi_gd32_data_##idx, ctx), \
|
|
SPI_CONTEXT_INIT_SYNC(spi_gd32_data_##idx, ctx), \
|
|
SPI_CONTEXT_CS_GPIOS_INITIALIZE(DT_DRV_INST(idx), ctx) }; \
|
|
static struct spi_gd32_config spi_gd32_config_##idx = { \
|
|
.reg = DT_INST_REG_ADDR(idx), \
|
|
.clkid = DT_INST_CLOCKS_CELL(idx, id), \
|
|
.reset = RESET_DT_SPEC_INST_GET(idx), \
|
|
.pcfg = PINCTRL_DT_INST_DEV_CONFIG_GET(idx), \
|
|
IF_ENABLED(CONFIG_SPI_GD32_INTERRUPT, \
|
|
(.irq_configure = spi_gd32_irq_configure_##idx)) }; \
|
|
DEVICE_DT_INST_DEFINE(idx, &spi_gd32_init, NULL, \
|
|
&spi_gd32_data_##idx, &spi_gd32_config_##idx, \
|
|
POST_KERNEL, CONFIG_SPI_INIT_PRIORITY, \
|
|
&spi_gd32_driver_api);
|
|
|
|
DT_INST_FOREACH_STATUS_OKAY(GD32_SPI_INIT)
|