305 lines
7.3 KiB
C
305 lines
7.3 KiB
C
/*
|
|
* Copyright (c) 2018 SiFive Inc.
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*/
|
|
|
|
#define DT_DRV_COMPAT sifive_spi0
|
|
|
|
#define LOG_LEVEL CONFIG_SPI_LOG_LEVEL
|
|
#include <zephyr/logging/log.h>
|
|
LOG_MODULE_REGISTER(spi_sifive);
|
|
|
|
#include "spi_sifive.h"
|
|
|
|
#include <soc.h>
|
|
#include <stdbool.h>
|
|
|
|
/* Helper Functions */
|
|
|
|
static ALWAYS_INLINE
|
|
void sys_set_mask(mem_addr_t addr, uint32_t mask, uint32_t value)
|
|
{
|
|
uint32_t temp = sys_read32(addr);
|
|
|
|
temp &= ~(mask);
|
|
temp |= value;
|
|
|
|
sys_write32(temp, addr);
|
|
}
|
|
|
|
static int spi_config(const struct device *dev, uint32_t frequency,
|
|
uint16_t operation)
|
|
{
|
|
uint32_t div;
|
|
uint32_t fmt_len;
|
|
|
|
if (operation & SPI_HALF_DUPLEX) {
|
|
return -ENOTSUP;
|
|
}
|
|
|
|
if (SPI_OP_MODE_GET(operation) != SPI_OP_MODE_MASTER) {
|
|
return -ENOTSUP;
|
|
}
|
|
|
|
if (operation & SPI_MODE_LOOP) {
|
|
return -ENOTSUP;
|
|
}
|
|
|
|
/* Set the SPI frequency */
|
|
div = (SPI_CFG(dev)->f_sys / (frequency * 2U)) - 1;
|
|
sys_write32((SF_SCKDIV_DIV_MASK & div), SPI_REG(dev, REG_SCKDIV));
|
|
|
|
/* Set the polarity */
|
|
if (operation & SPI_MODE_CPOL) {
|
|
/* If CPOL is set, then SCK idles at logical 1 */
|
|
sys_set_bit(SPI_REG(dev, REG_SCKMODE), SF_SCKMODE_POL);
|
|
} else {
|
|
/* SCK idles at logical 0 */
|
|
sys_clear_bit(SPI_REG(dev, REG_SCKMODE), SF_SCKMODE_POL);
|
|
}
|
|
|
|
/* Set the phase */
|
|
if (operation & SPI_MODE_CPHA) {
|
|
/*
|
|
* If CPHA is set, then data is sampled
|
|
* on the trailing SCK edge
|
|
*/
|
|
sys_set_bit(SPI_REG(dev, REG_SCKMODE), SF_SCKMODE_PHA);
|
|
} else {
|
|
/* Data is sampled on the leading SCK edge */
|
|
sys_clear_bit(SPI_REG(dev, REG_SCKMODE), SF_SCKMODE_PHA);
|
|
}
|
|
|
|
/* Get the frame length */
|
|
fmt_len = SPI_WORD_SIZE_GET(operation);
|
|
if (fmt_len > SF_FMT_LEN_MASK) {
|
|
return -ENOTSUP;
|
|
}
|
|
|
|
/* Set the frame length */
|
|
fmt_len = fmt_len << SF_FMT_LEN;
|
|
fmt_len &= SF_FMT_LEN_MASK;
|
|
sys_set_mask(SPI_REG(dev, REG_FMT), SF_FMT_LEN_MASK, fmt_len);
|
|
|
|
if (IS_ENABLED(CONFIG_SPI_EXTENDED_MODES) &&
|
|
(operation & SPI_LINES_MASK) != SPI_LINES_SINGLE) {
|
|
return -ENOTSUP;
|
|
}
|
|
/* Set single line operation */
|
|
sys_set_mask(SPI_REG(dev, REG_FMT),
|
|
SF_FMT_PROTO_MASK,
|
|
SF_FMT_PROTO_SINGLE);
|
|
|
|
/* Set the endianness */
|
|
if (operation & SPI_TRANSFER_LSB) {
|
|
sys_set_bit(SPI_REG(dev, REG_FMT), SF_FMT_ENDIAN);
|
|
} else {
|
|
sys_clear_bit(SPI_REG(dev, REG_FMT), SF_FMT_ENDIAN);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static ALWAYS_INLINE bool spi_sifive_send_available(const struct device *dev)
|
|
{
|
|
return !(sys_read32(SPI_REG(dev, REG_TXDATA)) & SF_TXDATA_FULL);
|
|
}
|
|
|
|
static ALWAYS_INLINE
|
|
void spi_sifive_send(const struct device *dev, uint8_t frame)
|
|
{
|
|
sys_write32((uint32_t) frame, SPI_REG(dev, REG_TXDATA));
|
|
}
|
|
|
|
static ALWAYS_INLINE
|
|
bool spi_sifive_recv(const struct device *dev, uint8_t *val)
|
|
{
|
|
uint32_t reg = sys_read32(SPI_REG(dev, REG_RXDATA));
|
|
|
|
if (reg & SF_RXDATA_EMPTY) {
|
|
return false;
|
|
}
|
|
*val = (uint8_t) reg;
|
|
return true;
|
|
}
|
|
|
|
static void spi_sifive_xfer(const struct device *dev, const bool hw_cs_control)
|
|
{
|
|
struct spi_context *ctx = &SPI_DATA(dev)->ctx;
|
|
uint8_t txd, rxd;
|
|
int queued_frames = 0;
|
|
|
|
while (spi_context_tx_on(ctx) || spi_context_rx_on(ctx) || queued_frames > 0) {
|
|
bool send = false;
|
|
|
|
/* As long as frames remain to be sent, attempt to queue them on Tx FIFO. If
|
|
* the FIFO is full then another attempt will be made next pass. If Rx length
|
|
* > Tx length then queue dummy Tx in order to read the requested Rx data.
|
|
*/
|
|
if (spi_context_tx_buf_on(ctx)) {
|
|
send = true;
|
|
txd = *ctx->tx_buf;
|
|
} else if (queued_frames == 0) { /* Implies spi_context_rx_on(). */
|
|
send = true;
|
|
txd = 0U;
|
|
}
|
|
|
|
if (send && spi_sifive_send_available(dev)) {
|
|
spi_sifive_send(dev, txd);
|
|
queued_frames++;
|
|
spi_context_update_tx(ctx, 1, 1);
|
|
}
|
|
|
|
if (queued_frames > 0 && spi_sifive_recv(dev, &rxd)) {
|
|
if (spi_context_rx_buf_on(ctx)) {
|
|
*ctx->rx_buf = rxd;
|
|
}
|
|
queued_frames--;
|
|
spi_context_update_rx(ctx, 1, 1);
|
|
}
|
|
}
|
|
|
|
/* Deassert the CS line */
|
|
if (!hw_cs_control) {
|
|
spi_context_cs_control(&SPI_DATA(dev)->ctx, false);
|
|
} else {
|
|
sys_write32(SF_CSMODE_OFF, SPI_REG(dev, REG_CSMODE));
|
|
}
|
|
|
|
spi_context_complete(ctx, dev, 0);
|
|
}
|
|
|
|
/* API Functions */
|
|
|
|
static int spi_sifive_init(const struct device *dev)
|
|
{
|
|
int err;
|
|
#ifdef CONFIG_PINCTRL
|
|
struct spi_sifive_cfg *cfg = (struct spi_sifive_cfg *)dev->config;
|
|
#endif
|
|
/* Disable SPI Flash mode */
|
|
sys_clear_bit(SPI_REG(dev, REG_FCTRL), SF_FCTRL_EN);
|
|
|
|
err = spi_context_cs_configure_all(&SPI_DATA(dev)->ctx);
|
|
if (err < 0) {
|
|
return err;
|
|
}
|
|
|
|
#ifdef CONFIG_PINCTRL
|
|
err = pinctrl_apply_state(cfg->pcfg, PINCTRL_STATE_DEFAULT);
|
|
if (err < 0) {
|
|
return err;
|
|
}
|
|
#endif
|
|
|
|
/* Make sure the context is unlocked */
|
|
spi_context_unlock_unconditionally(&SPI_DATA(dev)->ctx);
|
|
return 0;
|
|
}
|
|
|
|
static int spi_sifive_transceive(const struct device *dev,
|
|
const struct spi_config *config,
|
|
const struct spi_buf_set *tx_bufs,
|
|
const struct spi_buf_set *rx_bufs)
|
|
{
|
|
int rc = 0;
|
|
bool hw_cs_control = false;
|
|
|
|
/* Lock the SPI Context */
|
|
spi_context_lock(&SPI_DATA(dev)->ctx, false, NULL, NULL, config);
|
|
|
|
/* Configure the SPI bus */
|
|
SPI_DATA(dev)->ctx.config = config;
|
|
|
|
/*
|
|
* If the chip select configuration is not present, we'll ask the
|
|
* SPI peripheral itself to control the CS line
|
|
*/
|
|
if (!spi_cs_is_gpio(config)) {
|
|
hw_cs_control = true;
|
|
}
|
|
|
|
if (!hw_cs_control) {
|
|
/*
|
|
* If the user has requested manual GPIO control, ask the
|
|
* context for control and disable HW control
|
|
*/
|
|
sys_write32(SF_CSMODE_OFF, SPI_REG(dev, REG_CSMODE));
|
|
} else {
|
|
/*
|
|
* Tell the hardware to control the requested CS pin.
|
|
* NOTE:
|
|
* For the SPI peripheral, the pin number is not the
|
|
* GPIO pin, but the index into the list of available
|
|
* CS lines for the SPI peripheral.
|
|
*/
|
|
sys_write32(config->slave, SPI_REG(dev, REG_CSID));
|
|
sys_write32(SF_CSMODE_OFF, SPI_REG(dev, REG_CSMODE));
|
|
}
|
|
|
|
rc = spi_config(dev, config->frequency, config->operation);
|
|
if (rc < 0) {
|
|
spi_context_release(&SPI_DATA(dev)->ctx, rc);
|
|
return rc;
|
|
}
|
|
|
|
spi_context_buffers_setup(&SPI_DATA(dev)->ctx, tx_bufs, rx_bufs, 1);
|
|
|
|
/* Assert the CS line */
|
|
if (!hw_cs_control) {
|
|
spi_context_cs_control(&SPI_DATA(dev)->ctx, true);
|
|
} else {
|
|
sys_write32(SF_CSMODE_HOLD, SPI_REG(dev, REG_CSMODE));
|
|
}
|
|
|
|
/* Perform transfer */
|
|
spi_sifive_xfer(dev, hw_cs_control);
|
|
|
|
rc = spi_context_wait_for_completion(&SPI_DATA(dev)->ctx);
|
|
|
|
spi_context_release(&SPI_DATA(dev)->ctx, rc);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int spi_sifive_release(const struct device *dev,
|
|
const struct spi_config *config)
|
|
{
|
|
spi_context_unlock_unconditionally(&SPI_DATA(dev)->ctx);
|
|
return 0;
|
|
}
|
|
|
|
/* Device Instantiation */
|
|
|
|
static const struct spi_driver_api spi_sifive_api = {
|
|
.transceive = spi_sifive_transceive,
|
|
#ifdef CONFIG_SPI_RTIO
|
|
.iodev_submit = spi_rtio_iodev_default_submit,
|
|
#endif
|
|
.release = spi_sifive_release,
|
|
};
|
|
|
|
#define SPI_INIT(n) \
|
|
PINCTRL_DT_INST_DEFINE(n); \
|
|
static struct spi_sifive_data spi_sifive_data_##n = { \
|
|
SPI_CONTEXT_INIT_LOCK(spi_sifive_data_##n, ctx), \
|
|
SPI_CONTEXT_INIT_SYNC(spi_sifive_data_##n, ctx), \
|
|
SPI_CONTEXT_CS_GPIOS_INITIALIZE(DT_DRV_INST(n), ctx) \
|
|
}; \
|
|
static struct spi_sifive_cfg spi_sifive_cfg_##n = { \
|
|
.base = DT_INST_REG_ADDR_BY_NAME(n, control), \
|
|
.f_sys = SIFIVE_PERIPHERAL_CLOCK_FREQUENCY, \
|
|
.pcfg = PINCTRL_DT_INST_DEV_CONFIG_GET(n), \
|
|
}; \
|
|
DEVICE_DT_INST_DEFINE(n, \
|
|
spi_sifive_init, \
|
|
NULL, \
|
|
&spi_sifive_data_##n, \
|
|
&spi_sifive_cfg_##n, \
|
|
POST_KERNEL, \
|
|
CONFIG_SPI_INIT_PRIORITY, \
|
|
&spi_sifive_api);
|
|
|
|
DT_INST_FOREACH_STATUS_OKAY(SPI_INIT)
|