You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
932 lines
26 KiB
932 lines
26 KiB
/* |
|
* Copyright (c) 2016, Freescale Semiconductor, Inc. |
|
* Copyright (c) 2017, 2020-2021, NXP |
|
* |
|
* SPDX-License-Identifier: Apache-2.0 |
|
*/ |
|
|
|
#define DT_DRV_COMPAT nxp_kinetis_dspi |
|
|
|
#include <errno.h> |
|
#include <zephyr/drivers/spi.h> |
|
#include <zephyr/drivers/clock_control.h> |
|
#include <fsl_dspi.h> |
|
#include <zephyr/drivers/pinctrl.h> |
|
#ifdef CONFIG_DSPI_MCUX_EDMA |
|
#include <zephyr/drivers/dma.h> |
|
#include <fsl_edma.h> |
|
#endif |
|
|
|
#define LOG_LEVEL CONFIG_SPI_LOG_LEVEL |
|
#include <zephyr/logging/log.h> |
|
#include <zephyr/irq.h> |
|
LOG_MODULE_REGISTER(spi_mcux_dspi); |
|
|
|
#include "spi_context.h" |
|
|
|
#ifdef CONFIG_DSPI_MCUX_EDMA |
|
|
|
struct spi_edma_config { |
|
const struct device *dma_dev; |
|
int32_t state; |
|
uint32_t dma_channel; |
|
void (*irq_call_back)(void); |
|
struct dma_config dma_cfg; |
|
}; |
|
#endif |
|
|
|
struct spi_mcux_config { |
|
SPI_Type *base; |
|
const struct device *clock_dev; |
|
clock_control_subsys_t clock_subsys; |
|
void (*irq_config_func)(const struct device *dev); |
|
uint32_t pcs_sck_delay; |
|
uint32_t sck_pcs_delay; |
|
uint32_t transfer_delay; |
|
uint32_t which_ctar; |
|
uint32_t samplePoint; |
|
bool enable_continuous_sck; |
|
bool enable_rxfifo_overwrite; |
|
bool enable_modified_timing_format; |
|
bool is_dma_chn_shared; |
|
const struct pinctrl_dev_config *pincfg; |
|
}; |
|
|
|
struct spi_mcux_data { |
|
const struct device *dev; |
|
dspi_master_handle_t handle; |
|
struct spi_context ctx; |
|
size_t transfer_len; |
|
#ifdef CONFIG_DSPI_MCUX_EDMA |
|
struct dma_block_config tx_dma_block; |
|
struct dma_block_config tx_dma_block_end; |
|
struct dma_block_config rx_dma_block; |
|
struct spi_edma_config rx_dma_config; |
|
struct spi_edma_config tx_dma_config; |
|
int frame_size; |
|
int tx_transfer_count; |
|
int rx_transfer_count; |
|
uint32_t which_pcs; |
|
struct spi_buf *inner_tx_buffer; |
|
struct spi_buf *inner_rx_buffer; |
|
#endif |
|
}; |
|
|
|
#ifdef CONFIG_DSPI_MCUX_EDMA |
|
static int get_size_byte_by_frame_size(int len, int frame_size) |
|
{ |
|
if (frame_size == 8) { |
|
return (len * 4); |
|
} else { /* frame_size == 16*/ |
|
return (len * 2); |
|
} |
|
} |
|
#endif |
|
|
|
static int spi_mcux_transfer_next_packet(const struct device *dev) |
|
{ |
|
const struct spi_mcux_config *config = dev->config; |
|
struct spi_mcux_data *data = dev->data; |
|
SPI_Type *base = config->base; |
|
struct spi_context *ctx = &data->ctx; |
|
dspi_transfer_t transfer; |
|
status_t status; |
|
|
|
if ((ctx->tx_len == 0) && (ctx->rx_len == 0)) { |
|
/* nothing left to rx or tx, we're done! */ |
|
LOG_DBG("spi transceive done"); |
|
spi_context_cs_control(&data->ctx, false); |
|
spi_context_complete(&data->ctx, dev, 0); |
|
return 0; |
|
} |
|
|
|
#ifdef CONFIG_DSPI_MCUX_EDMA |
|
|
|
if (!config->is_dma_chn_shared) { |
|
/* start dma directly in not shared mode */ |
|
if (ctx->tx_len != 0) { |
|
int ret = 0; |
|
|
|
LOG_DBG("Starting DMA Ch%u", |
|
data->tx_dma_config.dma_channel); |
|
ret = dma_start(data->tx_dma_config.dma_dev, |
|
data->tx_dma_config.dma_channel); |
|
if (ret < 0) { |
|
LOG_ERR("Failed to start DMA Ch%d (%d)", |
|
data->tx_dma_config.dma_channel, ret); |
|
return ret; |
|
} |
|
} |
|
|
|
if (ctx->rx_len != 0) { |
|
int ret = 0; |
|
|
|
LOG_DBG("Starting DMA Ch%u", |
|
data->rx_dma_config.dma_channel); |
|
ret = dma_start(data->rx_dma_config.dma_dev, |
|
data->rx_dma_config.dma_channel); |
|
if (ret < 0) { |
|
LOG_ERR("Failed to start DMA Ch%d (%d)", |
|
data->rx_dma_config.dma_channel, ret); |
|
return ret; |
|
} |
|
} |
|
} |
|
|
|
DSPI_EnableDMA(base, (uint32_t)kDSPI_RxDmaEnable | |
|
(uint32_t)kDSPI_TxDmaEnable); |
|
DSPI_StartTransfer(base); |
|
|
|
if (config->is_dma_chn_shared) { |
|
/* in master mode start tx */ |
|
dma_start(data->tx_dma_config.dma_dev, data->tx_dma_config.dma_channel); |
|
/* TBD kDSPI_TxFifoFillRequestFlag */ |
|
DSPI_EnableInterrupts(base, |
|
(uint32_t)kDSPI_RxFifoDrainRequestFlag); |
|
LOG_DBG("trigger tx to start master"); |
|
} |
|
|
|
return 0; |
|
#endif |
|
|
|
transfer.configFlags = kDSPI_MasterCtar0 | kDSPI_MasterPcsContinuous | |
|
(ctx->config->slave << DSPI_MASTER_PCS_SHIFT); |
|
|
|
if (ctx->tx_len == 0) { |
|
/* rx only, nothing to tx */ |
|
transfer.txData = NULL; |
|
transfer.rxData = ctx->rx_buf; |
|
transfer.dataSize = ctx->rx_len; |
|
} else if (ctx->rx_len == 0) { |
|
/* tx only, nothing to rx */ |
|
transfer.txData = (uint8_t *) ctx->tx_buf; |
|
transfer.rxData = NULL; |
|
transfer.dataSize = ctx->tx_len; |
|
} else if (ctx->tx_len == ctx->rx_len) { |
|
/* rx and tx are the same length */ |
|
transfer.txData = (uint8_t *) ctx->tx_buf; |
|
transfer.rxData = ctx->rx_buf; |
|
transfer.dataSize = ctx->tx_len; |
|
} else if (ctx->tx_len > ctx->rx_len) { |
|
/* Break up the tx into multiple transfers so we don't have to |
|
* rx into a longer intermediate buffer. Leave chip select |
|
* active between transfers. |
|
*/ |
|
transfer.txData = (uint8_t *) ctx->tx_buf; |
|
transfer.rxData = ctx->rx_buf; |
|
transfer.dataSize = ctx->rx_len; |
|
transfer.configFlags |= kDSPI_MasterActiveAfterTransfer; |
|
} else { |
|
/* Break up the rx into multiple transfers so we don't have to |
|
* tx from a longer intermediate buffer. Leave chip select |
|
* active between transfers. |
|
*/ |
|
transfer.txData = (uint8_t *) ctx->tx_buf; |
|
transfer.rxData = ctx->rx_buf; |
|
transfer.dataSize = ctx->tx_len; |
|
transfer.configFlags |= kDSPI_MasterActiveAfterTransfer; |
|
} |
|
|
|
if (!(ctx->tx_count <= 1 && ctx->rx_count <= 1)) { |
|
transfer.configFlags |= kDSPI_MasterActiveAfterTransfer; |
|
} |
|
|
|
data->transfer_len = transfer.dataSize; |
|
|
|
status = DSPI_MasterTransferNonBlocking(base, &data->handle, &transfer); |
|
if (status != kStatus_Success) { |
|
LOG_ERR("Transfer could not start on %s: %d", dev->name, status); |
|
return status == kDSPI_Busy ? -EBUSY : -EINVAL; |
|
} |
|
|
|
return 0; |
|
} |
|
|
|
static void spi_mcux_isr(const struct device *dev) |
|
{ |
|
const struct spi_mcux_config *config = dev->config; |
|
struct spi_mcux_data *data = dev->data; |
|
SPI_Type *base = config->base; |
|
|
|
#ifdef CONFIG_DSPI_MCUX_EDMA |
|
LOG_DBG("isr is called"); |
|
|
|
if (0U != (DSPI_GetStatusFlags(base) & |
|
(uint32_t)kDSPI_RxFifoDrainRequestFlag)) { |
|
/* start rx */ |
|
dma_start(data->rx_dma_config.dma_dev, data->rx_dma_config.dma_channel); |
|
} |
|
#else |
|
DSPI_MasterTransferHandleIRQ(base, &data->handle); |
|
#endif |
|
} |
|
|
|
#ifdef CONFIG_DSPI_MCUX_EDMA |
|
|
|
static void mcux_init_inner_buffer_with_cmd(const struct device *dev, |
|
uint16_t dummy) |
|
{ |
|
const struct spi_mcux_config *config = dev->config; |
|
struct spi_mcux_data *data = dev->data; |
|
dspi_command_data_config_t commandStruct; |
|
uint32_t *pbuf = data->inner_tx_buffer->buf; |
|
uint32_t command; |
|
int i = 0; |
|
|
|
commandStruct.whichPcs = data->which_pcs; |
|
|
|
commandStruct.isEndOfQueue = false; |
|
commandStruct.clearTransferCount = false; |
|
commandStruct.whichCtar = config->which_ctar; |
|
commandStruct.isPcsContinuous = config->enable_continuous_sck; |
|
command = DSPI_MasterGetFormattedCommand(&(commandStruct)); |
|
for (i = 0; i < data->inner_tx_buffer->len / 4; i++) { |
|
*pbuf = command | dummy; |
|
pbuf++; |
|
} |
|
} |
|
|
|
/** |
|
* @brief update the tx data to internal buffer with command embedded, |
|
* if no tx data, use dummy value. |
|
* tx data frame size shall not bigger than 16 bits |
|
* the overall transfer data in one batch shall not larger than FIFO size |
|
*/ |
|
static int mcux_spi_context_data_update(const struct device *dev) |
|
{ |
|
struct spi_mcux_data *data = dev->data; |
|
uint32_t frame_size_bit = data->frame_size; |
|
struct spi_context *ctx = (struct spi_context *)&data->ctx; |
|
uint32_t *pcdata = data->inner_tx_buffer->buf; |
|
|
|
if (frame_size_bit > FSL_FEATURE_DSPI_MAX_DATA_WIDTH) { |
|
/* TODO need set to continues PCS to have frame size larger than 16 */ |
|
LOG_ERR("frame size is larger than 16"); |
|
return -EINVAL; |
|
} |
|
|
|
#ifdef CONFIG_MCUX_DSPI_EDMA_SHUFFLE_DATA |
|
/* only used when use inner buffer to translate tx format */ |
|
if (CONFIG_MCUX_DSPI_BUFFER_SIZE * 4 < |
|
get_size_byte_by_frame_size(ctx->current_tx->len, frame_size_bit)) { |
|
/* inner buffer can not hold all transferred data */ |
|
LOG_ERR("inner buffer is too small to hold all data esp %d, act %d", |
|
ctx->current_tx->len * 8 / frame_size_bit, |
|
(CONFIG_MCUX_DSPI_BUFFER_SIZE * 4 / frame_size_bit)); |
|
return -EINVAL; |
|
} |
|
|
|
if (frame_size_bit == 8) { |
|
int i = 0; |
|
uint8_t *pdata = (uint8_t *)ctx->tx_buf; |
|
|
|
if (pdata) { |
|
do { |
|
uint16_t temp_data = 0; |
|
|
|
temp_data = *pdata; |
|
pdata++; |
|
*pcdata |= temp_data; |
|
pcdata++; |
|
i++; |
|
} while (i < ctx->current_tx->len && |
|
i < data->inner_tx_buffer->len); |
|
} |
|
/* indicate it is the last data */ |
|
if (i == ctx->current_tx->len) { |
|
--pcdata; |
|
*pcdata |= SPI_PUSHR_EOQ(1) | SPI_PUSHR_CTCNT(1); |
|
LOG_DBG("last pcdata is %x", *pcdata); |
|
} |
|
} else if (frame_size_bit == 16) { |
|
int i = 0; |
|
uint16_t *pdata = (uint16_t *)ctx->tx_buf; |
|
|
|
if (pdata) { |
|
do { |
|
*pcdata |= *pdata; |
|
LOG_DBG("pcdata %d is %x", i / 2, *pcdata); |
|
pdata++; |
|
pcdata++; |
|
i += 2; |
|
} while (i < ctx->current_tx->len && |
|
i < data->inner_tx_buffer->len); |
|
} |
|
if (i == ctx->current_tx->len) { |
|
/* indicate it is the last data */ |
|
--pcdata; |
|
*pcdata |= SPI_PUSHR_EOQ(1); |
|
LOG_DBG("last pcdata is %x", *pcdata); |
|
} |
|
} else { |
|
/* TODO for other size */ |
|
LOG_ERR("DMA mode only support 8/16 bits frame size"); |
|
return -EINVAL; |
|
} |
|
|
|
#endif /* CONFIG_MCUX_DSPI_EDMA_SHUFFLE_DATA */ |
|
|
|
return 0; |
|
} |
|
|
|
static int update_tx_dma(const struct device *dev) |
|
{ |
|
uint32_t tx_size = 0; |
|
uint8_t *tx_buf; |
|
struct spi_mcux_data *data = dev->data; |
|
const struct spi_mcux_config *config = dev->config; |
|
SPI_Type *base = config->base; |
|
uint32_t frame_size = data->frame_size; |
|
bool rx_only = false; |
|
|
|
DSPI_DisableDMA(base, (uint32_t)kDSPI_TxDmaEnable); |
|
if (data->ctx.tx_len == 0) { |
|
LOG_DBG("empty data no need to setup DMA"); |
|
return 0; |
|
} |
|
|
|
if (data->ctx.current_tx && data->ctx.current_tx->len > 0 && |
|
data->ctx.current_tx->buf != NULL) { |
|
#ifdef CONFIG_MCUX_DSPI_EDMA_SHUFFLE_DATA |
|
tx_size = get_size_byte_by_frame_size(data->transfer_len, |
|
frame_size); |
|
tx_buf = data->inner_tx_buffer->buf; |
|
#else |
|
/* expect the buffer is pre-set */ |
|
tx_size = get_size_byte_by_frame_size(data->ctx.current_tx->len, |
|
frame_size); |
|
LOG_DBG("tx size is %d", tx_size); |
|
tx_buf = data->ctx.current_tx->buf; |
|
#endif |
|
} else { |
|
tx_buf = data->inner_tx_buffer->buf; |
|
tx_size = get_size_byte_by_frame_size(data->transfer_len, |
|
frame_size); |
|
rx_only = true; |
|
LOG_DBG("rx only 0x%x, size %d", (uint32_t)tx_buf, tx_size); |
|
} |
|
|
|
data->tx_dma_block.source_address = (uint32_t)tx_buf; |
|
data->tx_dma_block.dest_address = |
|
DSPI_MasterGetTxRegisterAddress(base); |
|
data->tx_dma_block.next_block = NULL; |
|
if (config->is_dma_chn_shared) { |
|
/* transfer FIFO size data */ |
|
data->tx_dma_block.block_size = 4; |
|
} else { |
|
data->tx_dma_block.block_size = tx_size; |
|
} |
|
|
|
data->tx_dma_config.dma_cfg.user_data = (void *) dev; |
|
dma_config(data->tx_dma_config.dma_dev, data->tx_dma_config.dma_channel, |
|
(struct dma_config *)&data->tx_dma_config.dma_cfg); |
|
|
|
return 0; |
|
} |
|
|
|
static int update_rx_dma(const struct device *dev) |
|
{ |
|
uint32_t rx_size = 0; |
|
uint8_t *rx_buf; |
|
struct spi_mcux_data *data = dev->data; |
|
const struct spi_mcux_config *config = dev->config; |
|
SPI_Type *base = config->base; |
|
uint32_t frame_size_byte = (data->frame_size >> 3); |
|
bool tx_only = false; |
|
|
|
DSPI_DisableDMA(base, (uint32_t)kDSPI_RxDmaEnable); |
|
if (data->ctx.rx_len == 0) { |
|
LOG_DBG("empty data no need to setup DMA"); |
|
return 0; |
|
} |
|
|
|
if (data->ctx.current_rx) { |
|
rx_size = data->transfer_len; |
|
if (data->ctx.rx_buf != NULL) { |
|
rx_buf = data->ctx.rx_buf; |
|
} else { |
|
rx_buf = data->inner_rx_buffer->buf; |
|
} |
|
} else { |
|
/* tx only */ |
|
rx_buf = data->inner_rx_buffer->buf; |
|
rx_size = data->transfer_len; |
|
tx_only = true; |
|
LOG_DBG("tx only 0x%x, size %d", (uint32_t)rx_buf, rx_size); |
|
} |
|
|
|
if (config->is_dma_chn_shared) { |
|
if (data->ctx.rx_len == 1) { |
|
/* do not link tx on last frame*/ |
|
LOG_DBG("do not link tx/rx channel for last one"); |
|
data->rx_dma_config.dma_cfg.source_chaining_en = 0; |
|
data->rx_dma_config.dma_cfg.dest_chaining_en = 0; |
|
} else { |
|
LOG_DBG("shared mux mode, link tx/rx channel"); |
|
data->rx_dma_config.dma_cfg.source_chaining_en = 1; |
|
data->rx_dma_config.dma_cfg.dest_chaining_en = 1; |
|
data->rx_dma_config.dma_cfg.linked_channel = |
|
data->tx_dma_config.dma_channel; |
|
} |
|
|
|
data->rx_dma_block.dest_address = (uint32_t)rx_buf; |
|
data->rx_dma_block.source_address = |
|
DSPI_GetRxRegisterAddress(base); |
|
/* do once in share mode */ |
|
data->rx_dma_block.block_size = frame_size_byte; |
|
data->rx_dma_config.dma_cfg.source_burst_length = |
|
frame_size_byte; |
|
data->rx_dma_config.dma_cfg.dest_burst_length = frame_size_byte; |
|
data->rx_dma_config.dma_cfg.source_data_size = frame_size_byte; |
|
data->rx_dma_config.dma_cfg.dest_data_size = frame_size_byte; |
|
|
|
} else { |
|
data->rx_dma_block.dest_address = (uint32_t)rx_buf; |
|
data->rx_dma_block.source_address = |
|
DSPI_GetRxRegisterAddress(base); |
|
data->rx_dma_block.block_size = rx_size; |
|
data->rx_dma_config.dma_cfg.source_burst_length = |
|
frame_size_byte; |
|
data->rx_dma_config.dma_cfg.dest_burst_length = frame_size_byte; |
|
data->rx_dma_config.dma_cfg.source_data_size = frame_size_byte; |
|
data->rx_dma_config.dma_cfg.dest_data_size = frame_size_byte; |
|
} |
|
|
|
data->rx_dma_config.dma_cfg.user_data = (void *) dev; |
|
dma_config(data->rx_dma_config.dma_dev, data->rx_dma_config.dma_channel, |
|
(struct dma_config *)&data->rx_dma_config.dma_cfg); |
|
|
|
return 0; |
|
} |
|
|
|
static int configure_dma(const struct device *dev) |
|
{ |
|
const struct spi_mcux_config *config = dev->config; |
|
|
|
if (config->is_dma_chn_shared) { |
|
LOG_DBG("shard DMA request"); |
|
} |
|
update_tx_dma(dev); |
|
update_rx_dma(dev); |
|
|
|
return 0; |
|
} |
|
|
|
static void dma_callback(const struct device *dma_dev, void *callback_arg, |
|
uint32_t channel, int error_code) |
|
{ |
|
const struct device *dev = (const struct device *)callback_arg; |
|
const struct spi_mcux_config *config = dev->config; |
|
SPI_Type *base = config->base; |
|
struct spi_mcux_data *data = dev->data; |
|
|
|
LOG_DBG("=dma call back @channel %d=", channel); |
|
|
|
if (error_code < 0) { |
|
LOG_ERR("error happened no callback process %d", error_code); |
|
return; |
|
} |
|
|
|
if (channel == data->tx_dma_config.dma_channel) { |
|
LOG_DBG("ctx.tx_len is %d", data->ctx.tx_len); |
|
LOG_DBG("tx count %d", data->ctx.tx_count); |
|
|
|
spi_context_update_tx(&data->ctx, 1, data->transfer_len); |
|
LOG_DBG("tx count %d", data->ctx.tx_count); |
|
LOG_DBG("tx buf/len %p/%zu", data->ctx.tx_buf, |
|
data->ctx.tx_len); |
|
data->tx_transfer_count++; |
|
/* tx done */ |
|
} else { |
|
LOG_DBG("ctx.rx_len is %d", data->ctx.rx_len); |
|
LOG_DBG("rx count %d", data->ctx.rx_count); |
|
spi_context_update_rx(&data->ctx, 1, data->transfer_len); |
|
LOG_DBG("rx count %d", data->ctx.rx_count); |
|
/* setup the inner tx buffer */ |
|
LOG_DBG("rx buf/len %p/%zu", data->ctx.rx_buf, |
|
data->ctx.rx_len); |
|
data->rx_transfer_count++; |
|
} |
|
|
|
if (data->tx_transfer_count == data->rx_transfer_count) { |
|
LOG_DBG("start next packet"); |
|
DSPI_StopTransfer(base); |
|
DSPI_FlushFifo(base, true, true); |
|
DSPI_ClearStatusFlags(base, |
|
(uint32_t)kDSPI_AllStatusFlag); |
|
mcux_init_inner_buffer_with_cmd(dev, 0); |
|
mcux_spi_context_data_update(dev); |
|
|
|
if (config->is_dma_chn_shared) { |
|
data->transfer_len = data->frame_size >> 3; |
|
} else { |
|
if (data->ctx.tx_len == 0) { |
|
data->transfer_len = data->ctx.rx_len; |
|
} else if (data->ctx.rx_len == 0) { |
|
data->transfer_len = data->ctx.tx_len; |
|
} else { |
|
data->transfer_len = |
|
data->ctx.tx_len > data->ctx.rx_len ? |
|
data->ctx.rx_len : |
|
data->ctx.tx_len; |
|
} |
|
} |
|
update_tx_dma(dev); |
|
update_rx_dma(dev); |
|
spi_mcux_transfer_next_packet(dev); |
|
} else if (data->ctx.rx_len == 0 && data->ctx.tx_len == 0) { |
|
LOG_DBG("end of transfer"); |
|
DSPI_StopTransfer(base); |
|
DSPI_FlushFifo(base, true, true); |
|
DSPI_ClearStatusFlags(base, |
|
(uint32_t)kDSPI_AllStatusFlag); |
|
data->transfer_len = 0; |
|
spi_mcux_transfer_next_packet(dev); |
|
} |
|
LOG_DBG("TX/RX DMA callback done"); |
|
} |
|
|
|
#else |
|
|
|
static void spi_mcux_master_transfer_callback(SPI_Type *base, |
|
dspi_master_handle_t *handle, status_t status, void *userData) |
|
{ |
|
struct spi_mcux_data *data = userData; |
|
|
|
spi_context_update_tx(&data->ctx, 1, data->transfer_len); |
|
spi_context_update_rx(&data->ctx, 1, data->transfer_len); |
|
|
|
spi_mcux_transfer_next_packet(data->dev); |
|
} |
|
|
|
#endif /* CONFIG_DSPI_MCUX_EDMA */ |
|
|
|
static int spi_mcux_configure(const struct device *dev, |
|
const struct spi_config *spi_cfg) |
|
{ |
|
const struct spi_mcux_config *config = dev->config; |
|
struct spi_mcux_data *data = dev->data; |
|
SPI_Type *base = config->base; |
|
dspi_master_config_t master_config; |
|
uint32_t clock_freq; |
|
uint32_t word_size; |
|
|
|
dspi_master_ctar_config_t *ctar_config = &master_config.ctarConfig; |
|
|
|
if (spi_context_configured(&data->ctx, spi_cfg)) { |
|
/* This configuration is already in use */ |
|
return 0; |
|
} |
|
|
|
if (spi_cfg->operation & SPI_HALF_DUPLEX) { |
|
LOG_ERR("Half-duplex not supported"); |
|
return -ENOTSUP; |
|
} |
|
|
|
DSPI_MasterGetDefaultConfig(&master_config); |
|
|
|
master_config.whichPcs = 1U << spi_cfg->slave; |
|
master_config.whichCtar = config->which_ctar; |
|
master_config.pcsActiveHighOrLow = |
|
(spi_cfg->operation & SPI_CS_ACTIVE_HIGH) ? |
|
kDSPI_PcsActiveHigh : |
|
kDSPI_PcsActiveLow; |
|
master_config.samplePoint = config->samplePoint; |
|
master_config.enableContinuousSCK = config->enable_continuous_sck; |
|
master_config.enableRxFifoOverWrite = config->enable_rxfifo_overwrite; |
|
master_config.enableModifiedTimingFormat = |
|
config->enable_modified_timing_format; |
|
|
|
if (spi_cfg->slave > FSL_FEATURE_DSPI_CHIP_SELECT_COUNT) { |
|
LOG_ERR("Slave %d is greater than %d", |
|
spi_cfg->slave, FSL_FEATURE_DSPI_CHIP_SELECT_COUNT); |
|
return -EINVAL; |
|
} |
|
|
|
word_size = SPI_WORD_SIZE_GET(spi_cfg->operation); |
|
if (word_size > FSL_FEATURE_DSPI_MAX_DATA_WIDTH) { |
|
LOG_ERR("Word size %d is greater than %d", |
|
word_size, FSL_FEATURE_DSPI_MAX_DATA_WIDTH); |
|
return -EINVAL; |
|
} |
|
|
|
ctar_config->bitsPerFrame = word_size; |
|
|
|
ctar_config->cpol = |
|
(SPI_MODE_GET(spi_cfg->operation) & SPI_MODE_CPOL) |
|
? kDSPI_ClockPolarityActiveLow |
|
: kDSPI_ClockPolarityActiveHigh; |
|
|
|
ctar_config->cpha = |
|
(SPI_MODE_GET(spi_cfg->operation) & SPI_MODE_CPHA) |
|
? kDSPI_ClockPhaseSecondEdge |
|
: kDSPI_ClockPhaseFirstEdge; |
|
|
|
ctar_config->direction = |
|
(spi_cfg->operation & SPI_TRANSFER_LSB) |
|
? kDSPI_LsbFirst |
|
: kDSPI_MsbFirst; |
|
|
|
ctar_config->baudRate = spi_cfg->frequency; |
|
|
|
ctar_config->pcsToSckDelayInNanoSec = config->pcs_sck_delay; |
|
ctar_config->lastSckToPcsDelayInNanoSec = config->sck_pcs_delay; |
|
ctar_config->betweenTransferDelayInNanoSec = config->transfer_delay; |
|
|
|
if (!device_is_ready(config->clock_dev)) { |
|
LOG_ERR("clock control device not ready"); |
|
return -ENODEV; |
|
} |
|
|
|
if (clock_control_get_rate(config->clock_dev, config->clock_subsys, |
|
&clock_freq)) { |
|
return -EINVAL; |
|
} |
|
|
|
LOG_DBG("clock_freq is %d", clock_freq); |
|
|
|
DSPI_MasterInit(base, &master_config, clock_freq); |
|
|
|
#ifdef CONFIG_DSPI_MCUX_EDMA |
|
DSPI_StopTransfer(base); |
|
DSPI_FlushFifo(base, true, true); |
|
DSPI_ClearStatusFlags(base, (uint32_t)kDSPI_AllStatusFlag); |
|
/* record frame_size setting for DMA */ |
|
data->frame_size = word_size; |
|
/* keep the pcs settings */ |
|
data->which_pcs = 1U << spi_cfg->slave; |
|
#ifdef CONFIG_MCUX_DSPI_EDMA_SHUFFLE_DATA |
|
mcux_init_inner_buffer_with_cmd(dev, 0); |
|
#endif |
|
#else |
|
DSPI_MasterTransferCreateHandle(base, &data->handle, |
|
spi_mcux_master_transfer_callback, |
|
data); |
|
|
|
DSPI_SetDummyData(base, 0); |
|
#endif |
|
|
|
data->ctx.config = spi_cfg; |
|
|
|
return 0; |
|
} |
|
|
|
static int transceive(const struct device *dev, |
|
const struct spi_config *spi_cfg, |
|
const struct spi_buf_set *tx_bufs, |
|
const struct spi_buf_set *rx_bufs, |
|
bool asynchronous, |
|
spi_callback_t cb, |
|
void *userdata) |
|
{ |
|
struct spi_mcux_data *data = dev->data; |
|
int ret; |
|
#ifdef CONFIG_DSPI_MCUX_EDMA |
|
const struct spi_mcux_config *config = dev->config; |
|
SPI_Type *base = config->base; |
|
#endif |
|
|
|
spi_context_lock(&data->ctx, asynchronous, cb, userdata, spi_cfg); |
|
|
|
ret = spi_mcux_configure(dev, spi_cfg); |
|
if (ret) { |
|
goto out; |
|
} |
|
|
|
spi_context_buffers_setup(&data->ctx, tx_bufs, rx_bufs, 1); |
|
|
|
spi_context_cs_control(&data->ctx, true); |
|
|
|
#ifdef CONFIG_DSPI_MCUX_EDMA |
|
DSPI_StopTransfer(base); |
|
DSPI_FlushFifo(base, true, true); |
|
DSPI_ClearStatusFlags(base, (uint32_t)kDSPI_AllStatusFlag); |
|
/* setup the tx buffer with end */ |
|
mcux_init_inner_buffer_with_cmd(dev, 0); |
|
mcux_spi_context_data_update(dev); |
|
if (config->is_dma_chn_shared) { |
|
data->transfer_len = data->frame_size >> 3; |
|
} else { |
|
data->transfer_len = data->ctx.tx_len > data->ctx.rx_len ? |
|
data->ctx.rx_len : |
|
data->ctx.tx_len; |
|
} |
|
data->tx_transfer_count = 0; |
|
data->rx_transfer_count = 0; |
|
configure_dma(dev); |
|
#endif |
|
|
|
ret = spi_mcux_transfer_next_packet(dev); |
|
if (ret) { |
|
goto out; |
|
} |
|
|
|
ret = spi_context_wait_for_completion(&data->ctx); |
|
out: |
|
spi_context_release(&data->ctx, ret); |
|
|
|
return ret; |
|
} |
|
|
|
static int spi_mcux_transceive(const struct device *dev, |
|
const struct spi_config *spi_cfg, |
|
const struct spi_buf_set *tx_bufs, |
|
const struct spi_buf_set *rx_bufs) |
|
{ |
|
return transceive(dev, spi_cfg, tx_bufs, rx_bufs, false, NULL, NULL); |
|
} |
|
|
|
#ifdef CONFIG_SPI_ASYNC |
|
static int spi_mcux_transceive_async(const struct device *dev, |
|
const struct spi_config *spi_cfg, |
|
const struct spi_buf_set *tx_bufs, |
|
const struct spi_buf_set *rx_bufs, |
|
spi_callback_t cb, |
|
void *userdata) |
|
{ |
|
return transceive(dev, spi_cfg, tx_bufs, rx_bufs, true, cb, userdata); |
|
} |
|
#endif /* CONFIG_SPI_ASYNC */ |
|
|
|
static int spi_mcux_release(const struct device *dev, |
|
const struct spi_config *spi_cfg) |
|
{ |
|
struct spi_mcux_data *data = dev->data; |
|
|
|
spi_context_unlock_unconditionally(&data->ctx); |
|
|
|
return 0; |
|
} |
|
|
|
static int spi_mcux_init(const struct device *dev) |
|
{ |
|
int err; |
|
struct spi_mcux_data *data = dev->data; |
|
const struct spi_mcux_config *config = dev->config; |
|
|
|
#ifdef CONFIG_DSPI_MCUX_EDMA |
|
enum dma_channel_filter spi_filter = DMA_CHANNEL_NORMAL; |
|
const struct device *dma_dev; |
|
|
|
dma_dev = data->rx_dma_config.dma_dev; |
|
data->rx_dma_config.dma_channel = |
|
dma_request_channel(dma_dev, (void *)&spi_filter); |
|
dma_dev = data->tx_dma_config.dma_dev; |
|
data->tx_dma_config.dma_channel = |
|
dma_request_channel(dma_dev, (void *)&spi_filter); |
|
#else |
|
config->irq_config_func(dev); |
|
#endif |
|
err = pinctrl_apply_state(config->pincfg, PINCTRL_STATE_DEFAULT); |
|
if (err != 0) { |
|
return err; |
|
} |
|
|
|
data->dev = dev; |
|
|
|
err = spi_context_cs_configure_all(&data->ctx); |
|
if (err < 0) { |
|
return err; |
|
} |
|
|
|
spi_context_unlock_unconditionally(&data->ctx); |
|
|
|
return 0; |
|
} |
|
|
|
static const struct spi_driver_api spi_mcux_driver_api = { |
|
.transceive = spi_mcux_transceive, |
|
#ifdef CONFIG_SPI_ASYNC |
|
.transceive_async = spi_mcux_transceive_async, |
|
#endif |
|
.release = spi_mcux_release, |
|
}; |
|
|
|
|
|
/* if a then b otherwise return 1 */ |
|
#define _UTIL_AND2(a, b) COND_CODE_1(UTIL_BOOL(a), (b), (1)) |
|
|
|
#ifdef CONFIG_DSPI_MCUX_EDMA |
|
|
|
#define TX_BUFFER(id) \ |
|
static uint32_t \ |
|
edma_tx_buffer_##id[CONFIG_MCUX_DSPI_BUFFER_SIZE >> 2]; \ |
|
static struct spi_buf spi_edma_tx_buffer_##id = { \ |
|
.buf = edma_tx_buffer_##id, \ |
|
.len = CONFIG_MCUX_DSPI_BUFFER_SIZE, \ |
|
} |
|
|
|
#define RX_BUFFER(id) \ |
|
static uint32_t \ |
|
edma_rx_buffer_##id[CONFIG_MCUX_DSPI_BUFFER_SIZE >> 2]; \ |
|
static struct spi_buf spi_edma_rx_buffer_##id = { \ |
|
.buf = edma_rx_buffer_##id, \ |
|
.len = CONFIG_MCUX_DSPI_BUFFER_SIZE, \ |
|
} |
|
|
|
#define TX_DMA_CONFIG(id) \ |
|
.inner_tx_buffer = &spi_edma_tx_buffer_##id, \ |
|
.tx_dma_config = { \ |
|
.dma_dev = \ |
|
DEVICE_DT_GET(DT_INST_DMAS_CTLR_BY_NAME(id, tx)), \ |
|
.dma_cfg = { \ |
|
.source_burst_length = 4, \ |
|
.dest_burst_length = 4, \ |
|
.source_data_size = 4, \ |
|
.dest_data_size = 4, \ |
|
.dma_callback = dma_callback, \ |
|
.complete_callback_en = 1, \ |
|
.error_callback_dis = 0, \ |
|
.block_count = 1, \ |
|
.head_block = &spi_mcux_data_##id.tx_dma_block, \ |
|
.channel_direction = MEMORY_TO_PERIPHERAL, \ |
|
.dma_slot = DT_INST_DMAS_CELL_BY_NAME( \ |
|
id, tx, source), \ |
|
}, \ |
|
}, |
|
|
|
#define RX_DMA_CONFIG(id) \ |
|
.inner_rx_buffer = &spi_edma_rx_buffer_##id, \ |
|
.rx_dma_config = { \ |
|
.dma_dev = \ |
|
DEVICE_DT_GET(DT_INST_DMAS_CTLR_BY_NAME(id, rx)), \ |
|
.dma_cfg = { \ |
|
.source_burst_length = 2, \ |
|
.dest_burst_length = 2, \ |
|
.source_data_size = 2, \ |
|
.dest_data_size = 2, \ |
|
.dma_callback = dma_callback, \ |
|
.complete_callback_en = 1, \ |
|
.error_callback_dis = 0, \ |
|
.block_count = \ |
|
_UTIL_AND2(DT_INST_NODE_HAS_PROP( \ |
|
id, nxp_rx_tx_chn_share), 2), \ |
|
.head_block = &spi_mcux_data_##id.rx_dma_block, \ |
|
.channel_direction = PERIPHERAL_TO_MEMORY, \ |
|
.dma_slot = DT_INST_DMAS_CELL_BY_NAME( \ |
|
id, rx, source), \ |
|
}, \ |
|
}, |
|
#else |
|
#define TX_BUFFER(id) |
|
#define RX_BUFFER(id) |
|
#define TX_DMA_CONFIG(id) |
|
#define RX_DMA_CONFIG(id) |
|
|
|
#endif |
|
|
|
#define SPI_MCUX_DSPI_DEVICE(id) \ |
|
PINCTRL_DT_INST_DEFINE(id); \ |
|
static void spi_mcux_config_func_##id(const struct device *dev);\ |
|
TX_BUFFER(id); \ |
|
RX_BUFFER(id); \ |
|
static struct spi_mcux_data spi_mcux_data_##id = { \ |
|
SPI_CONTEXT_INIT_LOCK(spi_mcux_data_##id, ctx), \ |
|
SPI_CONTEXT_INIT_SYNC(spi_mcux_data_##id, ctx), \ |
|
SPI_CONTEXT_CS_GPIOS_INITIALIZE(DT_DRV_INST(id), ctx) \ |
|
TX_DMA_CONFIG(id) RX_DMA_CONFIG(id) \ |
|
}; \ |
|
static const struct spi_mcux_config spi_mcux_config_##id = { \ |
|
.base = (SPI_Type *)DT_INST_REG_ADDR(id), \ |
|
.clock_dev = DEVICE_DT_GET(DT_INST_CLOCKS_CTLR(id)), \ |
|
.clock_subsys = \ |
|
(clock_control_subsys_t)DT_INST_CLOCKS_CELL(id, name), \ |
|
.irq_config_func = spi_mcux_config_func_##id, \ |
|
.pcs_sck_delay = \ |
|
DT_INST_PROP_OR(id, pcs_sck_delay, 0), \ |
|
.sck_pcs_delay = \ |
|
DT_INST_PROP_OR(id, sck_pcs_delay, 0), \ |
|
.transfer_delay = \ |
|
DT_INST_PROP_OR(id, transfer_delay, 0), \ |
|
.which_ctar = \ |
|
DT_INST_PROP_OR(id, ctar, 0), \ |
|
.samplePoint = \ |
|
DT_INST_PROP_OR(id, sample_point, 0), \ |
|
.enable_continuous_sck = \ |
|
DT_INST_PROP(id, continuous_sck), \ |
|
.enable_rxfifo_overwrite = \ |
|
DT_INST_PROP(id, rx_fifo_overwrite), \ |
|
.enable_modified_timing_format = \ |
|
DT_INST_PROP(id, modified_timing_format), \ |
|
.is_dma_chn_shared = \ |
|
DT_INST_PROP(id, nxp_rx_tx_chn_share), \ |
|
.pincfg = PINCTRL_DT_INST_DEV_CONFIG_GET(id), \ |
|
}; \ |
|
DEVICE_DT_INST_DEFINE(id, \ |
|
&spi_mcux_init, \ |
|
NULL, \ |
|
&spi_mcux_data_##id, \ |
|
&spi_mcux_config_##id, \ |
|
POST_KERNEL, \ |
|
CONFIG_SPI_INIT_PRIORITY, \ |
|
&spi_mcux_driver_api); \ |
|
static void spi_mcux_config_func_##id(const struct device *dev) \ |
|
{ \ |
|
IRQ_CONNECT(DT_INST_IRQN(id), \ |
|
DT_INST_IRQ(id, priority), \ |
|
spi_mcux_isr, DEVICE_DT_INST_GET(id), \ |
|
0); \ |
|
irq_enable(DT_INST_IRQN(id)); \ |
|
} |
|
|
|
DT_INST_FOREACH_STATUS_OKAY(SPI_MCUX_DSPI_DEVICE)
|
|
|