/* * Copyright (c) 2023 Fabian Blatz * Copyright (c) 2024 grandcentrix GmbH * * SPDX-License-Identifier: Apache-2.0 */ #define DT_DRV_COMPAT zephyr_uart_emul #include #include #include #include #include #include #include #include LOG_MODULE_REGISTER(uart_emul, CONFIG_UART_LOG_LEVEL); struct uart_emul_config { /* emul_list has to be the first member */ struct emul_list_for_bus emul_list; bool loopback; size_t latch_buffer_size; }; BUILD_ASSERT(offsetof(struct uart_emul_config, emul_list) == 0); /* Device run time data */ struct uart_emul_data { /* List of struct uart_emul associated with the device */ sys_slist_t emuls; const struct device *dev; struct uart_config cfg; int errors; struct ring_buf *rx_rb; struct k_spinlock rx_lock; uart_emul_callback_tx_data_ready_t tx_data_ready_cb; void *user_data; struct ring_buf *tx_rb; struct k_spinlock tx_lock; #ifdef CONFIG_UART_INTERRUPT_DRIVEN bool rx_irq_en; bool tx_irq_en; struct k_work irq_work; uart_irq_callback_user_data_t irq_cb; void *irq_cb_udata; #endif /* CONFIG_UART_INTERRUPT_DRIVEN */ #ifdef CONFIG_UART_ASYNC_API bool rx_async_en; bool rx_stopping; bool rx_release_on_timeout; struct k_work tx_work; struct k_work rx_work; struct k_work rx_disable_work; struct k_work_delayable rx_timeout_work; uart_callback_t uart_callback; void *callback_user_data; const uint8_t *tx_buf; size_t tx_buf_len; size_t tx_buf_offset; uint8_t *rx_buf; size_t rx_buf_len; size_t rx_buf_offset; size_t rx_buf_data_len; int32_t rx_buf_timeout; uint8_t *rx_buf_next; size_t rx_buf_next_len; #endif /* CONFIG_UART_ASYNC_API */ }; /* * Define local thread to emulate different thread priorities. * * A UART driver may call back from within a thread with higher or lower priority * than the thread calling the UART API. This can hide potential concurrency issues, * especially if the thread priorities are the same, or even using the same thread * in case the system work queue. */ K_THREAD_STACK_DEFINE(uart_emul_stack_area, CONFIG_UART_EMUL_WORK_Q_STACK_SIZE); struct k_work_q uart_emul_work_q; int uart_emul_init_work_q(void) { k_work_queue_init(&uart_emul_work_q); k_work_queue_start(&uart_emul_work_q, uart_emul_stack_area, K_THREAD_STACK_SIZEOF(uart_emul_stack_area), CONFIG_UART_EMUL_WORK_Q_PRIORITY, NULL); return 0; } SYS_INIT(uart_emul_init_work_q, POST_KERNEL, 0); static void uart_emul_tx_data_ready(const struct device *dev) { struct uart_emul_data *data = dev->data; sys_snode_t *node; if (data->tx_data_ready_cb) { (data->tx_data_ready_cb)(dev, ring_buf_size_get(data->tx_rb), data->user_data); } SYS_SLIST_FOR_EACH_NODE(&data->emuls, node) { struct uart_emul *emul = CONTAINER_OF(node, struct uart_emul, node); __ASSERT_NO_MSG(emul->api != NULL); __ASSERT_NO_MSG(emul->api->tx_data_ready != NULL); emul->api->tx_data_ready(dev, ring_buf_size_get(data->tx_rb), emul->target); } } static int uart_emul_poll_in(const struct device *dev, unsigned char *p_char) { struct uart_emul_data *drv_data = dev->data; k_spinlock_key_t key; uint32_t read; key = k_spin_lock(&drv_data->rx_lock); read = ring_buf_get(drv_data->rx_rb, p_char, 1); k_spin_unlock(&drv_data->rx_lock, key); if (!read) { LOG_DBG("Rx buffer is empty"); return -1; } return 0; } static void uart_emul_poll_out(const struct device *dev, unsigned char out_char) { struct uart_emul_data *drv_data = dev->data; const struct uart_emul_config *drv_cfg = dev->config; k_spinlock_key_t key; uint32_t written; key = k_spin_lock(&drv_data->tx_lock); written = ring_buf_put(drv_data->tx_rb, &out_char, 1); k_spin_unlock(&drv_data->tx_lock, key); if (!written) { LOG_DBG("Tx buffer is full"); return; } if (drv_cfg->loopback) { uart_emul_put_rx_data(dev, &out_char, 1); } uart_emul_tx_data_ready(dev); } static int uart_emul_err_check(const struct device *dev) { struct uart_emul_data *drv_data = dev->data; int errors = drv_data->errors; drv_data->errors = 0; return errors; } #ifdef CONFIG_UART_USE_RUNTIME_CONFIGURE static int uart_emul_configure(const struct device *dev, const struct uart_config *cfg) { struct uart_emul_data *drv_data = dev->data; memcpy(&drv_data->cfg, cfg, sizeof(struct uart_config)); return 0; } static int uart_emul_config_get(const struct device *dev, struct uart_config *cfg) { const struct uart_emul_data *drv_data = dev->data; memcpy(cfg, &drv_data->cfg, sizeof(struct uart_config)); return 0; } #endif /* CONFIG_UART_USE_RUNTIME_CONFIGURE */ #ifdef CONFIG_UART_INTERRUPT_DRIVEN static int uart_emul_fifo_fill(const struct device *dev, const uint8_t *tx_data, int size) { int ret; struct uart_emul_data *data = dev->data; const struct uart_emul_config *config = dev->config; uint32_t put_size = MIN(config->latch_buffer_size, size); K_SPINLOCK(&data->tx_lock) { ret = ring_buf_put(data->tx_rb, tx_data, put_size); } if (config->loopback) { uart_emul_put_rx_data(dev, (uint8_t *)tx_data, put_size); } uart_emul_tx_data_ready(dev); return ret; } static int uart_emul_fifo_read(const struct device *dev, uint8_t *rx_data, int size) { struct uart_emul_data *data = dev->data; const struct uart_emul_config *config = dev->config; uint32_t bytes_to_read; K_SPINLOCK(&data->rx_lock) { bytes_to_read = MIN(config->latch_buffer_size, ring_buf_size_get(data->rx_rb)); bytes_to_read = MIN(bytes_to_read, size); ring_buf_get(data->rx_rb, rx_data, bytes_to_read); } return bytes_to_read; } static int uart_emul_irq_tx_ready(const struct device *dev) { bool ready = false; struct uart_emul_data *data = dev->data; K_SPINLOCK(&data->tx_lock) { if (!data->tx_irq_en) { K_SPINLOCK_BREAK; } ready = ring_buf_space_get(data->tx_rb) > 0; } return ready; } static int uart_emul_irq_rx_ready(const struct device *dev) { bool ready = false; struct uart_emul_data *data = dev->data; K_SPINLOCK(&data->rx_lock) { if (!data->rx_irq_en) { K_SPINLOCK_BREAK; } ready = !ring_buf_is_empty(data->rx_rb); } return ready; } static void uart_emul_irq_handler(struct k_work *work) { struct uart_emul_data *data = CONTAINER_OF(work, struct uart_emul_data, irq_work); const struct device *dev = data->dev; uart_irq_callback_user_data_t cb = data->irq_cb; void *udata = data->irq_cb_udata; if (cb == NULL) { LOG_DBG("No IRQ callback configured for uart_emul device %p", dev); return; } while (true) { bool have_work = false; K_SPINLOCK(&data->tx_lock) { if (!data->tx_irq_en) { K_SPINLOCK_BREAK; } have_work = have_work || ring_buf_space_get(data->tx_rb) > 0; } K_SPINLOCK(&data->rx_lock) { if (!data->rx_irq_en) { K_SPINLOCK_BREAK; } have_work = have_work || !ring_buf_is_empty(data->rx_rb); } if (!have_work) { break; } cb(dev, udata); } } static int uart_emul_irq_is_pending(const struct device *dev) { return uart_emul_irq_tx_ready(dev) || uart_emul_irq_rx_ready(dev); } static void uart_emul_irq_tx_enable(const struct device *dev) { bool submit_irq_work; struct uart_emul_data *const data = dev->data; K_SPINLOCK(&data->tx_lock) { data->tx_irq_en = true; submit_irq_work = ring_buf_space_get(data->tx_rb) > 0; } if (submit_irq_work) { (void)k_work_submit_to_queue(&uart_emul_work_q, &data->irq_work); } } static void uart_emul_irq_rx_enable(const struct device *dev) { bool submit_irq_work; struct uart_emul_data *const data = dev->data; K_SPINLOCK(&data->rx_lock) { data->rx_irq_en = true; submit_irq_work = !ring_buf_is_empty(data->rx_rb); } if (submit_irq_work) { (void)k_work_submit_to_queue(&uart_emul_work_q, &data->irq_work); } } static void uart_emul_irq_tx_disable(const struct device *dev) { struct uart_emul_data *const data = dev->data; K_SPINLOCK(&data->tx_lock) { data->tx_irq_en = false; } } static void uart_emul_irq_rx_disable(const struct device *dev) { struct uart_emul_data *const data = dev->data; K_SPINLOCK(&data->rx_lock) { data->rx_irq_en = false; } } static int uart_emul_irq_tx_complete(const struct device *dev) { bool tx_complete = false; struct uart_emul_data *const data = dev->data; K_SPINLOCK(&data->tx_lock) { tx_complete = ring_buf_is_empty(data->tx_rb); } return tx_complete; } static void uart_emul_irq_callback_set(const struct device *dev, uart_irq_callback_user_data_t cb, void *user_data) { struct uart_emul_data *const data = dev->data; data->irq_cb = cb; data->irq_cb_udata = user_data; } static int uart_emul_irq_update(const struct device *dev) { return 1; } #endif /* CONFIG_UART_INTERRUPT_DRIVEN */ #ifdef CONFIG_UART_ASYNC_API static void uart_emul_post_event(const struct device *dev, struct uart_event *evt) { struct uart_emul_data *data = dev->data; if (!data->uart_callback) { LOG_DBG("No async callback configured for uart_emul device %p", dev); } data->uart_callback(dev, evt, data->callback_user_data); } static void uart_emul_simple_event(const struct device *dev, enum uart_event_type type) { uart_emul_post_event(dev, &(struct uart_event){.type = type}); } static void uart_emul_async_switch_buf_nolock(struct uart_emul_data *data) { data->rx_buf = data->rx_buf_next; data->rx_buf_len = data->rx_buf_next_len; data->rx_buf_offset = 0; data->rx_buf_data_len = 0; data->rx_buf_next = NULL; data->rx_buf_next_len = 0; } static void uart_emul_async_rx_timeout_handler(struct k_work *_work) { struct k_work_delayable *work = k_work_delayable_from_work(_work); struct uart_emul_data *data = CONTAINER_OF(work, struct uart_emul_data, rx_timeout_work); const struct device *dev = data->dev; uint8_t *rx_buf; size_t rx_buf_len; size_t rx_buf_offset; size_t rx_buf_data_len; bool rx_en; bool rx_buf_released = false; bool rx_stopped = false; K_SPINLOCK(&data->rx_lock) { rx_en = data->rx_async_en; rx_buf = data->rx_buf; rx_buf_len = data->rx_buf_len; rx_buf_offset = data->rx_buf_offset; rx_buf_data_len = data->rx_buf_data_len; data->rx_buf_offset += rx_buf_data_len; data->rx_buf_data_len = 0; if (data->rx_buf_offset >= rx_buf_len || (rx_buf_data_len > 0 && data->rx_release_on_timeout)) { rx_buf_released = true; uart_emul_async_switch_buf_nolock(data); if (data->rx_buf == NULL) { /* There was no second buffer scheduled, so stop receiving */ rx_stopped = true; data->rx_async_en = false; } } } if (!rx_en || rx_buf == NULL || rx_buf_data_len == 0) { return; } struct uart_event rx_rdy_event = { .type = UART_RX_RDY, .data.rx = { .buf = rx_buf, .offset = rx_buf_offset, .len = rx_buf_data_len, }, }; uart_emul_post_event(dev, &rx_rdy_event); if (rx_buf_released) { struct uart_event rx_buf_released_event = { .type = UART_RX_BUF_RELEASED, .data.rx_buf.buf = rx_buf, }; uart_emul_post_event(dev, &rx_buf_released_event); } if (rx_stopped) { uart_emul_simple_event(dev, UART_RX_DISABLED); } } static void uart_emul_async_rx_handler(struct k_work *work) { struct uart_emul_data *data = CONTAINER_OF(work, struct uart_emul_data, rx_work); const struct device *dev = data->dev; bool rx_en = false; bool empty = true; do { bool rx_rdy = false; bool buf_request = false; uint8_t *rx_buf = NULL; size_t buf_len; size_t offset; size_t data_len; K_SPINLOCK(&data->rx_lock) { rx_en = data->rx_async_en; rx_buf = data->rx_buf; buf_len = data->rx_buf_len; offset = data->rx_buf_offset; data_len = data->rx_buf_data_len; empty = ring_buf_is_empty(data->rx_rb); if (!rx_en) { K_SPINLOCK_BREAK; } if (rx_buf == NULL) { uart_emul_async_switch_buf_nolock(data); rx_buf = data->rx_buf; buf_len = data->rx_buf_len; offset = data->rx_buf_offset; data_len = data->rx_buf_data_len; } if (rx_buf == NULL) { /* During the last iteration the buffer was released but the * application did not provide a new buffer. Stop RX and quit now. */ data->rx_async_en = false; K_SPINLOCK_BREAK; } if (empty) { K_SPINLOCK_BREAK; } buf_request = data_len == 0 && data->rx_buf_next == NULL; uint32_t read = ring_buf_get(data->rx_rb, &rx_buf[offset + data_len], buf_len - (offset + data_len)); data_len += read; data->rx_buf_data_len = data_len; if (offset + data_len >= data->rx_buf_len) { rx_rdy = true; data->rx_buf = NULL; data->rx_buf_len = 0; data->rx_buf_offset = 0; data->rx_buf_data_len = 0; } } if (!rx_en) { break; } if (rx_buf == NULL) { uart_emul_simple_event(dev, UART_RX_DISABLED); break; } if (empty && data->rx_buf_timeout != SYS_FOREVER_US) { (void)k_work_reschedule_for_queue(&uart_emul_work_q, &data->rx_timeout_work, K_USEC(data->rx_buf_timeout)); } if (buf_request) { uart_emul_simple_event(dev, UART_RX_BUF_REQUEST); } if (rx_rdy) { struct uart_event rx_rdy_event = { .type = UART_RX_RDY, .data.rx = { .buf = rx_buf, .offset = offset, .len = data_len, }, }; uart_emul_post_event(dev, &rx_rdy_event); struct uart_event rx_buf_released_event = { .type = UART_RX_BUF_RELEASED, .data.rx_buf.buf = rx_buf, }; uart_emul_post_event(dev, &rx_buf_released_event); } } while (rx_en && !empty); } static void uart_emul_async_tx_handler(struct k_work *work) { struct uart_emul_data *data = CONTAINER_OF(work, struct uart_emul_data, tx_work); const struct device *dev = data->dev; const struct uart_emul_config *config = dev->config; uint32_t written; const uint8_t *tx_buf = NULL; size_t tx_buf_len = 0; size_t tx_buf_offset = 0; bool tx_done = true; K_SPINLOCK(&data->tx_lock) { tx_buf = data->tx_buf; tx_buf_len = data->tx_buf_len; tx_buf_offset = data->tx_buf_offset; if (!tx_buf) { K_SPINLOCK_BREAK; } written = ring_buf_put(data->tx_rb, &data->tx_buf[tx_buf_offset], tx_buf_len - tx_buf_offset); tx_done = written == (tx_buf_len - tx_buf_offset); if (!tx_done) { data->tx_buf_offset += written; K_SPINLOCK_BREAK; } data->tx_buf = NULL; data->tx_buf_len = 0; data->tx_buf_offset = 0; } if (!tx_buf) { return; } if (config->loopback && written) { uint32_t loop_written = uart_emul_put_rx_data(dev, &tx_buf[tx_buf_offset], written); if (loop_written < written) { LOG_WRN("Lost %" PRIu32 " bytes on loopback", written - loop_written); } } uart_emul_tx_data_ready(dev); if ((config->loopback && written) || !written) { /* When using the loopback fixture, just allow to drop all bytes in the ring buffer * not consumed by tx_data_ready_cb(). */ uint32_t flushed = uart_emul_flush_tx_data(dev); if (flushed) { if (written) { LOG_DBG("Flushed %" PRIu32 " unused bytes from tx buffer", flushed); } else { LOG_WRN("Flushed %" PRIu32 " unused bytes from tx buffer to break out of infinite " "loop! Consume or flush the bytes from the tx ring buffer " "in your test case to prevent this!", flushed); } } } if (!tx_done) { /* We are not done yet, yield back into workqueue. * * This would basically be an infinite loop when tx_data_ready_cb() does not consume * the bytes in the tx ring buffer. */ k_work_submit_to_queue(&uart_emul_work_q, &data->tx_work); return; } struct uart_event tx_done_event = { .type = UART_TX_DONE, .data.tx = { .buf = tx_buf, .len = tx_buf_len, }, }; uart_emul_post_event(dev, &tx_done_event); } static void uart_emul_rx_stop(const struct device *dev, struct uart_emul_data *data) { uint8_t *rx_buf = NULL; size_t rx_buf_offset = 0; size_t rx_buf_data_len = 0; k_work_cancel_delayable(&data->rx_timeout_work); K_SPINLOCK(&data->rx_lock) { if (!data->rx_async_en) { K_SPINLOCK_BREAK; } rx_buf = data->rx_buf; rx_buf_offset = data->rx_buf_offset; rx_buf_data_len = data->rx_buf_data_len; data->rx_buf = NULL; data->rx_buf_len = 0; data->rx_buf_offset = 0; data->rx_buf_data_len = 0; data->rx_buf_next = NULL; data->rx_buf_next_len = 0; data->rx_async_en = false; data->rx_stopping = false; } if (rx_buf == NULL) { return; } if (rx_buf_data_len > 0) { struct uart_event rx_rdy_event = { .type = UART_RX_RDY, .data.rx = { .buf = rx_buf, .offset = rx_buf_offset, .len = rx_buf_data_len, }, }; uart_emul_post_event(dev, &rx_rdy_event); } struct uart_event rx_buf_released_event = { .type = UART_RX_BUF_RELEASED, .data.rx_buf.buf = rx_buf, }; uart_emul_post_event(dev, &rx_buf_released_event); uart_emul_simple_event(dev, UART_RX_DISABLED); } static void uart_emul_async_rx_disable_handler(struct k_work *work) { struct uart_emul_data *data = CONTAINER_OF(work, struct uart_emul_data, rx_disable_work); const struct device *dev = data->dev; uart_emul_rx_stop(dev, data); } static int uart_emul_callback_set(const struct device *dev, uart_callback_t callback, void *user_data) { struct uart_emul_data *data = dev->data; data->uart_callback = callback; data->callback_user_data = user_data; return 0; } static int uart_emul_tx(const struct device *dev, const uint8_t *buf, size_t len, int32_t timeout) { struct uart_emul_data *data = dev->data; int ret = 0; K_SPINLOCK(&data->tx_lock) { if (data->tx_buf) { ret = -EBUSY; K_SPINLOCK_BREAK; } data->tx_buf = buf; data->tx_buf_len = len; data->tx_buf_offset = 0; k_work_submit_to_queue(&uart_emul_work_q, &data->tx_work); } return ret; } static int uart_emul_tx_abort(const struct device *dev) { struct uart_emul_data *data = dev->data; const uint8_t *tx_buf = NULL; size_t tx_buf_sent; K_SPINLOCK(&data->tx_lock) { tx_buf = data->tx_buf; tx_buf_sent = data->tx_buf_offset; data->tx_buf = NULL; data->tx_buf_len = 0; data->tx_buf_offset = 0; k_work_cancel(&data->tx_work); } if (!tx_buf) { return -EFAULT; } struct uart_event tx_aborted_event = { .type = UART_TX_ABORTED, .data.tx = { .buf = tx_buf, .len = tx_buf_sent, }, }; uart_emul_post_event(dev, &tx_aborted_event); return 0; } static int uart_emul_rx_buf_rsp(const struct device *dev, uint8_t *buf, size_t len) { struct uart_emul_data *data = dev->data; int ret = 0; K_SPINLOCK(&data->rx_lock) { if (!data->rx_async_en) { ret = -EACCES; K_SPINLOCK_BREAK; } if (data->rx_buf_next != NULL) { ret = -EBUSY; K_SPINLOCK_BREAK; } data->rx_buf_next = buf; data->rx_buf_next_len = len; } return ret; } static int uart_emul_rx_enable(const struct device *dev, uint8_t *buf, size_t len, int32_t timeout) { struct uart_emul_data *data = dev->data; int ret = 0; bool rx_stopping; K_SPINLOCK(&data->rx_lock) { rx_stopping = data->rx_stopping; k_work_cancel(&data->rx_disable_work); } if (rx_stopping) { uart_emul_rx_stop(dev, data); } K_SPINLOCK(&data->rx_lock) { if (data->rx_async_en) { ret = -EBUSY; K_SPINLOCK_BREAK; } data->rx_async_en = true; data->rx_buf = buf; data->rx_buf_len = len; data->rx_buf_timeout = timeout; data->rx_buf_offset = 0; data->rx_buf_data_len = 0; data->rx_buf_next = NULL; data->rx_buf_next_len = 0; if (!ring_buf_is_empty(data->rx_rb)) { (void)k_work_submit_to_queue(&uart_emul_work_q, &data->rx_work); } } return ret; } static int uart_emul_rx_disable(const struct device *dev) { struct uart_emul_data *data = dev->data; int ret = 0; K_SPINLOCK(&data->rx_lock) { if (!data->rx_async_en) { ret = -EFAULT; K_SPINLOCK_BREAK; } data->rx_stopping = true; k_work_submit_to_queue(&uart_emul_work_q, &data->rx_disable_work); } return ret; } #endif /* CONFIG_UART_ASYNC_API */ static const struct uart_driver_api uart_emul_api = { .poll_in = uart_emul_poll_in, .poll_out = uart_emul_poll_out, #ifdef CONFIG_UART_USE_RUNTIME_CONFIGURE .config_get = uart_emul_config_get, .configure = uart_emul_configure, #endif /* CONFIG_UART_USE_RUNTIME_CONFIGURE */ .err_check = uart_emul_err_check, #ifdef CONFIG_UART_INTERRUPT_DRIVEN .fifo_fill = uart_emul_fifo_fill, .fifo_read = uart_emul_fifo_read, .irq_tx_enable = uart_emul_irq_tx_enable, .irq_rx_enable = uart_emul_irq_rx_enable, .irq_tx_disable = uart_emul_irq_tx_disable, .irq_rx_disable = uart_emul_irq_rx_disable, .irq_tx_ready = uart_emul_irq_tx_ready, .irq_rx_ready = uart_emul_irq_rx_ready, .irq_tx_complete = uart_emul_irq_tx_complete, .irq_callback_set = uart_emul_irq_callback_set, .irq_update = uart_emul_irq_update, .irq_is_pending = uart_emul_irq_is_pending, #endif /* CONFIG_UART_INTERRUPT_DRIVEN */ #ifdef CONFIG_UART_ASYNC_API .callback_set = uart_emul_callback_set, .tx = uart_emul_tx, .tx_abort = uart_emul_tx_abort, .rx_enable = uart_emul_rx_enable, .rx_buf_rsp = uart_emul_rx_buf_rsp, .rx_disable = uart_emul_rx_disable, #endif /* CONFIG_UART_ASYNC_API */ }; void uart_emul_callback_tx_data_ready_set(const struct device *dev, uart_emul_callback_tx_data_ready_t cb, void *user_data) { struct uart_emul_data *drv_data = dev->data; drv_data->tx_data_ready_cb = cb; drv_data->user_data = user_data; } uint32_t uart_emul_put_rx_data(const struct device *dev, const uint8_t *data, size_t size) { struct uart_emul_data *drv_data = dev->data; uint32_t count; __unused bool empty; __unused bool irq_en; __unused bool rx_en; K_SPINLOCK(&drv_data->rx_lock) { count = ring_buf_put(drv_data->rx_rb, data, size); empty = ring_buf_is_empty(drv_data->rx_rb); IF_ENABLED(CONFIG_UART_INTERRUPT_DRIVEN, (irq_en = drv_data->rx_irq_en;)); IF_ENABLED(CONFIG_UART_ASYNC_API, (rx_en = drv_data->rx_async_en;)); } if (count < size) { uart_emul_set_errors(dev, UART_ERROR_OVERRUN); } IF_ENABLED(CONFIG_UART_INTERRUPT_DRIVEN, ( if (count > 0 && irq_en && !empty) { (void)k_work_submit_to_queue(&uart_emul_work_q, &drv_data->irq_work); } )) IF_ENABLED(CONFIG_UART_ASYNC_API, ( if (count > 0 && rx_en && !empty) { (void)k_work_submit_to_queue(&uart_emul_work_q, &drv_data->rx_work); } )) return count; } uint32_t uart_emul_get_tx_data(const struct device *dev, uint8_t *data, size_t size) { struct uart_emul_data *drv_data = dev->data; k_spinlock_key_t key; uint32_t count; key = k_spin_lock(&drv_data->tx_lock); count = ring_buf_get(drv_data->tx_rb, data, size); k_spin_unlock(&drv_data->tx_lock, key); return count; } uint32_t uart_emul_flush_rx_data(const struct device *dev) { struct uart_emul_data *drv_data = dev->data; k_spinlock_key_t key; uint32_t count; key = k_spin_lock(&drv_data->rx_lock); count = ring_buf_size_get(drv_data->rx_rb); ring_buf_reset(drv_data->rx_rb); k_spin_unlock(&drv_data->rx_lock, key); return count; } uint32_t uart_emul_flush_tx_data(const struct device *dev) { struct uart_emul_data *drv_data = dev->data; k_spinlock_key_t key; uint32_t count; key = k_spin_lock(&drv_data->tx_lock); count = ring_buf_size_get(drv_data->tx_rb); ring_buf_reset(drv_data->tx_rb); k_spin_unlock(&drv_data->tx_lock, key); return count; } void uart_emul_set_errors(const struct device *dev, int errors) { struct uart_emul_data *drv_data = dev->data; drv_data->errors |= errors; } void uart_emul_set_release_buffer_on_timeout(const struct device *dev, bool release_on_timeout) { __unused struct uart_emul_data *drv_data = dev->data; IF_ENABLED(CONFIG_UART_ASYNC_API, (drv_data->rx_release_on_timeout = release_on_timeout;)); } int uart_emul_register(const struct device *dev, struct uart_emul *emul) { struct uart_emul_data *data = dev->data; sys_slist_append(&data->emuls, &emul->node); return 0; } #define UART_EMUL_RX_FIFO_SIZE(inst) (DT_INST_PROP(inst, rx_fifo_size)) #define UART_EMUL_TX_FIFO_SIZE(inst) (DT_INST_PROP(inst, tx_fifo_size)) #define EMUL_LINK_AND_COMMA(node_id) \ { \ .dev = DEVICE_DT_GET(node_id), \ }, #define DEFINE_UART_EMUL(inst) \ static const struct emul_link_for_bus emuls_##inst[] = { \ DT_FOREACH_CHILD_STATUS_OKAY(DT_DRV_INST(inst), EMUL_LINK_AND_COMMA)}; \ \ RING_BUF_DECLARE(uart_emul_##inst##_rx_rb, UART_EMUL_RX_FIFO_SIZE(inst)); \ RING_BUF_DECLARE(uart_emul_##inst##_tx_rb, UART_EMUL_TX_FIFO_SIZE(inst)); \ \ static const struct uart_emul_config uart_emul_cfg_##inst = { \ .loopback = DT_INST_PROP(inst, loopback), \ .latch_buffer_size = DT_INST_PROP(inst, latch_buffer_size), \ .emul_list = { \ .children = emuls_##inst, \ .num_children = ARRAY_SIZE(emuls_##inst), \ }, \ }; \ static struct uart_emul_data uart_emul_data_##inst = { \ .emuls = SYS_SLIST_STATIC_INIT(&_CONCAT(uart_emul_data_, inst).emuls), \ .dev = DEVICE_DT_INST_GET(inst), \ .rx_rb = &uart_emul_##inst##_rx_rb, \ .tx_rb = &uart_emul_##inst##_tx_rb, \ IF_ENABLED(CONFIG_UART_INTERRUPT_DRIVEN, \ (.irq_work = Z_WORK_INITIALIZER(uart_emul_irq_handler),)) \ IF_ENABLED(CONFIG_UART_ASYNC_API, \ (.tx_work = Z_WORK_INITIALIZER(uart_emul_async_tx_handler), \ .rx_timeout_work = Z_WORK_DELAYABLE_INITIALIZER( \ uart_emul_async_rx_timeout_handler), \ .rx_work = Z_WORK_INITIALIZER(uart_emul_async_rx_handler), \ .rx_disable_work = Z_WORK_INITIALIZER( \ uart_emul_async_rx_disable_handler),)) \ }; \ \ static int uart_emul_post_init_##inst(void) \ { \ return emul_init_for_bus(DEVICE_DT_INST_GET(inst)); \ } \ SYS_INIT(uart_emul_post_init_##inst, POST_KERNEL, CONFIG_UART_EMUL_DEVICE_INIT_PRIORITY); \ \ DEVICE_DT_INST_DEFINE(inst, NULL, NULL, &uart_emul_data_##inst, &uart_emul_cfg_##inst, \ PRE_KERNEL_1, CONFIG_SERIAL_INIT_PRIORITY, &uart_emul_api); DT_INST_FOREACH_STATUS_OKAY(DEFINE_UART_EMUL)