mirror of
				https://github.com/espressif/esp-idf.git
				synced 2025-10-29 14:11:42 +01:00 
			
		
		
		
	portBASE_TYPE is an internal macro defined by the porting layer. This commit changes all references to BaseType_t which is the official type exposed by FreeRTOS.
		
			
				
	
	
		
			819 lines
		
	
	
		
			27 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			819 lines
		
	
	
		
			27 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * SPDX-FileCopyrightText: 2015-2023 Espressif Systems (Shanghai) CO LTD
 | |
|  *
 | |
|  * SPDX-License-Identifier: Apache-2.0
 | |
|  */
 | |
| 
 | |
| /*
 | |
| Architecture:
 | |
| 
 | |
| The whole SDIO slave peripheral consists of three parts: the registers (including the control registers of
 | |
| interrupts and shared registers), the sending FIFO and the receiving FIFO. A document ``esp_slave_protocol.rst``
 | |
| describes the functionality of the peripheral detailedly.
 | |
| 
 | |
| The host can access only one of those parts at once, and the hardware functions of these parts are totally
 | |
| independent. Hence this driver is designed into these three independent parts. The shared registers are quite
 | |
| simple. As well as the interrupts: when a slave interrupt is written by the host, the slave gets an interrupt;
 | |
| when one of the host interrupt bits is active, slave hardware output interrupt signals on the DAT1 line.
 | |
| 
 | |
| For the FIFOs, the peripheral provides counters as registers so that the host can always know whether the slave
 | |
| is ready to send/receive data. The driver resets the counters during initialization, and the host should somehow
 | |
| inform the slave to reset the counters again if it should reboot (or lose the counter value for some reasons).
 | |
| Then the host can read/write the FIFOs by CMD53 commands according to the counters.
 | |
| 
 | |
| Since we don't want to copy all the data from the buffer each time we use sending/receiving buffer,
 | |
| the buffers are directly loaded onto the sending/receiving linked-list and taken off only after use.
 | |
| Hence the driver takes ownership of the buffer when the buffer is fed to the driver.
 | |
| 
 | |
| The driver returns the ownership of buffers when a "finish" function is called. When the hardware finishes
 | |
| the sending/receiving of a buffer, the ISR is invoked and it goes through the linked-list to see how many buffers
 | |
| are freed after last interrupt, and send corresponding signals to the app.
 | |
| 
 | |
| The driver of FIFOs works as below:
 | |
| 
 | |
| 1.  The receive driver requires application to "register" a buffer before it's used. The driver
 | |
|     dynamically allocate a linked-list descriptor for the buffer, and return the descriptor as a handle
 | |
|     to the app.
 | |
| 
 | |
|     Each time the app asks to receive by a buffer, the descriptor of the buffer is loaded onto the linked-list,
 | |
|     and the counter of receiving buffers is increased so that the host will know this by the receiving interrupt.
 | |
|     The hardware will automatically go through the linked list and write data into the buffers loaded on the
 | |
|     list.
 | |
| 
 | |
|     The receiving driver sends a counting semaphore to the app for each buffer finished receiving. A task can only
 | |
|     check the linked list and fetch one finished buffer for a received semaphore.
 | |
| 
 | |
| 2.  The sending driver is slightly different due to different hardware working styles. (See README.md in the same folder
 | |
|     for the diagram how task and ISR work concurrently)
 | |
|     (TODO: re-write this part if the stitch mode is released)
 | |
|     The hardware has a cache, so that once a descriptor is loaded onto the linked-list, it cannot be modified
 | |
|     until returned (used) by the hardware. This forbids us from loading descriptors onto the linked list during
 | |
|     the transfer (or the time waiting for host to start a transfer). However, we use a "ringbuffer" (different from
 | |
|     the one in ``freertos/`` folder) holding descriptors to solve this:
 | |
| 
 | |
|     1.  The driver allocates continuous memory for several buffer descriptors (the maximum buffer number) during
 | |
|         initialization. Then the driver points the STAILQ_NEXT pointer of all the descriptors except the last one
 | |
|         to the next descriptor of each of them. Then the pointer of the last descriptor points back to the first one:
 | |
|         now the descriptor is in a ring.
 | |
| 
 | |
|     2.  The "ringbuffer" has a write pointer points to where app can write new descriptor. The app writes the new descriptor
 | |
|         indicated by the write pointer without touching the STAILQ_NEXT pointer so that the descriptors are always in a
 | |
|         ring-like linked-list. The app never touches the part of linked-list being used by the hardware.
 | |
| 
 | |
|     3.  When the hardware needs some data to send, it automatically pick a part of connected descriptors. According to the mode:
 | |
|             - Buffer mode: only pick the next one of the last sent one;
 | |
|             - Stream mode: pick the one above to the latest one.
 | |
| 
 | |
|         The driver removes the STAILQ_NEXT pointer of the last descriptor and put the head of the part to the DMA controller so
 | |
|         that it looks like just a linear linked-list rather than a ring to the hardware.
 | |
| 
 | |
|     4.  The counter of sending FIFO can increase when app load new buffers (in STREAM_MODE) or when new transfer should
 | |
|         start (in PACKET_MODE).
 | |
| 
 | |
|     5.  When the sending transfer is finished, the driver goes through the descriptors just send in the ISR and push all
 | |
|         the ``arg`` member of descriptors to the queue back to the app, so that the app can handle finished buffers. The
 | |
|         driver also fix the STAILQ_NEXT pointer of the last descriptor so that the descriptors are now in a ring again.
 | |
| */
 | |
| 
 | |
| 
 | |
| 
 | |
| #include <string.h>
 | |
| #include "driver/sdio_slave.h"
 | |
| #include "soc/sdio_slave_periph.h"
 | |
| #include "esp_log.h"
 | |
| #include "esp_intr_alloc.h"
 | |
| #include "freertos/FreeRTOS.h"
 | |
| #include "soc/soc_memory_layout.h"
 | |
| #include "soc/gpio_periph.h"
 | |
| #include "esp_cpu.h"
 | |
| #include "freertos/semphr.h"
 | |
| #include "esp_private/periph_ctrl.h"
 | |
| #include "driver/gpio.h"
 | |
| #include "hal/sdio_slave_hal.h"
 | |
| #include "hal/gpio_hal.h"
 | |
| 
 | |
| 
 | |
| #define SDIO_SLAVE_CHECK(res, str, ret_val) do { if(!(res)){\
 | |
|     SDIO_SLAVE_LOGE("%s", str);\
 | |
|     return ret_val;\
 | |
| } }while (0)
 | |
| 
 | |
| static const char TAG[] = "sdio_slave";
 | |
| 
 | |
| #define SDIO_SLAVE_LOGE(s, ...) ESP_LOGE(TAG, "%s(%d): "s, __FUNCTION__,__LINE__,##__VA_ARGS__)
 | |
| #define SDIO_SLAVE_LOGW(s, ...) ESP_LOGW(TAG, "%s: "s, __FUNCTION__,##__VA_ARGS__)
 | |
| 
 | |
| 
 | |
| // sdio_slave_buf_handle_t is of type recv_desc_t*;
 | |
| typedef struct recv_desc_s {
 | |
|     union {
 | |
|         struct {
 | |
|             // the third word, pointer to next desc, is shared with the tailq entry.
 | |
|             sdio_slave_hal_recv_desc_t hal_desc;
 | |
|             // when the forth word is used (not NULL), means the tailq is used, not in the receiving state.
 | |
|             uint32_t not_receiving;
 | |
|         };
 | |
|         struct {
 | |
|             // first 3 WORDs of this struct is defined by and compatible to the DMA link list format.
 | |
|             uint32_t _reserved0;
 | |
|             uint32_t _reserved1;
 | |
|             TAILQ_ENTRY(recv_desc_s) te; // tailq used to store the registered descriptors.
 | |
|         };
 | |
|     };
 | |
| } recv_desc_t;
 | |
| 
 | |
| 
 | |
| typedef TAILQ_HEAD(recv_tailq_head_s, recv_desc_s) recv_tailq_t;
 | |
| 
 | |
| typedef struct {
 | |
|     sdio_slave_config_t     config;
 | |
|     sdio_slave_context_t    *hal;
 | |
|     intr_handle_t           intr_handle; //allocated interrupt handle
 | |
|     /*------- events ---------------*/
 | |
|     union {
 | |
|         SemaphoreHandle_t       events[9]; // 0-7 for gp intr
 | |
|         struct {
 | |
|             SemaphoreHandle_t   _events[8];
 | |
|             SemaphoreHandle_t   recv_event; // 8 for recv
 | |
|         };
 | |
|     };
 | |
|     portMUX_TYPE            reg_spinlock;
 | |
|     /*------- sending ---------------*/
 | |
|     //desc in the send_link_list are temporary, taken information and space from the ringbuf, return to ringbuf after use.
 | |
|     SemaphoreHandle_t   remain_cnt;
 | |
|     portMUX_TYPE        write_spinlock;
 | |
|     QueueHandle_t       ret_queue;
 | |
|     /*------- receiving ---------------*/
 | |
|     recv_tailq_t        recv_reg_list;  // removed from the link list, registered but not used now
 | |
|     portMUX_TYPE        recv_spinlock;
 | |
| } sdio_context_t;
 | |
| 
 | |
| #define CONTEXT_INIT_VAL { \
 | |
|     .intr_handle = NULL, \
 | |
|     .hal = NULL, \
 | |
|     /*------- events ---------------*/ \
 | |
|     .events     =   {}, \
 | |
|     .reg_spinlock = portMUX_INITIALIZER_UNLOCKED, \
 | |
|     /*------- sending ---------------*/ \
 | |
|     .ret_queue      =   NULL, \
 | |
|     .write_spinlock =   portMUX_INITIALIZER_UNLOCKED, \
 | |
|     /*------- receiving ---------------*/ \
 | |
|     .recv_reg_list  =   TAILQ_HEAD_INITIALIZER(context.recv_reg_list), \
 | |
|     .recv_spinlock  =   portMUX_INITIALIZER_UNLOCKED, \
 | |
| }
 | |
| 
 | |
| static sdio_context_t context = CONTEXT_INIT_VAL;
 | |
| 
 | |
| static void sdio_intr(void *);
 | |
| static void sdio_intr_host(void *);
 | |
| static void sdio_intr_send(void *);
 | |
| static void sdio_intr_recv(void *);
 | |
| 
 | |
| static esp_err_t send_flush_data(void);
 | |
| static esp_err_t recv_flush_data(void);
 | |
| 
 | |
| static inline void critical_enter_recv(void);
 | |
| static inline void critical_exit_recv(void);
 | |
| 
 | |
| static void deinit_context(void);
 | |
| 
 | |
| static inline void show_ll(sdio_slave_ll_desc_t *item)
 | |
| {
 | |
|     ESP_EARLY_LOGI(TAG, "=> %p: size: %d(%d), eof: %d, owner: %d", item, item->size, item->length, item->eof, item->owner);
 | |
|     ESP_EARLY_LOGI(TAG, "   buf: %p, stqe_next: %p", item->buf, item->qe.stqe_next);
 | |
| }
 | |
| 
 | |
| static void __attribute((unused)) dump_ll(sdio_slave_ll_desc_t *queue)
 | |
| {
 | |
|     int cnt = 0;
 | |
|     sdio_slave_ll_desc_t *item = queue;
 | |
|     while (item != NULL) {
 | |
|         cnt++;
 | |
|         show_ll(item);
 | |
|         item = STAILQ_NEXT(item, qe);
 | |
|     }
 | |
|     ESP_EARLY_LOGI(TAG, "total: %d", cnt);
 | |
| }
 | |
| 
 | |
| static inline void deinit_context(void)
 | |
| {
 | |
|     context.config = (sdio_slave_config_t) {};
 | |
|     for (int i = 0; i < 9; i++) {
 | |
|         if (context.events[i] != NULL) {
 | |
|             vSemaphoreDelete(context.events[i]);
 | |
|             context.events[i] = NULL;
 | |
|         }
 | |
|     }
 | |
|     if (context.ret_queue != NULL) {
 | |
|         vQueueDelete(context.ret_queue);
 | |
|         context.ret_queue = NULL;
 | |
|     }
 | |
|     if (context.remain_cnt != NULL) {
 | |
|         vSemaphoreDelete(context.remain_cnt);
 | |
|     }
 | |
|     free(context.hal->send_desc_queue.data);
 | |
|     context.hal->send_desc_queue.data = NULL;
 | |
|     free(context.hal);
 | |
|     context.hal = NULL;
 | |
| }
 | |
| 
 | |
| static esp_err_t init_context(const sdio_slave_config_t *config)
 | |
| {
 | |
|     SDIO_SLAVE_CHECK(*(uint32_t *)&context.config == 0, "sdio slave already initialized", ESP_ERR_INVALID_STATE);
 | |
|     context = (sdio_context_t)CONTEXT_INIT_VAL;
 | |
|     context.config = *config;
 | |
| 
 | |
|     //initialize and configure the HAL
 | |
|     context.hal = (sdio_slave_context_t *)heap_caps_calloc(sizeof(sdio_slave_context_t), 1, MALLOC_CAP_INTERNAL | MALLOC_CAP_8BIT);
 | |
|     if (context.hal == NULL) {
 | |
|         goto no_mem;
 | |
|     }
 | |
| 
 | |
|     context.hal->sending_mode = config->sending_mode;
 | |
|     context.hal->timing = config->timing;
 | |
|     context.hal->no_highspeed = (config->flags & SDIO_SLAVE_FLAG_DEFAULT_SPEED) == SDIO_SLAVE_FLAG_DEFAULT_SPEED;
 | |
|     context.hal->send_queue_size = config->send_queue_size;
 | |
|     context.hal->recv_buffer_size = config->recv_buffer_size;
 | |
|     //initialize ringbuffer resources
 | |
|     sdio_ringbuf_t *buf = &(context.hal->send_desc_queue);
 | |
|     //one item is not used.
 | |
|     buf->size = SDIO_SLAVE_SEND_DESC_SIZE * (config->send_queue_size + 1);
 | |
|     buf->data = (uint8_t *)heap_caps_malloc(buf->size, MALLOC_CAP_INTERNAL | MALLOC_CAP_8BIT | MALLOC_CAP_DMA);
 | |
|     if (buf->data == NULL) {
 | |
|         goto no_mem;
 | |
|     }
 | |
| 
 | |
|     sdio_slave_hal_init(context.hal);
 | |
| 
 | |
|     // in theory we can queue infinite buffers in the linked list, but for multi-core reason we have to use a queue to
 | |
|     // count the finished buffers.
 | |
|     context.recv_event = xSemaphoreCreateCounting(UINT32_MAX, 0);
 | |
|     for (int i = 0; i < 9; i++) {
 | |
|         if (i < 8) {
 | |
|             context.events[i] = xSemaphoreCreateBinary();
 | |
|         }   //for 8, already created.
 | |
|         if (context.events[i] == NULL) {
 | |
|             SDIO_SLAVE_LOGE("event initialize failed");
 | |
|             goto no_mem;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     context.remain_cnt = xSemaphoreCreateCounting(context.config.send_queue_size, context.config.send_queue_size);
 | |
|     if (context.remain_cnt == NULL) {
 | |
|         goto no_mem;
 | |
|     }
 | |
| 
 | |
|     context.ret_queue = xQueueCreate(config->send_queue_size, sizeof(void *));
 | |
|     if (context.ret_queue == NULL) {
 | |
|         goto no_mem;
 | |
|     }
 | |
| 
 | |
|     return ESP_OK;
 | |
| 
 | |
| no_mem:
 | |
|     deinit_context();
 | |
|     return ESP_ERR_NO_MEM;
 | |
| }
 | |
| 
 | |
| static void configure_pin(int pin, uint32_t func, bool pullup)
 | |
| {
 | |
|     const int sdmmc_func = func;
 | |
|     const int drive_strength = 3;
 | |
|     assert(pin != -1);
 | |
|     uint32_t reg = GPIO_PIN_MUX_REG[pin];
 | |
|     assert(reg != UINT32_MAX);
 | |
| 
 | |
|     PIN_INPUT_ENABLE(reg);
 | |
|     gpio_hal_iomux_func_sel(reg, sdmmc_func);
 | |
|     PIN_SET_DRV(reg, drive_strength);
 | |
|     gpio_pulldown_dis(pin);
 | |
|     if (pullup) {
 | |
|         gpio_pullup_en(pin);
 | |
|     }
 | |
| }
 | |
| 
 | |
| static inline esp_err_t sdio_slave_hw_init(sdio_slave_config_t *config)
 | |
| {
 | |
|     //initialize pin
 | |
|     const sdio_slave_slot_info_t *slot = &sdio_slave_slot_info[0];
 | |
| 
 | |
|     bool pullup = config->flags & SDIO_SLAVE_FLAG_INTERNAL_PULLUP;
 | |
|     configure_pin(slot->clk_gpio, slot->func, false);   //clk doesn't need a pullup
 | |
|     configure_pin(slot->cmd_gpio, slot->func, pullup);
 | |
|     configure_pin(slot->d0_gpio, slot->func, pullup);
 | |
|     if ((config->flags & SDIO_SLAVE_FLAG_HOST_INTR_DISABLED) == 0) {
 | |
|         configure_pin(slot->d1_gpio, slot->func, pullup);
 | |
|     }
 | |
|     if ((config->flags & SDIO_SLAVE_FLAG_DAT2_DISABLED) == 0) {
 | |
|         configure_pin(slot->d2_gpio, slot->func, pullup);
 | |
|     }
 | |
|     configure_pin(slot->d3_gpio, slot->func, pullup);
 | |
| 
 | |
|     //enable module and config
 | |
|     periph_module_reset(PERIPH_SDIO_SLAVE_MODULE);
 | |
|     periph_module_enable(PERIPH_SDIO_SLAVE_MODULE);
 | |
| 
 | |
|     sdio_slave_hal_hw_init(context.hal);
 | |
|     return ESP_OK;
 | |
| }
 | |
| 
 | |
| static void recover_pin(int pin, int sdio_func)
 | |
| {
 | |
|     uint32_t reg = GPIO_PIN_MUX_REG[pin];
 | |
|     assert(reg != UINT32_MAX);
 | |
| 
 | |
|     int func = REG_GET_FIELD(reg, MCU_SEL);
 | |
|     if (func == sdio_func) {
 | |
|         gpio_set_direction(pin, GPIO_MODE_INPUT);
 | |
|         gpio_hal_iomux_func_sel(reg, PIN_FUNC_GPIO);
 | |
|     }
 | |
| }
 | |
| 
 | |
| static void sdio_slave_hw_deinit(void)
 | |
| {
 | |
|     const sdio_slave_slot_info_t *slot = &sdio_slave_slot_info[0];
 | |
|     recover_pin(slot->clk_gpio, slot->func);
 | |
|     recover_pin(slot->cmd_gpio, slot->func);
 | |
|     recover_pin(slot->d0_gpio, slot->func);
 | |
|     recover_pin(slot->d1_gpio, slot->func);
 | |
|     recover_pin(slot->d2_gpio, slot->func);
 | |
|     recover_pin(slot->d3_gpio, slot->func);
 | |
| }
 | |
| 
 | |
| esp_err_t sdio_slave_initialize(sdio_slave_config_t *config)
 | |
| {
 | |
|     esp_err_t r;
 | |
|     intr_handle_t intr_handle = NULL;
 | |
|     const int flags = 0;
 | |
|     r = esp_intr_alloc(ETS_SLC0_INTR_SOURCE, flags, sdio_intr, NULL, &intr_handle);
 | |
|     if (r != ESP_OK) {
 | |
|         return r;
 | |
|     }
 | |
| 
 | |
|     r = init_context(config);
 | |
|     if (r != ESP_OK) {
 | |
|         return r;
 | |
|     }
 | |
|     context.intr_handle = intr_handle;
 | |
| 
 | |
|     r = sdio_slave_hw_init(config);
 | |
|     if (r != ESP_OK) {
 | |
|         return r;
 | |
|     }
 | |
| 
 | |
|     sdio_slave_reset();
 | |
|     return ESP_OK;
 | |
| }
 | |
| 
 | |
| void sdio_slave_deinit(void)
 | |
| {
 | |
|     sdio_slave_hw_deinit();
 | |
| 
 | |
|     //unregister all buffers registered but returned (not loaded)
 | |
|     recv_desc_t *temp_desc;
 | |
|     recv_desc_t *desc;
 | |
|     TAILQ_FOREACH_SAFE(desc, &context.recv_reg_list, te, temp_desc) {
 | |
|         TAILQ_REMOVE(&context.recv_reg_list, desc, te);
 | |
|         free(desc);
 | |
|     }
 | |
|     //unregister all buffers that is loaded and not returned
 | |
|     while (1) {
 | |
|         desc = (recv_desc_t *)sdio_slave_hal_recv_unload_desc(context.hal);
 | |
|         if (desc == NULL) {
 | |
|             break;
 | |
|         }
 | |
|         free(desc);
 | |
|     }
 | |
|     esp_err_t ret = esp_intr_free(context.intr_handle);
 | |
|     assert(ret == ESP_OK);
 | |
|     (void)ret;
 | |
|     context.intr_handle = NULL;
 | |
|     deinit_context();
 | |
| }
 | |
| 
 | |
| esp_err_t sdio_slave_start(void)
 | |
| {
 | |
|     esp_err_t ret;
 | |
|     sdio_slave_hostint_t intr = (sdio_slave_hostint_t)UINT32_MAX;
 | |
|     sdio_slave_hal_hostint_clear(context.hal, &intr);
 | |
|     ret = sdio_slave_hal_send_start(context.hal);
 | |
|     if (ret != ESP_OK) {
 | |
|         return ret;
 | |
|     }
 | |
| 
 | |
|     critical_enter_recv();
 | |
|     sdio_slave_hal_recv_start(context.hal);
 | |
|     critical_exit_recv();
 | |
| 
 | |
|     sdio_slave_hal_set_ioready(context.hal, true);
 | |
|     return ESP_OK;
 | |
| }
 | |
| 
 | |
| esp_err_t sdio_slave_reset(void)
 | |
| {
 | |
|     esp_err_t err;
 | |
|     err = send_flush_data();
 | |
|     if (err != ESP_OK) {
 | |
|         return err;
 | |
|     }
 | |
| 
 | |
|     err = sdio_slave_hal_send_reset_counter(context.hal);
 | |
|     if (err != ESP_OK) {
 | |
|         return err;
 | |
|     }
 | |
| 
 | |
|     err = recv_flush_data();
 | |
|     if (err != ESP_OK) {
 | |
|         return err;
 | |
|     }
 | |
| 
 | |
|     critical_enter_recv();
 | |
|     sdio_slave_hal_recv_reset_counter(context.hal);
 | |
|     critical_exit_recv();
 | |
|     err = ESP_OK;
 | |
|     return err;
 | |
| }
 | |
| 
 | |
| void sdio_slave_stop(void)
 | |
| {
 | |
|     sdio_slave_hal_set_ioready(context.hal, false);
 | |
|     sdio_slave_hal_send_stop(context.hal);
 | |
|     sdio_slave_hal_recv_stop(context.hal);
 | |
| }
 | |
| 
 | |
| static void sdio_intr(void *arg)
 | |
| {
 | |
|     sdio_intr_send(arg);
 | |
|     sdio_intr_recv(arg);
 | |
|     sdio_intr_host(arg);
 | |
| }
 | |
| 
 | |
| /*---------------------------------------------------------------------------
 | |
|  *                  Host
 | |
|  *--------------------------------------------------------------------------*/
 | |
| static void sdio_intr_host(void *arg)
 | |
| {
 | |
|     sdio_slave_ll_slvint_t int_val;
 | |
|     sdio_slave_hal_slvint_fetch_clear(context.hal, &int_val);
 | |
|     BaseType_t yield = pdFALSE;
 | |
|     for (int i = 0; i < 8; i++) {
 | |
|         if (BIT(i) & int_val) {
 | |
|             if (context.config.event_cb != NULL) {
 | |
|                 (*context.config.event_cb)(i);
 | |
|             }
 | |
|             xSemaphoreGiveFromISR(context.events[i], &yield);
 | |
|         }
 | |
|     }
 | |
|     if (yield) {
 | |
|         portYIELD_FROM_ISR();
 | |
|     }
 | |
| }
 | |
| 
 | |
| esp_err_t sdio_slave_wait_int(int pos, TickType_t wait)
 | |
| {
 | |
|     SDIO_SLAVE_CHECK(pos >= 0 && pos < 8, "interrupt num invalid", ESP_ERR_INVALID_ARG);
 | |
|     return xSemaphoreTake(context.events[pos], wait);
 | |
| }
 | |
| 
 | |
| uint8_t sdio_slave_read_reg(int pos)
 | |
| {
 | |
|     if (pos >= 28 && pos <= 31) {
 | |
|         SDIO_SLAVE_LOGW("%s: interrupt reg, for reference", __FUNCTION__);
 | |
|     }
 | |
|     if (pos < 0 || pos >= 64) {
 | |
|         SDIO_SLAVE_LOGE("read register address wrong");
 | |
|     }
 | |
|     return sdio_slave_hal_host_get_reg(context.hal, pos);
 | |
| }
 | |
| 
 | |
| esp_err_t sdio_slave_write_reg(int pos, uint8_t reg)
 | |
| {
 | |
|     if (pos >= 28 && pos <= 31) {
 | |
|         SDIO_SLAVE_LOGE("interrupt reg, please use sdio_slave_clear_int");
 | |
|         return ESP_ERR_INVALID_ARG;
 | |
|     }
 | |
|     if (pos < 0 || pos >= 64) {
 | |
|         SDIO_SLAVE_LOGE("write register address wrong");
 | |
|         return ESP_ERR_INVALID_ARG;
 | |
|     }
 | |
| 
 | |
|     portENTER_CRITICAL(&context.reg_spinlock);
 | |
|     sdio_slave_hal_host_set_reg(context.hal, pos, reg);
 | |
|     portEXIT_CRITICAL(&context.reg_spinlock);
 | |
|     return ESP_OK;
 | |
| }
 | |
| 
 | |
| sdio_slave_hostint_t sdio_slave_get_host_intena(void)
 | |
| {
 | |
|     sdio_slave_hostint_t host_int;
 | |
|     sdio_slave_hal_hostint_get_ena(context.hal, &host_int);
 | |
|     return host_int;
 | |
| }
 | |
| 
 | |
| void sdio_slave_set_host_intena(sdio_slave_hostint_t mask)
 | |
| {
 | |
|     sdio_slave_hal_hostint_set_ena(context.hal, &mask);
 | |
| }
 | |
| 
 | |
| void sdio_slave_clear_host_int(sdio_slave_hostint_t mask)
 | |
| {
 | |
|     sdio_slave_hal_hostint_clear(context.hal, &mask);
 | |
| }
 | |
| 
 | |
| static inline sdio_slave_hostint_t get_hostint_by_pos(int pos)
 | |
| {
 | |
|     return (sdio_slave_hostint_t)BIT(pos);
 | |
| }
 | |
| 
 | |
| esp_err_t sdio_slave_send_host_int(uint8_t pos)
 | |
| {
 | |
|     SDIO_SLAVE_CHECK(pos < 8, "interrupt num invalid", ESP_ERR_INVALID_ARG);
 | |
|     sdio_slave_hostint_t intr = get_hostint_by_pos(pos);
 | |
|     sdio_slave_hal_hostint_send(context.hal, &intr);
 | |
|     return ESP_OK;
 | |
| }
 | |
| 
 | |
| /*---------------------------------------------------------------------------
 | |
|  *                  Send
 | |
|  *--------------------------------------------------------------------------*/
 | |
| 
 | |
| /* The link list is handled in the app, while counter and pointer processed in ISR.
 | |
|  * Driver abuse rx_done bit to invoke ISR.
 | |
|  * If driver is stopped, the link list is stopped as well as the ISR invoker.
 | |
|  */
 | |
| 
 | |
| static void sdio_intr_send(void *arg)
 | |
| {
 | |
|     ESP_EARLY_LOGV(TAG, "intr_send");
 | |
|     BaseType_t yield = pdFALSE;
 | |
| 
 | |
|     // this interrupt is abused to get ISR invoked by app
 | |
|     sdio_slave_hal_send_handle_isr_invoke(context.hal);
 | |
| 
 | |
|     uint32_t returned_cnt;
 | |
|     if (sdio_slave_hal_send_eof_happened(context.hal)) {
 | |
|         BaseType_t ret __attribute__((unused));
 | |
| 
 | |
|         esp_err_t err;
 | |
|         while (1) {
 | |
|             void *finished_arg;
 | |
|             err = sdio_slave_hal_send_get_next_finished_arg(context.hal, &finished_arg, &returned_cnt);
 | |
|             if (err != ESP_OK) {
 | |
|                 break;
 | |
|             }
 | |
| 
 | |
|             assert(returned_cnt == 0);
 | |
|             ESP_EARLY_LOGV(TAG, "end: %x", finished_arg);
 | |
|             ret = xQueueSendFromISR(context.ret_queue, &finished_arg, &yield);
 | |
|             assert(ret == pdTRUE);
 | |
|         }
 | |
|         //get_next_finished_arg returns the total amount of returned descs.
 | |
|         for (size_t i = 0; i < returned_cnt; i++) {
 | |
|             ret = xSemaphoreGiveFromISR(context.remain_cnt, &yield);
 | |
|             assert(ret == pdTRUE);
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     sdio_slave_hal_send_new_packet_if_exist(context.hal);
 | |
| 
 | |
|     if (yield) {
 | |
|         portYIELD_FROM_ISR();
 | |
|     }
 | |
| }
 | |
| 
 | |
| esp_err_t sdio_slave_send_queue(uint8_t *addr, size_t len, void *arg, TickType_t wait)
 | |
| {
 | |
|     SDIO_SLAVE_CHECK(len > 0, "len <= 0", ESP_ERR_INVALID_ARG);
 | |
|     SDIO_SLAVE_CHECK(esp_ptr_dma_capable(addr) && (uint32_t)addr % 4 == 0, "buffer to send should be DMA capable and 32-bit aligned",
 | |
|                      ESP_ERR_INVALID_ARG);
 | |
| 
 | |
|     BaseType_t cnt_ret = xSemaphoreTake(context.remain_cnt, wait);
 | |
|     if (cnt_ret != pdTRUE) {
 | |
|         return ESP_ERR_TIMEOUT;
 | |
|     }
 | |
| 
 | |
|     portENTER_CRITICAL(&context.write_spinlock);
 | |
|     esp_err_t ret = sdio_slave_hal_send_queue(context.hal, addr, len, arg);
 | |
|     portEXIT_CRITICAL(&context.write_spinlock);
 | |
|     if (ret != ESP_OK) {
 | |
|         return ret;
 | |
|     }
 | |
| 
 | |
|     return ESP_OK;
 | |
| }
 | |
| 
 | |
| esp_err_t sdio_slave_send_get_finished(void **out_arg, TickType_t wait)
 | |
| {
 | |
|     void *arg = NULL;
 | |
|     BaseType_t err = xQueueReceive(context.ret_queue, &arg, wait);
 | |
|     if (out_arg) {
 | |
|         *out_arg = arg;
 | |
|     }
 | |
|     if (err != pdTRUE) {
 | |
|         return ESP_ERR_TIMEOUT;
 | |
|     }
 | |
|     return ESP_OK;
 | |
| }
 | |
| 
 | |
| esp_err_t sdio_slave_transmit(uint8_t *addr, size_t len)
 | |
| {
 | |
|     uint32_t timestamp = esp_cpu_get_cycle_count();
 | |
|     uint32_t ret_stamp;
 | |
| 
 | |
|     esp_err_t err = sdio_slave_send_queue(addr, len, (void *)timestamp, portMAX_DELAY);
 | |
|     if (err != ESP_OK) {
 | |
|         return err;
 | |
|     }
 | |
|     err = sdio_slave_send_get_finished((void **)&ret_stamp, portMAX_DELAY);
 | |
|     if (err != ESP_OK) {
 | |
|         return err;
 | |
|     }
 | |
|     SDIO_SLAVE_CHECK(ret_stamp == timestamp, "already sent without return before", ESP_ERR_INVALID_STATE);
 | |
| 
 | |
|     return ESP_OK;
 | |
| }
 | |
| 
 | |
| //clear data but keep counter
 | |
| static esp_err_t send_flush_data(void)
 | |
| {
 | |
|     esp_err_t err;
 | |
|     BaseType_t ret __attribute__((unused));
 | |
| 
 | |
|     while (1) {
 | |
|         void *finished_arg;
 | |
|         uint32_t return_cnt = 0;
 | |
|         err = sdio_slave_hal_send_flush_next_buffer(context.hal, &finished_arg, &return_cnt);
 | |
|         if (err == ESP_OK) {
 | |
|             ret = xQueueSend(context.ret_queue, &finished_arg, portMAX_DELAY);
 | |
|             assert(ret == pdTRUE);
 | |
|             for (size_t i = 0; i < return_cnt; i++) {
 | |
|                 ret = xSemaphoreGive(context.remain_cnt);
 | |
|                 assert(ret == pdTRUE);
 | |
|             }
 | |
|         } else {
 | |
|             if (err == ESP_ERR_NOT_FOUND) {
 | |
|                 err = ESP_OK;
 | |
|             }
 | |
|             break;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     if (err == ESP_ERR_INVALID_STATE) {
 | |
|         ESP_LOGE(TAG, "flush data when transmission started");
 | |
|     }
 | |
|     return err;
 | |
| }
 | |
| 
 | |
| /*---------------------------------------------------------------------------
 | |
|  *                  Recv
 | |
|  *--------------------------------------------------------------------------*/
 | |
| #define CHECK_HANDLE_IDLE(desc) do { if (desc == NULL || !desc->not_receiving) {\
 | |
|     return ESP_ERR_INVALID_ARG; } } while(0)
 | |
| 
 | |
| static inline void critical_enter_recv(void)
 | |
| {
 | |
|     portENTER_CRITICAL(&context.recv_spinlock);
 | |
| }
 | |
| 
 | |
| static inline void critical_exit_recv(void)
 | |
| {
 | |
|     portEXIT_CRITICAL(&context.recv_spinlock);
 | |
| }
 | |
| 
 | |
| // remove data, still increase the counter
 | |
| static esp_err_t recv_flush_data(void)
 | |
| {
 | |
|     while (1) {
 | |
|         BaseType_t ret = xSemaphoreTake(context.recv_event, 0);
 | |
|         if (ret == pdFALSE) {
 | |
|             break;
 | |
|         }
 | |
|         critical_enter_recv();
 | |
|         sdio_slave_hal_recv_flush_one_buffer(context.hal);
 | |
|         critical_exit_recv();
 | |
|     }
 | |
|     return ESP_OK;
 | |
| }
 | |
| 
 | |
| static void sdio_intr_recv(void *arg)
 | |
| {
 | |
|     BaseType_t yield = 0;
 | |
|     bool triggered = sdio_slave_hal_recv_done(context.hal);
 | |
|     while (triggered) {
 | |
|         portENTER_CRITICAL_ISR(&context.recv_spinlock);
 | |
|         bool has_next_item = sdio_slave_hal_recv_has_next_item(context.hal);
 | |
|         portEXIT_CRITICAL_ISR(&context.recv_spinlock);
 | |
|         if (has_next_item) {
 | |
|             ESP_EARLY_LOGV(TAG, "intr_recv: Give");
 | |
|             xSemaphoreGiveFromISR(context.recv_event, &yield);
 | |
|             continue;   //check the linked list again skip the interrupt checking
 | |
|         }
 | |
|         // if no more items on the list, check the interrupt again,
 | |
|         // will loop until the interrupt bit is kept cleared.
 | |
|         triggered = sdio_slave_hal_recv_done(context.hal);
 | |
|     }
 | |
|     if (yield) {
 | |
|         portYIELD_FROM_ISR();
 | |
|     }
 | |
| }
 | |
| 
 | |
| esp_err_t sdio_slave_recv_load_buf(sdio_slave_buf_handle_t handle)
 | |
| {
 | |
|     recv_desc_t *desc = (recv_desc_t *)handle;
 | |
|     CHECK_HANDLE_IDLE(desc);
 | |
|     assert(desc->not_receiving);
 | |
| 
 | |
|     critical_enter_recv();
 | |
|     TAILQ_REMOVE(&context.recv_reg_list, desc, te);
 | |
|     desc->not_receiving = 0; //manually remove the prev link (by set not_receiving=0), to indicate this is in the queue
 | |
|     sdio_slave_hal_load_buf(context.hal, &desc->hal_desc);
 | |
|     critical_exit_recv();
 | |
|     return ESP_OK;
 | |
| }
 | |
| 
 | |
| sdio_slave_buf_handle_t sdio_slave_recv_register_buf(uint8_t *start)
 | |
| {
 | |
|     SDIO_SLAVE_CHECK(esp_ptr_dma_capable(start) && (uint32_t)start % 4 == 0,
 | |
|                      "buffer to register should be DMA capable and 32-bit aligned", NULL);
 | |
|     recv_desc_t *desc = (recv_desc_t *)heap_caps_malloc(sizeof(recv_desc_t), MALLOC_CAP_INTERNAL | MALLOC_CAP_8BIT | MALLOC_CAP_DMA);
 | |
|     if (desc == NULL) {
 | |
|         SDIO_SLAVE_LOGE("cannot allocate lldesc for new buffer");
 | |
|         return NULL;
 | |
|     }
 | |
| 
 | |
|     //initially in the reg list
 | |
|     sdio_slave_hal_recv_init_desc(context.hal, &desc->hal_desc, start);
 | |
|     critical_enter_recv();
 | |
|     TAILQ_INSERT_TAIL(&context.recv_reg_list, desc, te);
 | |
|     critical_exit_recv();
 | |
|     return desc;
 | |
| }
 | |
| 
 | |
| esp_err_t sdio_slave_recv(sdio_slave_buf_handle_t *handle_ret, uint8_t **out_addr, size_t *out_len, TickType_t wait)
 | |
| {
 | |
|     esp_err_t ret = sdio_slave_recv_packet(handle_ret, wait);
 | |
|     if (ret == ESP_ERR_NOT_FINISHED) {
 | |
|         //This API was not awared of the EOF info, return ESP_OK to keep back-compatible.
 | |
|         ret = ESP_OK;
 | |
|     }
 | |
|     if (ret == ESP_OK) {
 | |
|         recv_desc_t *desc = (recv_desc_t *)(*handle_ret);
 | |
|         if (out_addr) {
 | |
|             *out_addr = (uint8_t *)desc->hal_desc.buf;
 | |
|         }
 | |
|         if (out_len) {
 | |
|             *out_len = desc->hal_desc.length;
 | |
|         }
 | |
|     }
 | |
|     return ret;
 | |
| }
 | |
| 
 | |
| esp_err_t sdio_slave_recv_packet(sdio_slave_buf_handle_t *handle_ret, TickType_t wait)
 | |
| {
 | |
|     SDIO_SLAVE_CHECK(handle_ret != NULL, "handle address cannot be 0", ESP_ERR_INVALID_ARG);
 | |
|     BaseType_t err = xSemaphoreTake(context.recv_event, wait);
 | |
|     if (err == pdFALSE) {
 | |
|         return ESP_ERR_TIMEOUT;
 | |
|     }
 | |
| 
 | |
|     esp_err_t ret = ESP_OK;
 | |
|     critical_enter_recv();
 | |
|     //remove from queue, add back to reg list.
 | |
|     recv_desc_t *desc = (recv_desc_t *)sdio_slave_hal_recv_unload_desc(context.hal);
 | |
|     assert(desc != NULL && desc->hal_desc.owner == 0);
 | |
|     TAILQ_INSERT_TAIL(&context.recv_reg_list, desc, te);
 | |
|     critical_exit_recv();
 | |
| 
 | |
|     *handle_ret = (sdio_slave_buf_handle_t)desc;
 | |
| 
 | |
|     if (!desc->hal_desc.eof) {
 | |
|         ret = ESP_ERR_NOT_FINISHED;
 | |
|     }
 | |
|     return ret;
 | |
| }
 | |
| 
 | |
| esp_err_t sdio_slave_recv_unregister_buf(sdio_slave_buf_handle_t handle)
 | |
| {
 | |
|     recv_desc_t *desc = (recv_desc_t *)handle;
 | |
|     CHECK_HANDLE_IDLE(desc); //in the queue, fail.
 | |
| 
 | |
|     critical_enter_recv();
 | |
|     TAILQ_REMOVE(&context.recv_reg_list, desc, te);
 | |
|     critical_exit_recv();
 | |
|     free(desc);
 | |
|     return ESP_OK;
 | |
| }
 | |
| 
 | |
| uint8_t *sdio_slave_recv_get_buf(sdio_slave_buf_handle_t handle, size_t *len_o)
 | |
| {
 | |
|     if (handle == NULL) {
 | |
|         return NULL;
 | |
|     }
 | |
|     recv_desc_t *desc = (recv_desc_t *)handle;
 | |
| 
 | |
|     if (len_o != NULL) {
 | |
|         *len_o = desc->hal_desc.length;
 | |
|     }
 | |
|     return (uint8_t *)desc->hal_desc.buf;
 | |
| }
 |