389 lines
8.2 KiB
C
389 lines
8.2 KiB
C
/** @file
|
|
* @brief Modem socket / packet size handler
|
|
*
|
|
* Generic modem socket and packet size implementation for modem context
|
|
*/
|
|
|
|
/*
|
|
* Copyright (c) 2019-2020 Foundries.io
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*/
|
|
|
|
#include <kernel.h>
|
|
#include <sys/fdtable.h>
|
|
|
|
#include "modem_socket.h"
|
|
|
|
/*
|
|
* Packet Size Support Functions
|
|
*/
|
|
|
|
uint16_t modem_socket_next_packet_size(struct modem_socket_config *cfg,
|
|
struct modem_socket *sock)
|
|
{
|
|
uint16_t total = 0U;
|
|
|
|
k_sem_take(&cfg->sem_lock, K_FOREVER);
|
|
|
|
if (!sock || !sock->packet_count) {
|
|
goto exit;
|
|
}
|
|
|
|
total = sock->packet_sizes[0];
|
|
|
|
exit:
|
|
k_sem_give(&cfg->sem_lock);
|
|
return total;
|
|
}
|
|
|
|
static uint16_t modem_socket_packet_get_total(struct modem_socket *sock)
|
|
{
|
|
int i;
|
|
uint16_t total = 0U;
|
|
|
|
if (!sock || !sock->packet_count) {
|
|
return 0U;
|
|
}
|
|
|
|
for (i = 0; i < sock->packet_count; i++) {
|
|
total += sock->packet_sizes[i];
|
|
}
|
|
|
|
return total;
|
|
}
|
|
|
|
static int modem_socket_packet_drop_first(struct modem_socket *sock)
|
|
{
|
|
int i;
|
|
|
|
if (!sock || !sock->packet_count) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
sock->packet_count--;
|
|
for (i = 0; i < sock->packet_count; i++) {
|
|
sock->packet_sizes[i] =
|
|
sock->packet_sizes[i + 1];
|
|
}
|
|
|
|
sock->packet_sizes[sock->packet_count] = 0U;
|
|
return 0;
|
|
}
|
|
|
|
int modem_socket_packet_size_update(struct modem_socket_config *cfg,
|
|
struct modem_socket *sock, int new_total)
|
|
{
|
|
uint16_t old_total = 0U;
|
|
|
|
if (!sock) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
k_sem_take(&cfg->sem_lock, K_FOREVER);
|
|
|
|
if (new_total < 0) {
|
|
new_total += modem_socket_packet_get_total(sock);
|
|
}
|
|
|
|
if (new_total <= 0) {
|
|
/* reset outstanding value here */
|
|
sock->packet_count = 0U;
|
|
sock->packet_sizes[0] = 0U;
|
|
k_sem_give(&cfg->sem_lock);
|
|
return 0;
|
|
}
|
|
|
|
old_total = modem_socket_packet_get_total(sock);
|
|
if (new_total == old_total) {
|
|
goto data_ready;
|
|
}
|
|
|
|
/* remove sent packets */
|
|
if (new_total < old_total) {
|
|
/* remove packets that are not included in new_size */
|
|
while (old_total > new_total && sock->packet_count > 0) {
|
|
/* handle partial read */
|
|
if (old_total - new_total < sock->packet_sizes[0]) {
|
|
sock->packet_sizes[0] -= old_total - new_total;
|
|
break;
|
|
}
|
|
|
|
old_total -= sock->packet_sizes[0];
|
|
modem_socket_packet_drop_first(sock);
|
|
}
|
|
|
|
goto data_ready;
|
|
}
|
|
|
|
/* new packet to add */
|
|
if (sock->packet_count >= CONFIG_MODEM_SOCKET_PACKET_COUNT) {
|
|
k_sem_give(&cfg->sem_lock);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
if (new_total - old_total > 0) {
|
|
sock->packet_sizes[sock->packet_count] = new_total - old_total;
|
|
sock->packet_count++;
|
|
} else {
|
|
k_sem_give(&cfg->sem_lock);
|
|
return -EINVAL;
|
|
}
|
|
|
|
data_ready:
|
|
k_sem_give(&cfg->sem_lock);
|
|
return new_total;
|
|
}
|
|
|
|
/*
|
|
* Socket Support Functions
|
|
*/
|
|
|
|
int modem_socket_get(struct modem_socket_config *cfg,
|
|
int family, int type, int proto)
|
|
{
|
|
int i;
|
|
|
|
k_sem_take(&cfg->sem_lock, K_FOREVER);
|
|
|
|
for (i = 0; i < cfg->sockets_len; i++) {
|
|
if (cfg->sockets[i].id < cfg->base_socket_num) {
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (i >= cfg->sockets_len) {
|
|
k_sem_give(&cfg->sem_lock);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* FIXME: 4 fds max now due to POSIX_OS conflict */
|
|
cfg->sockets[i].sock_fd = z_reserve_fd();
|
|
if (cfg->sockets[i].sock_fd < 0) {
|
|
k_sem_give(&cfg->sem_lock);
|
|
return -errno;
|
|
}
|
|
|
|
cfg->sockets[i].family = family;
|
|
cfg->sockets[i].type = type;
|
|
cfg->sockets[i].ip_proto = proto;
|
|
/* socket # needs assigning */
|
|
cfg->sockets[i].id = cfg->sockets_len + 1;
|
|
z_finalize_fd(cfg->sockets[i].sock_fd, &cfg->sockets[i],
|
|
(const struct fd_op_vtable *)cfg->vtable);
|
|
|
|
k_sem_give(&cfg->sem_lock);
|
|
return cfg->sockets[i].sock_fd;
|
|
}
|
|
|
|
struct modem_socket *modem_socket_from_fd(struct modem_socket_config *cfg,
|
|
int sock_fd)
|
|
{
|
|
int i;
|
|
|
|
k_sem_take(&cfg->sem_lock, K_FOREVER);
|
|
|
|
for (i = 0; i < cfg->sockets_len; i++) {
|
|
if (cfg->sockets[i].sock_fd == sock_fd) {
|
|
k_sem_give(&cfg->sem_lock);
|
|
return &cfg->sockets[i];
|
|
}
|
|
}
|
|
|
|
k_sem_give(&cfg->sem_lock);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
struct modem_socket *modem_socket_from_id(struct modem_socket_config *cfg,
|
|
int id)
|
|
{
|
|
int i;
|
|
|
|
if (id < cfg->base_socket_num) {
|
|
return NULL;
|
|
}
|
|
|
|
k_sem_take(&cfg->sem_lock, K_FOREVER);
|
|
|
|
for (i = 0; i < cfg->sockets_len; i++) {
|
|
if (cfg->sockets[i].id == id) {
|
|
k_sem_give(&cfg->sem_lock);
|
|
return &cfg->sockets[i];
|
|
}
|
|
}
|
|
|
|
k_sem_give(&cfg->sem_lock);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
struct modem_socket *modem_socket_from_newid(struct modem_socket_config *cfg)
|
|
{
|
|
return modem_socket_from_id(cfg, cfg->sockets_len + 1);
|
|
}
|
|
|
|
void modem_socket_put(struct modem_socket_config *cfg, int sock_fd)
|
|
{
|
|
struct modem_socket *sock = modem_socket_from_fd(cfg, sock_fd);
|
|
|
|
if (!sock) {
|
|
return;
|
|
}
|
|
|
|
k_sem_take(&cfg->sem_lock, K_FOREVER);
|
|
|
|
sock->id = cfg->base_socket_num - 1;
|
|
sock->sock_fd = -1;
|
|
sock->is_waiting = false;
|
|
sock->is_polled = false;
|
|
sock->is_connected = false;
|
|
(void)memset(&sock->src, 0, sizeof(struct sockaddr));
|
|
(void)memset(&sock->dst, 0, sizeof(struct sockaddr));
|
|
memset(&sock->packet_sizes, 0, sizeof(sock->packet_sizes));
|
|
sock->packet_count = 0;
|
|
k_sem_reset(&sock->sem_data_ready);
|
|
|
|
k_sem_give(&cfg->sem_lock);
|
|
}
|
|
|
|
/*
|
|
* Generic Poll Function
|
|
*/
|
|
|
|
/*
|
|
* FIXME: The design here makes the poll function non-reentrant. If two
|
|
* different threads poll on two different sockets we'll end up with unexpected
|
|
* behavior - the higher priority thread will be unblocked, regardless on which
|
|
* socket it polled. I think we could live with such limitation though in the
|
|
* initial implementation, but this should be improved in the future.
|
|
*/
|
|
int modem_socket_poll(struct modem_socket_config *cfg,
|
|
struct zsock_pollfd *fds, int nfds, int msecs)
|
|
{
|
|
struct modem_socket *sock;
|
|
int ret, i;
|
|
uint8_t found_count = 0;
|
|
|
|
if (!cfg) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
k_sem_reset(&cfg->sem_poll);
|
|
|
|
for (i = 0; i < nfds; i++) {
|
|
sock = modem_socket_from_fd(cfg, fds[i].fd);
|
|
if (sock) {
|
|
/*
|
|
* Handle user check for POLLOUT events:
|
|
* we consider the socket to always be writable.
|
|
*/
|
|
if (fds[i].events & ZSOCK_POLLOUT) {
|
|
found_count++;
|
|
break;
|
|
} else if (fds[i].events & ZSOCK_POLLIN) {
|
|
sock->is_polled = true;
|
|
|
|
/*
|
|
* Handle check done after data reception on
|
|
* the socket. In this case that was received
|
|
* but as the socket wasn't polled, no sem_poll
|
|
* semaphore was given at that time. Therefore
|
|
* if there is a polled socket with data,
|
|
* increment found_count to escape the
|
|
* k_sem_take().
|
|
*/
|
|
if (sock->packet_sizes[0] > 0U) {
|
|
found_count++;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Avoid waiting on semaphore if we have already found an event */
|
|
ret = 0;
|
|
if (!found_count) {
|
|
ret = k_sem_take(&cfg->sem_poll, K_MSEC(msecs));
|
|
}
|
|
/* Reset counter as we reiterate on all polled sockets */
|
|
found_count = 0;
|
|
|
|
for (i = 0; i < nfds; i++) {
|
|
sock = modem_socket_from_fd(cfg, fds[i].fd);
|
|
if (!sock) {
|
|
continue;
|
|
}
|
|
|
|
/*
|
|
* Handle user check for ZSOCK_POLLOUT events:
|
|
* we consider the socket to always be writable.
|
|
*/
|
|
if (fds[i].events & ZSOCK_POLLOUT) {
|
|
fds[i].revents |= ZSOCK_POLLOUT;
|
|
found_count++;
|
|
} else if ((fds[i].events & ZSOCK_POLLIN) &&
|
|
(sock->packet_sizes[0] > 0U)) {
|
|
fds[i].revents |= ZSOCK_POLLIN;
|
|
found_count++;
|
|
}
|
|
|
|
sock->is_polled = false;
|
|
}
|
|
|
|
/* EBUSY, EAGAIN and ETIMEDOUT aren't true errors */
|
|
if (ret < 0 && ret != -EBUSY && ret != -EAGAIN && ret != -ETIMEDOUT) {
|
|
errno = ret;
|
|
return -1;
|
|
}
|
|
|
|
errno = 0;
|
|
return found_count;
|
|
}
|
|
|
|
void modem_socket_wait_data(struct modem_socket_config *cfg,
|
|
struct modem_socket *sock)
|
|
{
|
|
k_sem_take(&cfg->sem_lock, K_FOREVER);
|
|
sock->is_waiting = true;
|
|
k_sem_give(&cfg->sem_lock);
|
|
|
|
k_sem_take(&sock->sem_data_ready, K_FOREVER);
|
|
}
|
|
|
|
void modem_socket_data_ready(struct modem_socket_config *cfg,
|
|
struct modem_socket *sock)
|
|
{
|
|
k_sem_take(&cfg->sem_lock, K_FOREVER);
|
|
|
|
if (sock->is_waiting) {
|
|
/* unblock sockets waiting on recv() */
|
|
sock->is_waiting = false;
|
|
k_sem_give(&sock->sem_data_ready);
|
|
}
|
|
|
|
if (sock->is_polled) {
|
|
/* unblock poll() */
|
|
k_sem_give(&cfg->sem_poll);
|
|
}
|
|
|
|
k_sem_give(&cfg->sem_lock);
|
|
}
|
|
|
|
int modem_socket_init(struct modem_socket_config *cfg,
|
|
const struct socket_op_vtable *vtable)
|
|
{
|
|
int i;
|
|
|
|
k_sem_init(&cfg->sem_poll, 0, 1);
|
|
k_sem_init(&cfg->sem_lock, 1, 1);
|
|
for (i = 0; i < cfg->sockets_len; i++) {
|
|
k_sem_init(&cfg->sockets[i].sem_data_ready, 0, 1);
|
|
cfg->sockets[i].id = cfg->base_socket_num - 1;
|
|
}
|
|
|
|
cfg->vtable = vtable;
|
|
|
|
return 0;
|
|
}
|