|  | /* | 
|  | * Event loop based on select() loop | 
|  | * Copyright (c) 2002-2009, Jouni Malinen <j@w1.fi> | 
|  | * | 
|  | * This software may be distributed under the terms of the BSD license. | 
|  | * See README for more details. | 
|  | */ | 
|  |  | 
|  | #include "includes.h" | 
|  | #include <assert.h> | 
|  |  | 
|  | #include "common.h" | 
|  | #include "trace.h" | 
|  | #include "list.h" | 
|  | #include "eloop.h" | 
|  |  | 
|  | #if defined(CONFIG_ELOOP_POLL) && defined(CONFIG_ELOOP_EPOLL) | 
|  | #error Do not define both of poll and epoll | 
|  | #endif | 
|  |  | 
|  | #if defined(CONFIG_ELOOP_POLL) && defined(CONFIG_ELOOP_KQUEUE) | 
|  | #error Do not define both of poll and kqueue | 
|  | #endif | 
|  |  | 
|  | #if !defined(CONFIG_ELOOP_POLL) && !defined(CONFIG_ELOOP_EPOLL) && \ | 
|  | !defined(CONFIG_ELOOP_KQUEUE) | 
|  | #define CONFIG_ELOOP_SELECT | 
|  | #endif | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_POLL | 
|  | #include <poll.h> | 
|  | #endif /* CONFIG_ELOOP_POLL */ | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_EPOLL | 
|  | #include <sys/epoll.h> | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | #include <sys/event.h> | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  |  | 
|  | struct eloop_sock { | 
|  | int sock; | 
|  | void *eloop_data; | 
|  | void *user_data; | 
|  | eloop_sock_handler handler; | 
|  | WPA_TRACE_REF(eloop); | 
|  | WPA_TRACE_REF(user); | 
|  | WPA_TRACE_INFO | 
|  | }; | 
|  |  | 
|  | struct eloop_timeout { | 
|  | struct dl_list list; | 
|  | struct os_reltime time; | 
|  | void *eloop_data; | 
|  | void *user_data; | 
|  | eloop_timeout_handler handler; | 
|  | WPA_TRACE_REF(eloop); | 
|  | WPA_TRACE_REF(user); | 
|  | WPA_TRACE_INFO | 
|  | }; | 
|  |  | 
|  | struct eloop_signal { | 
|  | int sig; | 
|  | void *user_data; | 
|  | eloop_signal_handler handler; | 
|  | int signaled; | 
|  | }; | 
|  |  | 
|  | struct eloop_sock_table { | 
|  | size_t count; | 
|  | struct eloop_sock *table; | 
|  | eloop_event_type type; | 
|  | int changed; | 
|  | }; | 
|  |  | 
|  | struct eloop_data { | 
|  | int max_sock; | 
|  |  | 
|  | size_t count; /* sum of all table counts */ | 
|  | #ifdef CONFIG_ELOOP_POLL | 
|  | size_t max_pollfd_map; /* number of pollfds_map currently allocated */ | 
|  | size_t max_poll_fds; /* number of pollfds currently allocated */ | 
|  | struct pollfd *pollfds; | 
|  | struct pollfd **pollfds_map; | 
|  | #endif /* CONFIG_ELOOP_POLL */ | 
|  | #if defined(CONFIG_ELOOP_EPOLL) || defined(CONFIG_ELOOP_KQUEUE) | 
|  | int max_fd; | 
|  | struct eloop_sock *fd_table; | 
|  | #endif /* CONFIG_ELOOP_EPOLL || CONFIG_ELOOP_KQUEUE */ | 
|  | #ifdef CONFIG_ELOOP_EPOLL | 
|  | int epollfd; | 
|  | size_t epoll_max_event_num; | 
|  | struct epoll_event *epoll_events; | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | int kqueuefd; | 
|  | size_t kqueue_nevents; | 
|  | struct kevent *kqueue_events; | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  | struct eloop_sock_table readers; | 
|  | struct eloop_sock_table writers; | 
|  | struct eloop_sock_table exceptions; | 
|  |  | 
|  | struct dl_list timeout; | 
|  |  | 
|  | size_t signal_count; | 
|  | struct eloop_signal *signals; | 
|  | int signaled; | 
|  | int pending_terminate; | 
|  |  | 
|  | int terminate; | 
|  | }; | 
|  |  | 
|  | static struct eloop_data eloop; | 
|  |  | 
|  |  | 
|  | #ifdef WPA_TRACE | 
|  |  | 
|  | static void eloop_sigsegv_handler(int sig) | 
|  | { | 
|  | wpa_trace_show("eloop SIGSEGV"); | 
|  | abort(); | 
|  | } | 
|  |  | 
|  | static void eloop_trace_sock_add_ref(struct eloop_sock_table *table) | 
|  | { | 
|  | size_t i; | 
|  |  | 
|  | if (table == NULL || table->table == NULL) | 
|  | return; | 
|  | for (i = 0; i < table->count; i++) { | 
|  | wpa_trace_add_ref(&table->table[i], eloop, | 
|  | table->table[i].eloop_data); | 
|  | wpa_trace_add_ref(&table->table[i], user, | 
|  | table->table[i].user_data); | 
|  | } | 
|  | } | 
|  |  | 
|  |  | 
|  | static void eloop_trace_sock_remove_ref(struct eloop_sock_table *table) | 
|  | { | 
|  | size_t i; | 
|  |  | 
|  | if (table == NULL || table->table == NULL) | 
|  | return; | 
|  | for (i = 0; i < table->count; i++) { | 
|  | wpa_trace_remove_ref(&table->table[i], eloop, | 
|  | table->table[i].eloop_data); | 
|  | wpa_trace_remove_ref(&table->table[i], user, | 
|  | table->table[i].user_data); | 
|  | } | 
|  | } | 
|  |  | 
|  | #else /* WPA_TRACE */ | 
|  |  | 
|  | #define eloop_trace_sock_add_ref(table) do { } while (0) | 
|  | #define eloop_trace_sock_remove_ref(table) do { } while (0) | 
|  |  | 
|  | #endif /* WPA_TRACE */ | 
|  |  | 
|  |  | 
|  | int eloop_init(void) | 
|  | { | 
|  | os_memset(&eloop, 0, sizeof(eloop)); | 
|  | dl_list_init(&eloop.timeout); | 
|  | #ifdef CONFIG_ELOOP_EPOLL | 
|  | eloop.epollfd = epoll_create1(0); | 
|  | if (eloop.epollfd < 0) { | 
|  | wpa_printf(MSG_ERROR, "%s: epoll_create1 failed. %s", | 
|  | __func__, strerror(errno)); | 
|  | return -1; | 
|  | } | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | eloop.kqueuefd = kqueue(); | 
|  | if (eloop.kqueuefd < 0) { | 
|  | wpa_printf(MSG_ERROR, "%s: kqueue failed: %s", | 
|  | __func__, strerror(errno)); | 
|  | return -1; | 
|  | } | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  | #if defined(CONFIG_ELOOP_EPOLL) || defined(CONFIG_ELOOP_KQUEUE) | 
|  | eloop.readers.type = EVENT_TYPE_READ; | 
|  | eloop.writers.type = EVENT_TYPE_WRITE; | 
|  | eloop.exceptions.type = EVENT_TYPE_EXCEPTION; | 
|  | #endif /* CONFIG_ELOOP_EPOLL || CONFIG_ELOOP_KQUEUE */ | 
|  | #ifdef WPA_TRACE | 
|  | signal(SIGSEGV, eloop_sigsegv_handler); | 
|  | #endif /* WPA_TRACE */ | 
|  | return 0; | 
|  | } | 
|  |  | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_EPOLL | 
|  | static int eloop_sock_queue(int sock, eloop_event_type type) | 
|  | { | 
|  | struct epoll_event ev; | 
|  |  | 
|  | os_memset(&ev, 0, sizeof(ev)); | 
|  | switch (type) { | 
|  | case EVENT_TYPE_READ: | 
|  | ev.events = EPOLLIN; | 
|  | break; | 
|  | case EVENT_TYPE_WRITE: | 
|  | ev.events = EPOLLOUT; | 
|  | break; | 
|  | /* | 
|  | * Exceptions are always checked when using epoll, but I suppose it's | 
|  | * possible that someone registered a socket *only* for exception | 
|  | * handling. | 
|  | */ | 
|  | case EVENT_TYPE_EXCEPTION: | 
|  | ev.events = EPOLLERR | EPOLLHUP; | 
|  | break; | 
|  | } | 
|  | ev.data.fd = sock; | 
|  | if (epoll_ctl(eloop.epollfd, EPOLL_CTL_ADD, sock, &ev) < 0) { | 
|  | wpa_printf(MSG_ERROR, "%s: epoll_ctl(ADD) for fd=%d failed: %s", | 
|  | __func__, sock, strerror(errno)); | 
|  | return -1; | 
|  | } | 
|  | return 0; | 
|  | } | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  |  | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  |  | 
|  | static short event_type_kevent_filter(eloop_event_type type) | 
|  | { | 
|  | switch (type) { | 
|  | case EVENT_TYPE_READ: | 
|  | return EVFILT_READ; | 
|  | case EVENT_TYPE_WRITE: | 
|  | return EVFILT_WRITE; | 
|  | default: | 
|  | return 0; | 
|  | } | 
|  | } | 
|  |  | 
|  |  | 
|  | static int eloop_sock_queue(int sock, eloop_event_type type) | 
|  | { | 
|  | struct kevent ke; | 
|  |  | 
|  | EV_SET(&ke, sock, event_type_kevent_filter(type), EV_ADD, 0, 0, 0); | 
|  | if (kevent(eloop.kqueuefd, &ke, 1, NULL, 0, NULL) == -1) { | 
|  | wpa_printf(MSG_ERROR, "%s: kevent(ADD) for fd=%d failed: %s", | 
|  | __func__, sock, strerror(errno)); | 
|  | return -1; | 
|  | } | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  |  | 
|  |  | 
|  | static int eloop_sock_table_add_sock(struct eloop_sock_table *table, | 
|  | int sock, eloop_sock_handler handler, | 
|  | void *eloop_data, void *user_data) | 
|  | { | 
|  | #ifdef CONFIG_ELOOP_EPOLL | 
|  | struct epoll_event *temp_events; | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | struct kevent *temp_events; | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  | #if defined(CONFIG_ELOOP_EPOLL) || defined(CONFIG_ELOOP_KQUEUE) | 
|  | struct eloop_sock *temp_table; | 
|  | size_t next; | 
|  | #endif /* CONFIG_ELOOP_EPOLL || CONFIG_ELOOP_KQUEUE */ | 
|  | struct eloop_sock *tmp; | 
|  | int new_max_sock; | 
|  |  | 
|  | if (sock > eloop.max_sock) | 
|  | new_max_sock = sock; | 
|  | else | 
|  | new_max_sock = eloop.max_sock; | 
|  |  | 
|  | if (table == NULL) | 
|  | return -1; | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_POLL | 
|  | if ((size_t) new_max_sock >= eloop.max_pollfd_map) { | 
|  | struct pollfd **nmap; | 
|  | nmap = os_realloc_array(eloop.pollfds_map, new_max_sock + 50, | 
|  | sizeof(struct pollfd *)); | 
|  | if (nmap == NULL) | 
|  | return -1; | 
|  |  | 
|  | eloop.max_pollfd_map = new_max_sock + 50; | 
|  | eloop.pollfds_map = nmap; | 
|  | } | 
|  |  | 
|  | if (eloop.count + 1 > eloop.max_poll_fds) { | 
|  | struct pollfd *n; | 
|  | size_t nmax = eloop.count + 1 + 50; | 
|  |  | 
|  | n = os_realloc_array(eloop.pollfds, nmax, | 
|  | sizeof(struct pollfd)); | 
|  | if (n == NULL) | 
|  | return -1; | 
|  |  | 
|  | eloop.max_poll_fds = nmax; | 
|  | eloop.pollfds = n; | 
|  | } | 
|  | #endif /* CONFIG_ELOOP_POLL */ | 
|  | #if defined(CONFIG_ELOOP_EPOLL) || defined(CONFIG_ELOOP_KQUEUE) | 
|  | if (new_max_sock >= eloop.max_fd) { | 
|  | next = new_max_sock + 16; | 
|  | temp_table = os_realloc_array(eloop.fd_table, next, | 
|  | sizeof(struct eloop_sock)); | 
|  | if (temp_table == NULL) | 
|  | return -1; | 
|  |  | 
|  | eloop.max_fd = next; | 
|  | eloop.fd_table = temp_table; | 
|  | } | 
|  | #endif /* CONFIG_ELOOP_EPOLL || CONFIG_ELOOP_KQUEUE */ | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_EPOLL | 
|  | if (eloop.count + 1 > eloop.epoll_max_event_num) { | 
|  | next = eloop.epoll_max_event_num == 0 ? 8 : | 
|  | eloop.epoll_max_event_num * 2; | 
|  | temp_events = os_realloc_array(eloop.epoll_events, next, | 
|  | sizeof(struct epoll_event)); | 
|  | if (temp_events == NULL) { | 
|  | wpa_printf(MSG_ERROR, "%s: malloc for epoll failed: %s", | 
|  | __func__, strerror(errno)); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | eloop.epoll_max_event_num = next; | 
|  | eloop.epoll_events = temp_events; | 
|  | } | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | if (eloop.count + 1 > eloop.kqueue_nevents) { | 
|  | next = eloop.kqueue_nevents == 0 ? 8 : eloop.kqueue_nevents * 2; | 
|  | temp_events = os_malloc(next * sizeof(*temp_events)); | 
|  | if (!temp_events) { | 
|  | wpa_printf(MSG_ERROR, | 
|  | "%s: malloc for kqueue failed: %s", | 
|  | __func__, strerror(errno)); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | os_free(eloop.kqueue_events); | 
|  | eloop.kqueue_events = temp_events; | 
|  | eloop.kqueue_nevents = next; | 
|  | } | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  |  | 
|  | eloop_trace_sock_remove_ref(table); | 
|  | tmp = os_realloc_array(table->table, table->count + 1, | 
|  | sizeof(struct eloop_sock)); | 
|  | if (tmp == NULL) { | 
|  | eloop_trace_sock_add_ref(table); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | tmp[table->count].sock = sock; | 
|  | tmp[table->count].eloop_data = eloop_data; | 
|  | tmp[table->count].user_data = user_data; | 
|  | tmp[table->count].handler = handler; | 
|  | wpa_trace_record(&tmp[table->count]); | 
|  | table->count++; | 
|  | table->table = tmp; | 
|  | eloop.max_sock = new_max_sock; | 
|  | eloop.count++; | 
|  | table->changed = 1; | 
|  | eloop_trace_sock_add_ref(table); | 
|  |  | 
|  | #if defined(CONFIG_ELOOP_EPOLL) || defined(CONFIG_ELOOP_KQUEUE) | 
|  | if (eloop_sock_queue(sock, table->type) < 0) | 
|  | return -1; | 
|  | os_memcpy(&eloop.fd_table[sock], &table->table[table->count - 1], | 
|  | sizeof(struct eloop_sock)); | 
|  | #endif /* CONFIG_ELOOP_EPOLL || CONFIG_ELOOP_KQUEUE */ | 
|  | return 0; | 
|  | } | 
|  |  | 
|  |  | 
|  | static void eloop_sock_table_remove_sock(struct eloop_sock_table *table, | 
|  | int sock) | 
|  | { | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | struct kevent ke; | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  | size_t i; | 
|  |  | 
|  | if (table == NULL || table->table == NULL || table->count == 0) | 
|  | return; | 
|  |  | 
|  | for (i = 0; i < table->count; i++) { | 
|  | if (table->table[i].sock == sock) | 
|  | break; | 
|  | } | 
|  | if (i == table->count) | 
|  | return; | 
|  | eloop_trace_sock_remove_ref(table); | 
|  | if (i != table->count - 1) { | 
|  | os_memmove(&table->table[i], &table->table[i + 1], | 
|  | (table->count - i - 1) * | 
|  | sizeof(struct eloop_sock)); | 
|  | } | 
|  | table->count--; | 
|  | eloop.count--; | 
|  | table->changed = 1; | 
|  | eloop_trace_sock_add_ref(table); | 
|  | #ifdef CONFIG_ELOOP_EPOLL | 
|  | if (epoll_ctl(eloop.epollfd, EPOLL_CTL_DEL, sock, NULL) < 0) { | 
|  | wpa_printf(MSG_ERROR, "%s: epoll_ctl(DEL) for fd=%d failed: %s", | 
|  | __func__, sock, strerror(errno)); | 
|  | return; | 
|  | } | 
|  | os_memset(&eloop.fd_table[sock], 0, sizeof(struct eloop_sock)); | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | EV_SET(&ke, sock, event_type_kevent_filter(table->type), EV_DELETE, 0, | 
|  | 0, 0); | 
|  | if (kevent(eloop.kqueuefd, &ke, 1, NULL, 0, NULL) < 0) { | 
|  | wpa_printf(MSG_ERROR, "%s: kevent(DEL) for fd=%d failed: %s", | 
|  | __func__, sock, strerror(errno)); | 
|  | return; | 
|  | } | 
|  | os_memset(&eloop.fd_table[sock], 0, sizeof(struct eloop_sock)); | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  | } | 
|  |  | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_POLL | 
|  |  | 
|  | static struct pollfd * find_pollfd(struct pollfd **pollfds_map, int fd, int mx) | 
|  | { | 
|  | if (fd < mx && fd >= 0) | 
|  | return pollfds_map[fd]; | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  |  | 
|  | static int eloop_sock_table_set_fds(struct eloop_sock_table *readers, | 
|  | struct eloop_sock_table *writers, | 
|  | struct eloop_sock_table *exceptions, | 
|  | struct pollfd *pollfds, | 
|  | struct pollfd **pollfds_map, | 
|  | int max_pollfd_map) | 
|  | { | 
|  | size_t i; | 
|  | int nxt = 0; | 
|  | int fd; | 
|  | struct pollfd *pfd; | 
|  |  | 
|  | /* Clear pollfd lookup map. It will be re-populated below. */ | 
|  | os_memset(pollfds_map, 0, sizeof(struct pollfd *) * max_pollfd_map); | 
|  |  | 
|  | if (readers && readers->table) { | 
|  | for (i = 0; i < readers->count; i++) { | 
|  | fd = readers->table[i].sock; | 
|  | assert(fd >= 0 && fd < max_pollfd_map); | 
|  | pollfds[nxt].fd = fd; | 
|  | pollfds[nxt].events = POLLIN; | 
|  | pollfds[nxt].revents = 0; | 
|  | pollfds_map[fd] = &(pollfds[nxt]); | 
|  | nxt++; | 
|  | } | 
|  | } | 
|  |  | 
|  | if (writers && writers->table) { | 
|  | for (i = 0; i < writers->count; i++) { | 
|  | /* | 
|  | * See if we already added this descriptor, update it | 
|  | * if so. | 
|  | */ | 
|  | fd = writers->table[i].sock; | 
|  | assert(fd >= 0 && fd < max_pollfd_map); | 
|  | pfd = pollfds_map[fd]; | 
|  | if (!pfd) { | 
|  | pfd = &(pollfds[nxt]); | 
|  | pfd->events = 0; | 
|  | pfd->fd = fd; | 
|  | pollfds[i].revents = 0; | 
|  | pollfds_map[fd] = pfd; | 
|  | nxt++; | 
|  | } | 
|  | pfd->events |= POLLOUT; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Exceptions are always checked when using poll, but I suppose it's | 
|  | * possible that someone registered a socket *only* for exception | 
|  | * handling. Set the POLLIN bit in this case. | 
|  | */ | 
|  | if (exceptions && exceptions->table) { | 
|  | for (i = 0; i < exceptions->count; i++) { | 
|  | /* | 
|  | * See if we already added this descriptor, just use it | 
|  | * if so. | 
|  | */ | 
|  | fd = exceptions->table[i].sock; | 
|  | assert(fd >= 0 && fd < max_pollfd_map); | 
|  | pfd = pollfds_map[fd]; | 
|  | if (!pfd) { | 
|  | pfd = &(pollfds[nxt]); | 
|  | pfd->events = POLLIN; | 
|  | pfd->fd = fd; | 
|  | pollfds[i].revents = 0; | 
|  | pollfds_map[fd] = pfd; | 
|  | nxt++; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | return nxt; | 
|  | } | 
|  |  | 
|  |  | 
|  | static int eloop_sock_table_dispatch_table(struct eloop_sock_table *table, | 
|  | struct pollfd **pollfds_map, | 
|  | int max_pollfd_map, | 
|  | short int revents) | 
|  | { | 
|  | size_t i; | 
|  | struct pollfd *pfd; | 
|  |  | 
|  | if (!table || !table->table) | 
|  | return 0; | 
|  |  | 
|  | table->changed = 0; | 
|  | for (i = 0; i < table->count; i++) { | 
|  | pfd = find_pollfd(pollfds_map, table->table[i].sock, | 
|  | max_pollfd_map); | 
|  | if (!pfd) | 
|  | continue; | 
|  |  | 
|  | if (!(pfd->revents & revents)) | 
|  | continue; | 
|  |  | 
|  | table->table[i].handler(table->table[i].sock, | 
|  | table->table[i].eloop_data, | 
|  | table->table[i].user_data); | 
|  | if (table->changed) | 
|  | return 1; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  |  | 
|  | static void eloop_sock_table_dispatch(struct eloop_sock_table *readers, | 
|  | struct eloop_sock_table *writers, | 
|  | struct eloop_sock_table *exceptions, | 
|  | struct pollfd **pollfds_map, | 
|  | int max_pollfd_map) | 
|  | { | 
|  | if (eloop_sock_table_dispatch_table(readers, pollfds_map, | 
|  | max_pollfd_map, POLLIN | POLLERR | | 
|  | POLLHUP)) | 
|  | return; /* pollfds may be invalid at this point */ | 
|  |  | 
|  | if (eloop_sock_table_dispatch_table(writers, pollfds_map, | 
|  | max_pollfd_map, POLLOUT)) | 
|  | return; /* pollfds may be invalid at this point */ | 
|  |  | 
|  | eloop_sock_table_dispatch_table(exceptions, pollfds_map, | 
|  | max_pollfd_map, POLLERR | POLLHUP); | 
|  | } | 
|  |  | 
|  | #endif /* CONFIG_ELOOP_POLL */ | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_SELECT | 
|  |  | 
|  | static void eloop_sock_table_set_fds(struct eloop_sock_table *table, | 
|  | fd_set *fds) | 
|  | { | 
|  | size_t i; | 
|  |  | 
|  | FD_ZERO(fds); | 
|  |  | 
|  | if (table->table == NULL) | 
|  | return; | 
|  |  | 
|  | for (i = 0; i < table->count; i++) { | 
|  | assert(table->table[i].sock >= 0); | 
|  | FD_SET(table->table[i].sock, fds); | 
|  | } | 
|  | } | 
|  |  | 
|  |  | 
|  | static void eloop_sock_table_dispatch(struct eloop_sock_table *table, | 
|  | fd_set *fds) | 
|  | { | 
|  | size_t i; | 
|  |  | 
|  | if (table == NULL || table->table == NULL) | 
|  | return; | 
|  |  | 
|  | table->changed = 0; | 
|  | for (i = 0; i < table->count; i++) { | 
|  | if (FD_ISSET(table->table[i].sock, fds)) { | 
|  | table->table[i].handler(table->table[i].sock, | 
|  | table->table[i].eloop_data, | 
|  | table->table[i].user_data); | 
|  | if (table->changed) | 
|  | break; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | #endif /* CONFIG_ELOOP_SELECT */ | 
|  |  | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_EPOLL | 
|  | static void eloop_sock_table_dispatch(struct epoll_event *events, int nfds) | 
|  | { | 
|  | struct eloop_sock *table; | 
|  | int i; | 
|  |  | 
|  | for (i = 0; i < nfds; i++) { | 
|  | table = &eloop.fd_table[events[i].data.fd]; | 
|  | if (table->handler == NULL) | 
|  | continue; | 
|  | table->handler(table->sock, table->eloop_data, | 
|  | table->user_data); | 
|  | if (eloop.readers.changed || | 
|  | eloop.writers.changed || | 
|  | eloop.exceptions.changed) | 
|  | break; | 
|  | } | 
|  | } | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  |  | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  |  | 
|  | static void eloop_sock_table_dispatch(struct kevent *events, int nfds) | 
|  | { | 
|  | struct eloop_sock *table; | 
|  | int i; | 
|  |  | 
|  | for (i = 0; i < nfds; i++) { | 
|  | table = &eloop.fd_table[events[i].ident]; | 
|  | if (table->handler == NULL) | 
|  | continue; | 
|  | table->handler(table->sock, table->eloop_data, | 
|  | table->user_data); | 
|  | if (eloop.readers.changed || | 
|  | eloop.writers.changed || | 
|  | eloop.exceptions.changed) | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  |  | 
|  | static int eloop_sock_table_requeue(struct eloop_sock_table *table) | 
|  | { | 
|  | size_t i; | 
|  | int r; | 
|  |  | 
|  | r = 0; | 
|  | for (i = 0; i < table->count && table->table; i++) { | 
|  | if (eloop_sock_queue(table->table[i].sock, table->type) == -1) | 
|  | r = -1; | 
|  | } | 
|  | return r; | 
|  | } | 
|  |  | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  |  | 
|  |  | 
|  | int eloop_sock_requeue(void) | 
|  | { | 
|  | int r = 0; | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | close(eloop.kqueuefd); | 
|  | eloop.kqueuefd = kqueue(); | 
|  | if (eloop.kqueuefd < 0) { | 
|  | wpa_printf(MSG_ERROR, "%s: kqueue failed: %s", | 
|  | __func__, strerror(errno)); | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | if (eloop_sock_table_requeue(&eloop.readers) < 0) | 
|  | r = -1; | 
|  | if (eloop_sock_table_requeue(&eloop.writers) < 0) | 
|  | r = -1; | 
|  | if (eloop_sock_table_requeue(&eloop.exceptions) < 0) | 
|  | r = -1; | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  |  | 
|  | return r; | 
|  | } | 
|  |  | 
|  |  | 
|  | static void eloop_sock_table_destroy(struct eloop_sock_table *table) | 
|  | { | 
|  | if (table) { | 
|  | size_t i; | 
|  |  | 
|  | for (i = 0; i < table->count && table->table; i++) { | 
|  | wpa_printf(MSG_INFO, "ELOOP: remaining socket: " | 
|  | "sock=%d eloop_data=%p user_data=%p " | 
|  | "handler=%p", | 
|  | table->table[i].sock, | 
|  | table->table[i].eloop_data, | 
|  | table->table[i].user_data, | 
|  | table->table[i].handler); | 
|  | wpa_trace_dump_funcname("eloop unregistered socket " | 
|  | "handler", | 
|  | table->table[i].handler); | 
|  | wpa_trace_dump("eloop sock", &table->table[i]); | 
|  | } | 
|  | os_free(table->table); | 
|  | } | 
|  | } | 
|  |  | 
|  |  | 
|  | int eloop_register_read_sock(int sock, eloop_sock_handler handler, | 
|  | void *eloop_data, void *user_data) | 
|  | { | 
|  | return eloop_register_sock(sock, EVENT_TYPE_READ, handler, | 
|  | eloop_data, user_data); | 
|  | } | 
|  |  | 
|  |  | 
|  | void eloop_unregister_read_sock(int sock) | 
|  | { | 
|  | eloop_unregister_sock(sock, EVENT_TYPE_READ); | 
|  | } | 
|  |  | 
|  |  | 
|  | static struct eloop_sock_table *eloop_get_sock_table(eloop_event_type type) | 
|  | { | 
|  | switch (type) { | 
|  | case EVENT_TYPE_READ: | 
|  | return &eloop.readers; | 
|  | case EVENT_TYPE_WRITE: | 
|  | return &eloop.writers; | 
|  | case EVENT_TYPE_EXCEPTION: | 
|  | return &eloop.exceptions; | 
|  | } | 
|  |  | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  |  | 
|  | int eloop_register_sock(int sock, eloop_event_type type, | 
|  | eloop_sock_handler handler, | 
|  | void *eloop_data, void *user_data) | 
|  | { | 
|  | struct eloop_sock_table *table; | 
|  |  | 
|  | assert(sock >= 0); | 
|  | table = eloop_get_sock_table(type); | 
|  | return eloop_sock_table_add_sock(table, sock, handler, | 
|  | eloop_data, user_data); | 
|  | } | 
|  |  | 
|  |  | 
|  | void eloop_unregister_sock(int sock, eloop_event_type type) | 
|  | { | 
|  | struct eloop_sock_table *table; | 
|  |  | 
|  | table = eloop_get_sock_table(type); | 
|  | eloop_sock_table_remove_sock(table, sock); | 
|  | } | 
|  |  | 
|  |  | 
|  | int eloop_register_timeout(unsigned int secs, unsigned int usecs, | 
|  | eloop_timeout_handler handler, | 
|  | void *eloop_data, void *user_data) | 
|  | { | 
|  | struct eloop_timeout *timeout, *tmp; | 
|  | os_time_t now_sec; | 
|  |  | 
|  | timeout = os_zalloc(sizeof(*timeout)); | 
|  | if (timeout == NULL) | 
|  | return -1; | 
|  | if (os_get_reltime(&timeout->time) < 0) { | 
|  | os_free(timeout); | 
|  | return -1; | 
|  | } | 
|  | now_sec = timeout->time.sec; | 
|  | timeout->time.sec += secs; | 
|  | if (timeout->time.sec < now_sec) | 
|  | goto overflow; | 
|  | timeout->time.usec += usecs; | 
|  | while (timeout->time.usec >= 1000000) { | 
|  | timeout->time.sec++; | 
|  | timeout->time.usec -= 1000000; | 
|  | } | 
|  | if (timeout->time.sec < now_sec) | 
|  | goto overflow; | 
|  | timeout->eloop_data = eloop_data; | 
|  | timeout->user_data = user_data; | 
|  | timeout->handler = handler; | 
|  | wpa_trace_add_ref(timeout, eloop, eloop_data); | 
|  | wpa_trace_add_ref(timeout, user, user_data); | 
|  | wpa_trace_record(timeout); | 
|  |  | 
|  | /* Maintain timeouts in order of increasing time */ | 
|  | dl_list_for_each(tmp, &eloop.timeout, struct eloop_timeout, list) { | 
|  | if (os_reltime_before(&timeout->time, &tmp->time)) { | 
|  | dl_list_add(tmp->list.prev, &timeout->list); | 
|  | return 0; | 
|  | } | 
|  | } | 
|  | dl_list_add_tail(&eloop.timeout, &timeout->list); | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | overflow: | 
|  | /* | 
|  | * Integer overflow - assume long enough timeout to be assumed | 
|  | * to be infinite, i.e., the timeout would never happen. | 
|  | */ | 
|  | wpa_printf(MSG_DEBUG, | 
|  | "ELOOP: Too long timeout (secs=%u usecs=%u) to ever happen - ignore it", | 
|  | secs,usecs); | 
|  | os_free(timeout); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  |  | 
|  | static void eloop_remove_timeout(struct eloop_timeout *timeout) | 
|  | { | 
|  | dl_list_del(&timeout->list); | 
|  | wpa_trace_remove_ref(timeout, eloop, timeout->eloop_data); | 
|  | wpa_trace_remove_ref(timeout, user, timeout->user_data); | 
|  | os_free(timeout); | 
|  | } | 
|  |  | 
|  |  | 
|  | int eloop_cancel_timeout(eloop_timeout_handler handler, | 
|  | void *eloop_data, void *user_data) | 
|  | { | 
|  | struct eloop_timeout *timeout, *prev; | 
|  | int removed = 0; | 
|  |  | 
|  | dl_list_for_each_safe(timeout, prev, &eloop.timeout, | 
|  | struct eloop_timeout, list) { | 
|  | if (timeout->handler == handler && | 
|  | (timeout->eloop_data == eloop_data || | 
|  | eloop_data == ELOOP_ALL_CTX) && | 
|  | (timeout->user_data == user_data || | 
|  | user_data == ELOOP_ALL_CTX)) { | 
|  | eloop_remove_timeout(timeout); | 
|  | removed++; | 
|  | } | 
|  | } | 
|  |  | 
|  | return removed; | 
|  | } | 
|  |  | 
|  |  | 
|  | int eloop_cancel_timeout_one(eloop_timeout_handler handler, | 
|  | void *eloop_data, void *user_data, | 
|  | struct os_reltime *remaining) | 
|  | { | 
|  | struct eloop_timeout *timeout, *prev; | 
|  | int removed = 0; | 
|  | struct os_reltime now; | 
|  |  | 
|  | os_get_reltime(&now); | 
|  | remaining->sec = remaining->usec = 0; | 
|  |  | 
|  | dl_list_for_each_safe(timeout, prev, &eloop.timeout, | 
|  | struct eloop_timeout, list) { | 
|  | if (timeout->handler == handler && | 
|  | (timeout->eloop_data == eloop_data) && | 
|  | (timeout->user_data == user_data)) { | 
|  | removed = 1; | 
|  | if (os_reltime_before(&now, &timeout->time)) | 
|  | os_reltime_sub(&timeout->time, &now, remaining); | 
|  | eloop_remove_timeout(timeout); | 
|  | break; | 
|  | } | 
|  | } | 
|  | return removed; | 
|  | } | 
|  |  | 
|  |  | 
|  | int eloop_is_timeout_registered(eloop_timeout_handler handler, | 
|  | void *eloop_data, void *user_data) | 
|  | { | 
|  | struct eloop_timeout *tmp; | 
|  |  | 
|  | dl_list_for_each(tmp, &eloop.timeout, struct eloop_timeout, list) { | 
|  | if (tmp->handler == handler && | 
|  | tmp->eloop_data == eloop_data && | 
|  | tmp->user_data == user_data) | 
|  | return 1; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  |  | 
|  | int eloop_deplete_timeout(unsigned int req_secs, unsigned int req_usecs, | 
|  | eloop_timeout_handler handler, void *eloop_data, | 
|  | void *user_data) | 
|  | { | 
|  | struct os_reltime now, requested, remaining; | 
|  | struct eloop_timeout *tmp; | 
|  |  | 
|  | dl_list_for_each(tmp, &eloop.timeout, struct eloop_timeout, list) { | 
|  | if (tmp->handler == handler && | 
|  | tmp->eloop_data == eloop_data && | 
|  | tmp->user_data == user_data) { | 
|  | requested.sec = req_secs; | 
|  | requested.usec = req_usecs; | 
|  | os_get_reltime(&now); | 
|  | os_reltime_sub(&tmp->time, &now, &remaining); | 
|  | if (os_reltime_before(&requested, &remaining)) { | 
|  | eloop_cancel_timeout(handler, eloop_data, | 
|  | user_data); | 
|  | eloop_register_timeout(requested.sec, | 
|  | requested.usec, | 
|  | handler, eloop_data, | 
|  | user_data); | 
|  | return 1; | 
|  | } | 
|  | return 0; | 
|  | } | 
|  | } | 
|  |  | 
|  | return -1; | 
|  | } | 
|  |  | 
|  |  | 
|  | int eloop_replenish_timeout(unsigned int req_secs, unsigned int req_usecs, | 
|  | eloop_timeout_handler handler, void *eloop_data, | 
|  | void *user_data) | 
|  | { | 
|  | struct os_reltime now, requested, remaining; | 
|  | struct eloop_timeout *tmp; | 
|  |  | 
|  | dl_list_for_each(tmp, &eloop.timeout, struct eloop_timeout, list) { | 
|  | if (tmp->handler == handler && | 
|  | tmp->eloop_data == eloop_data && | 
|  | tmp->user_data == user_data) { | 
|  | requested.sec = req_secs; | 
|  | requested.usec = req_usecs; | 
|  | os_get_reltime(&now); | 
|  | os_reltime_sub(&tmp->time, &now, &remaining); | 
|  | if (os_reltime_before(&remaining, &requested)) { | 
|  | eloop_cancel_timeout(handler, eloop_data, | 
|  | user_data); | 
|  | eloop_register_timeout(requested.sec, | 
|  | requested.usec, | 
|  | handler, eloop_data, | 
|  | user_data); | 
|  | return 1; | 
|  | } | 
|  | return 0; | 
|  | } | 
|  | } | 
|  |  | 
|  | return -1; | 
|  | } | 
|  |  | 
|  |  | 
|  | #ifndef CONFIG_NATIVE_WINDOWS | 
|  | static void eloop_handle_alarm(int sig) | 
|  | { | 
|  | wpa_printf(MSG_ERROR, "eloop: could not process SIGINT or SIGTERM in " | 
|  | "two seconds. Looks like there\n" | 
|  | "is a bug that ends up in a busy loop that " | 
|  | "prevents clean shutdown.\n" | 
|  | "Killing program forcefully.\n"); | 
|  | exit(1); | 
|  | } | 
|  | #endif /* CONFIG_NATIVE_WINDOWS */ | 
|  |  | 
|  |  | 
|  | static void eloop_handle_signal(int sig) | 
|  | { | 
|  | size_t i; | 
|  |  | 
|  | #ifndef CONFIG_NATIVE_WINDOWS | 
|  | if ((sig == SIGINT || sig == SIGTERM) && !eloop.pending_terminate) { | 
|  | /* Use SIGALRM to break out from potential busy loops that | 
|  | * would not allow the program to be killed. */ | 
|  | eloop.pending_terminate = 1; | 
|  | signal(SIGALRM, eloop_handle_alarm); | 
|  | alarm(2); | 
|  | } | 
|  | #endif /* CONFIG_NATIVE_WINDOWS */ | 
|  |  | 
|  | eloop.signaled++; | 
|  | for (i = 0; i < eloop.signal_count; i++) { | 
|  | if (eloop.signals[i].sig == sig) { | 
|  | eloop.signals[i].signaled++; | 
|  | break; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  |  | 
|  | static void eloop_process_pending_signals(void) | 
|  | { | 
|  | size_t i; | 
|  |  | 
|  | if (eloop.signaled == 0) | 
|  | return; | 
|  | eloop.signaled = 0; | 
|  |  | 
|  | if (eloop.pending_terminate) { | 
|  | #ifndef CONFIG_NATIVE_WINDOWS | 
|  | alarm(0); | 
|  | #endif /* CONFIG_NATIVE_WINDOWS */ | 
|  | eloop.pending_terminate = 0; | 
|  | } | 
|  |  | 
|  | for (i = 0; i < eloop.signal_count; i++) { | 
|  | if (eloop.signals[i].signaled) { | 
|  | eloop.signals[i].signaled = 0; | 
|  | eloop.signals[i].handler(eloop.signals[i].sig, | 
|  | eloop.signals[i].user_data); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  |  | 
|  | int eloop_register_signal(int sig, eloop_signal_handler handler, | 
|  | void *user_data) | 
|  | { | 
|  | struct eloop_signal *tmp; | 
|  |  | 
|  | tmp = os_realloc_array(eloop.signals, eloop.signal_count + 1, | 
|  | sizeof(struct eloop_signal)); | 
|  | if (tmp == NULL) | 
|  | return -1; | 
|  |  | 
|  | tmp[eloop.signal_count].sig = sig; | 
|  | tmp[eloop.signal_count].user_data = user_data; | 
|  | tmp[eloop.signal_count].handler = handler; | 
|  | tmp[eloop.signal_count].signaled = 0; | 
|  | eloop.signal_count++; | 
|  | eloop.signals = tmp; | 
|  | signal(sig, eloop_handle_signal); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  |  | 
|  | int eloop_register_signal_terminate(eloop_signal_handler handler, | 
|  | void *user_data) | 
|  | { | 
|  | int ret = eloop_register_signal(SIGINT, handler, user_data); | 
|  | if (ret == 0) | 
|  | ret = eloop_register_signal(SIGTERM, handler, user_data); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  |  | 
|  | int eloop_register_signal_reconfig(eloop_signal_handler handler, | 
|  | void *user_data) | 
|  | { | 
|  | #ifdef CONFIG_NATIVE_WINDOWS | 
|  | return 0; | 
|  | #else /* CONFIG_NATIVE_WINDOWS */ | 
|  | return eloop_register_signal(SIGHUP, handler, user_data); | 
|  | #endif /* CONFIG_NATIVE_WINDOWS */ | 
|  | } | 
|  |  | 
|  |  | 
|  | void eloop_run(void) | 
|  | { | 
|  | #ifdef CONFIG_ELOOP_POLL | 
|  | int num_poll_fds; | 
|  | int timeout_ms = 0; | 
|  | #endif /* CONFIG_ELOOP_POLL */ | 
|  | #ifdef CONFIG_ELOOP_SELECT | 
|  | fd_set *rfds, *wfds, *efds; | 
|  | struct timeval _tv; | 
|  | #endif /* CONFIG_ELOOP_SELECT */ | 
|  | #ifdef CONFIG_ELOOP_EPOLL | 
|  | int timeout_ms = -1; | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | struct timespec ts; | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  | int res; | 
|  | struct os_reltime tv, now; | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_SELECT | 
|  | rfds = os_malloc(sizeof(*rfds)); | 
|  | wfds = os_malloc(sizeof(*wfds)); | 
|  | efds = os_malloc(sizeof(*efds)); | 
|  | if (rfds == NULL || wfds == NULL || efds == NULL) | 
|  | goto out; | 
|  | #endif /* CONFIG_ELOOP_SELECT */ | 
|  |  | 
|  | while (!eloop.terminate && | 
|  | (!dl_list_empty(&eloop.timeout) || eloop.readers.count > 0 || | 
|  | eloop.writers.count > 0 || eloop.exceptions.count > 0)) { | 
|  | struct eloop_timeout *timeout; | 
|  |  | 
|  | if (eloop.pending_terminate) { | 
|  | /* | 
|  | * This may happen in some corner cases where a signal | 
|  | * is received during a blocking operation. We need to | 
|  | * process the pending signals and exit if requested to | 
|  | * avoid hitting the SIGALRM limit if the blocking | 
|  | * operation took more than two seconds. | 
|  | */ | 
|  | eloop_process_pending_signals(); | 
|  | if (eloop.terminate) | 
|  | break; | 
|  | } | 
|  |  | 
|  | timeout = dl_list_first(&eloop.timeout, struct eloop_timeout, | 
|  | list); | 
|  | if (timeout) { | 
|  | os_get_reltime(&now); | 
|  | if (os_reltime_before(&now, &timeout->time)) | 
|  | os_reltime_sub(&timeout->time, &now, &tv); | 
|  | else | 
|  | tv.sec = tv.usec = 0; | 
|  | #if defined(CONFIG_ELOOP_POLL) || defined(CONFIG_ELOOP_EPOLL) | 
|  | timeout_ms = tv.sec * 1000 + tv.usec / 1000; | 
|  | #endif /* defined(CONFIG_ELOOP_POLL) || defined(CONFIG_ELOOP_EPOLL) */ | 
|  | #ifdef CONFIG_ELOOP_SELECT | 
|  | _tv.tv_sec = tv.sec; | 
|  | _tv.tv_usec = tv.usec; | 
|  | #endif /* CONFIG_ELOOP_SELECT */ | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | ts.tv_sec = tv.sec; | 
|  | ts.tv_nsec = tv.usec * 1000L; | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_POLL | 
|  | num_poll_fds = eloop_sock_table_set_fds( | 
|  | &eloop.readers, &eloop.writers, &eloop.exceptions, | 
|  | eloop.pollfds, eloop.pollfds_map, | 
|  | eloop.max_pollfd_map); | 
|  | res = poll(eloop.pollfds, num_poll_fds, | 
|  | timeout ? timeout_ms : -1); | 
|  | #endif /* CONFIG_ELOOP_POLL */ | 
|  | #ifdef CONFIG_ELOOP_SELECT | 
|  | eloop_sock_table_set_fds(&eloop.readers, rfds); | 
|  | eloop_sock_table_set_fds(&eloop.writers, wfds); | 
|  | eloop_sock_table_set_fds(&eloop.exceptions, efds); | 
|  | res = select(eloop.max_sock + 1, rfds, wfds, efds, | 
|  | timeout ? &_tv : NULL); | 
|  | #endif /* CONFIG_ELOOP_SELECT */ | 
|  | #ifdef CONFIG_ELOOP_EPOLL | 
|  | if (eloop.count == 0) { | 
|  | res = 0; | 
|  | } else { | 
|  | res = epoll_wait(eloop.epollfd, eloop.epoll_events, | 
|  | eloop.count, timeout_ms); | 
|  | } | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | if (eloop.count == 0) { | 
|  | res = 0; | 
|  | } else { | 
|  | res = kevent(eloop.kqueuefd, NULL, 0, | 
|  | eloop.kqueue_events, eloop.kqueue_nevents, | 
|  | timeout ? &ts : NULL); | 
|  | } | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  | if (res < 0 && errno != EINTR && errno != 0) { | 
|  | wpa_printf(MSG_ERROR, "eloop: %s: %s", | 
|  | #ifdef CONFIG_ELOOP_POLL | 
|  | "poll" | 
|  | #endif /* CONFIG_ELOOP_POLL */ | 
|  | #ifdef CONFIG_ELOOP_SELECT | 
|  | "select" | 
|  | #endif /* CONFIG_ELOOP_SELECT */ | 
|  | #ifdef CONFIG_ELOOP_EPOLL | 
|  | "epoll" | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | "kqueue" | 
|  | #endif /* CONFIG_ELOOP_EKQUEUE */ | 
|  |  | 
|  | , strerror(errno)); | 
|  | goto out; | 
|  | } | 
|  |  | 
|  | eloop.readers.changed = 0; | 
|  | eloop.writers.changed = 0; | 
|  | eloop.exceptions.changed = 0; | 
|  |  | 
|  | eloop_process_pending_signals(); | 
|  |  | 
|  |  | 
|  | /* check if some registered timeouts have occurred */ | 
|  | timeout = dl_list_first(&eloop.timeout, struct eloop_timeout, | 
|  | list); | 
|  | if (timeout) { | 
|  | os_get_reltime(&now); | 
|  | if (!os_reltime_before(&now, &timeout->time)) { | 
|  | void *eloop_data = timeout->eloop_data; | 
|  | void *user_data = timeout->user_data; | 
|  | eloop_timeout_handler handler = | 
|  | timeout->handler; | 
|  | eloop_remove_timeout(timeout); | 
|  | handler(eloop_data, user_data); | 
|  | } | 
|  |  | 
|  | } | 
|  |  | 
|  | if (res <= 0) | 
|  | continue; | 
|  |  | 
|  | if (eloop.readers.changed || | 
|  | eloop.writers.changed || | 
|  | eloop.exceptions.changed) { | 
|  | /* | 
|  | * Sockets may have been closed and reopened with the | 
|  | * same FD in the signal or timeout handlers, so we | 
|  | * must skip the previous results and check again | 
|  | * whether any of the currently registered sockets have | 
|  | * events. | 
|  | */ | 
|  | continue; | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_POLL | 
|  | eloop_sock_table_dispatch(&eloop.readers, &eloop.writers, | 
|  | &eloop.exceptions, eloop.pollfds_map, | 
|  | eloop.max_pollfd_map); | 
|  | #endif /* CONFIG_ELOOP_POLL */ | 
|  | #ifdef CONFIG_ELOOP_SELECT | 
|  | eloop_sock_table_dispatch(&eloop.readers, rfds); | 
|  | eloop_sock_table_dispatch(&eloop.writers, wfds); | 
|  | eloop_sock_table_dispatch(&eloop.exceptions, efds); | 
|  | #endif /* CONFIG_ELOOP_SELECT */ | 
|  | #ifdef CONFIG_ELOOP_EPOLL | 
|  | eloop_sock_table_dispatch(eloop.epoll_events, res); | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | eloop_sock_table_dispatch(eloop.kqueue_events, res); | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  | } | 
|  |  | 
|  | eloop.terminate = 0; | 
|  | out: | 
|  | #ifdef CONFIG_ELOOP_SELECT | 
|  | os_free(rfds); | 
|  | os_free(wfds); | 
|  | os_free(efds); | 
|  | #endif /* CONFIG_ELOOP_SELECT */ | 
|  | return; | 
|  | } | 
|  |  | 
|  |  | 
|  | void eloop_terminate(void) | 
|  | { | 
|  | eloop.terminate = 1; | 
|  | } | 
|  |  | 
|  |  | 
|  | void eloop_destroy(void) | 
|  | { | 
|  | struct eloop_timeout *timeout, *prev; | 
|  | struct os_reltime now; | 
|  |  | 
|  | os_get_reltime(&now); | 
|  | dl_list_for_each_safe(timeout, prev, &eloop.timeout, | 
|  | struct eloop_timeout, list) { | 
|  | int sec, usec; | 
|  | sec = timeout->time.sec - now.sec; | 
|  | usec = timeout->time.usec - now.usec; | 
|  | if (timeout->time.usec < now.usec) { | 
|  | sec--; | 
|  | usec += 1000000; | 
|  | } | 
|  | wpa_printf(MSG_INFO, "ELOOP: remaining timeout: %d.%06d " | 
|  | "eloop_data=%p user_data=%p handler=%p", | 
|  | sec, usec, timeout->eloop_data, timeout->user_data, | 
|  | timeout->handler); | 
|  | wpa_trace_dump_funcname("eloop unregistered timeout handler", | 
|  | timeout->handler); | 
|  | wpa_trace_dump("eloop timeout", timeout); | 
|  | eloop_remove_timeout(timeout); | 
|  | } | 
|  | eloop_sock_table_destroy(&eloop.readers); | 
|  | eloop_sock_table_destroy(&eloop.writers); | 
|  | eloop_sock_table_destroy(&eloop.exceptions); | 
|  | os_free(eloop.signals); | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_POLL | 
|  | os_free(eloop.pollfds); | 
|  | os_free(eloop.pollfds_map); | 
|  | #endif /* CONFIG_ELOOP_POLL */ | 
|  | #if defined(CONFIG_ELOOP_EPOLL) || defined(CONFIG_ELOOP_KQUEUE) | 
|  | os_free(eloop.fd_table); | 
|  | #endif /* CONFIG_ELOOP_EPOLL || CONFIG_ELOOP_KQUEUE */ | 
|  | #ifdef CONFIG_ELOOP_EPOLL | 
|  | os_free(eloop.epoll_events); | 
|  | close(eloop.epollfd); | 
|  | #endif /* CONFIG_ELOOP_EPOLL */ | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | os_free(eloop.kqueue_events); | 
|  | close(eloop.kqueuefd); | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  | } | 
|  |  | 
|  |  | 
|  | int eloop_terminated(void) | 
|  | { | 
|  | return eloop.terminate || eloop.pending_terminate; | 
|  | } | 
|  |  | 
|  |  | 
|  | void eloop_wait_for_read_sock(int sock) | 
|  | { | 
|  | #ifdef CONFIG_ELOOP_POLL | 
|  | struct pollfd pfd; | 
|  |  | 
|  | if (sock < 0) | 
|  | return; | 
|  |  | 
|  | os_memset(&pfd, 0, sizeof(pfd)); | 
|  | pfd.fd = sock; | 
|  | pfd.events = POLLIN; | 
|  |  | 
|  | poll(&pfd, 1, -1); | 
|  | #endif /* CONFIG_ELOOP_POLL */ | 
|  | #if defined(CONFIG_ELOOP_SELECT) || defined(CONFIG_ELOOP_EPOLL) | 
|  | /* | 
|  | * We can use epoll() here. But epoll() requres 4 system calls. | 
|  | * epoll_create1(), epoll_ctl() for ADD, epoll_wait, and close() for | 
|  | * epoll fd. So select() is better for performance here. | 
|  | */ | 
|  | fd_set rfds; | 
|  |  | 
|  | if (sock < 0) | 
|  | return; | 
|  |  | 
|  | FD_ZERO(&rfds); | 
|  | FD_SET(sock, &rfds); | 
|  | select(sock + 1, &rfds, NULL, NULL, NULL); | 
|  | #endif /* defined(CONFIG_ELOOP_SELECT) || defined(CONFIG_ELOOP_EPOLL) */ | 
|  | #ifdef CONFIG_ELOOP_KQUEUE | 
|  | int kfd; | 
|  | struct kevent ke1, ke2; | 
|  |  | 
|  | kfd = kqueue(); | 
|  | if (kfd == -1) | 
|  | return; | 
|  | EV_SET(&ke1, sock, EVFILT_READ, EV_ADD | EV_ONESHOT, 0, 0, 0); | 
|  | kevent(kfd, &ke1, 1, &ke2, 1, NULL); | 
|  | close(kfd); | 
|  | #endif /* CONFIG_ELOOP_KQUEUE */ | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_ELOOP_SELECT | 
|  | #undef CONFIG_ELOOP_SELECT | 
|  | #endif /* CONFIG_ELOOP_SELECT */ |