/* ** GNU Pth - The GNU Portable Threads ** Copyright (c) 1999-2006 Ralf S. Engelschall ** ** This file is part of GNU Pth, a non-preemptive thread scheduling ** library which can be found at http://www.gnu.org/software/pth/. ** ** This library is free software; you can redistribute it and/or ** modify it under the terms of the GNU Lesser General Public ** License as published by the Free Software Foundation; either ** version 2.1 of the License, or (at your option) any later version. ** ** This library is distributed in the hope that it will be useful, ** but WITHOUT ANY WARRANTY; without even the implied warranty of ** MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ** Lesser General Public License for more details. ** ** You should have received a copy of the GNU Lesser General Public ** License along with this library; if not, write to the Free Software ** Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 ** USA, or contact Ralf S. Engelschall . ** ** pth_high.c: Pth high-level replacement functions */ /* ``The difference between a computer industry job and open-source software hacking is about 30 hours a week.'' -- Ralf S. Engelschall */ /* * These functions used by the applications instead of the * regular Unix/POSIX functions. When the regular functions would * block, these variants let only the thread sleep. */ #include "pth_p.h" /* Pth variant of nanosleep(2) */ int pth_nanosleep(const struct timespec *rqtp, struct timespec *rmtp) { pth_time_t until; pth_time_t offset; pth_time_t now; pth_event_t ev; static pth_key_t ev_key = PTH_KEY_INIT; /* consistency checks for POSIX conformance */ if (rqtp == NULL) return pth_error(-1, EFAULT); if (rqtp->tv_nsec < 0 || rqtp->tv_nsec > (1000*1000000)) return pth_error(-1, EINVAL); /* short-circuit */ if (rqtp->tv_sec == 0 && rqtp->tv_nsec == 0) return 0; /* calculate asleep time */ offset = pth_time((long)(rqtp->tv_sec), (long)(rqtp->tv_nsec) / 1000); pth_time_set(&until, PTH_TIME_NOW); pth_time_add(&until, &offset); /* and let thread sleep until this time is elapsed */ if ((ev = pth_event(PTH_EVENT_TIME|PTH_MODE_STATIC, &ev_key, until)) == NULL) return pth_error(-1, errno); pth_wait(ev); /* optionally provide amount of slept time */ if (rmtp != NULL) { pth_time_set(&now, PTH_TIME_NOW); pth_time_sub(&until, &now); rmtp->tv_sec = until.tv_sec; rmtp->tv_nsec = until.tv_usec * 1000; } return 0; } /* Pth variant of usleep(3) */ int pth_usleep(unsigned int usec) { pth_time_t until; pth_time_t offset; pth_event_t ev; static pth_key_t ev_key = PTH_KEY_INIT; /* short-circuit */ if (usec == 0) return 0; /* calculate asleep time */ offset = pth_time((long)(usec / 1000000), (long)(usec % 1000000)); pth_time_set(&until, PTH_TIME_NOW); pth_time_add(&until, &offset); /* and let thread sleep until this time is elapsed */ if ((ev = pth_event(PTH_EVENT_TIME|PTH_MODE_STATIC, &ev_key, until)) == NULL) return pth_error(-1, errno); pth_wait(ev); return 0; } /* Pth variant of sleep(3) */ unsigned int pth_sleep(unsigned int sec) { pth_time_t until; pth_time_t offset; pth_event_t ev; static pth_key_t ev_key = PTH_KEY_INIT; /* consistency check */ if (sec == 0) return 0; /* calculate asleep time */ offset = pth_time(sec, 0); pth_time_set(&until, PTH_TIME_NOW); pth_time_add(&until, &offset); /* and let thread sleep until this time is elapsed */ if ((ev = pth_event(PTH_EVENT_TIME|PTH_MODE_STATIC, &ev_key, until)) == NULL) return sec; pth_wait(ev); return 0; } #if !defined(DISABLE_PIPE) /* Pth variant of POSIX pthread_sigmask(3) */ int pth_sigmask(int how, const sigset_t *set, sigset_t *oset) { int rv; /* change the explicitly remembered signal mask copy for the scheduler */ if (set != NULL) pth_sc(sigprocmask)(how, &(pth_current->mctx.sigs), NULL); /* change the real (per-thread saved/restored) signal mask */ rv = pth_sc(sigprocmask)(how, set, oset); return rv; } /* Pth variant of POSIX sigwait(3) */ int pth_sigwait(const sigset_t *set, int *sigp) { return pth_sigwait_ev(set, sigp, NULL); } /* Pth variant of POSIX sigwait(3) with extra events */ int pth_sigwait_ev(const sigset_t *set, int *sigp, pth_event_t ev_extra) { pth_event_t ev; static pth_key_t ev_key = PTH_KEY_INIT; sigset_t pending; int sig; if (set == NULL || sigp == NULL) return pth_error(EINVAL, EINVAL); /* check whether signal is already pending */ if (sigpending(&pending) < 0) sigemptyset(&pending); for (sig = 1; sig < PTH_NSIG; sig++) { if (sigismember(set, sig) && sigismember(&pending, sig)) { pth_util_sigdelete(sig); *sigp = sig; return 0; } } /* create event and wait on it */ if ((ev = pth_event(PTH_EVENT_SIGS|PTH_MODE_STATIC, &ev_key, set, sigp)) == NULL) return pth_error(errno, errno); if (ev_extra != NULL) pth_event_concat(ev, ev_extra, NULL); pth_wait(ev); if (ev_extra != NULL) { pth_event_isolate(ev); if (pth_event_status(ev) != PTH_STATUS_OCCURRED) return pth_error(EINTR, EINTR); } /* nothing to do, scheduler has already set *sigp for us */ return 0; } #endif /* Pth variant of waitpid(2) */ pid_t pth_waitpid(pid_t wpid, int *status, int options) { pth_event_t ev; static pth_key_t ev_key = PTH_KEY_INIT; pid_t pid; pth_debug2("pth_waitpid: called from thread \"%s\"", pth_current->name); for (;;) { /* do a non-blocking poll for the pid */ while ( (pid = pth_sc(waitpid)(wpid, status, options|WNOHANG)) < 0 && errno == EINTR) ; /* if pid was found or caller requested a polling return immediately */ if (pid == -1 || pid > 0 || (pid == 0 && (options & WNOHANG))) break; /* else wait a little bit */ ev = pth_event(PTH_EVENT_TIME|PTH_MODE_STATIC, &ev_key, pth_timeout(0,250000)); pth_wait(ev); } pth_debug2("pth_waitpid: leave to thread \"%s\"", pth_current->name); return pid; } /* Pth variant of system(3) */ int pth_system(const char *cmd) { struct sigaction sa_ign, sa_int, sa_quit; sigset_t ss_block, ss_old; struct stat sb; pid_t pid; int pstat; /* POSIX calling convention: determine whether the Bourne Shell ("sh") is available on this platform */ if (cmd == NULL) { if (stat(PTH_PATH_BINSH, &sb) == -1) return 0; return 1; } /* temporarily ignore SIGINT and SIGQUIT actions */ sa_ign.sa_handler = SIG_IGN; sigemptyset(&sa_ign.sa_mask); sa_ign.sa_flags = 0; sigaction(SIGINT, &sa_ign, &sa_int); sigaction(SIGQUIT, &sa_ign, &sa_quit); #ifdef HAVE_SIGCHLD /* block SIGCHLD signal */ sigemptyset(&ss_block); sigaddset(&ss_block, SIGCHLD); pth_sc(sigprocmask)(SIG_BLOCK, &ss_block, &ss_old); #endif /* fork the current process */ pstat = -1; switch (pid = pth_fork()) { case -1: /* error */ break; case 0: /* child */ /* restore original signal dispositions and execute the command */ sigaction(SIGINT, &sa_int, NULL); sigaction(SIGQUIT, &sa_quit, NULL); #ifdef HAVE_SIGCHLD pth_sc(sigprocmask)(SIG_SETMASK, &ss_old, NULL); #endif /* stop the Pth scheduling */ pth_scheduler_kill(); /* execute the command through Bourne Shell */ execl(PTH_PATH_BINSH, "sh", "-c", cmd, (char *)NULL); /* POSIX compliant return in case execution failed */ exit(127); default: /* parent */ /* wait until child process terminates */ pid = pth_waitpid(pid, &pstat, 0); break; } /* restore original signal dispositions and execute the command */ sigaction(SIGINT, &sa_int, NULL); sigaction(SIGQUIT, &sa_quit, NULL); #ifdef HAVE_SIGCHLD pth_sc(sigprocmask)(SIG_SETMASK, &ss_old, NULL); #endif /* return error or child process result code */ return (pid == -1 ? -1 : pstat); } /* Pth variant of select(2) */ int pth_select(int nfds, fd_set *rfds, fd_set *wfds, fd_set *efds, struct timeval *timeout) { return pth_select_ev(nfds, rfds, wfds, efds, timeout, NULL); } /* Pth variant of select(2) with extra events */ int pth_select_ev(int nfd, fd_set *rfds, fd_set *wfds, fd_set *efds, struct timeval *timeout, pth_event_t ev_extra) { struct timeval delay; pth_event_t ev; pth_event_t ev_select; pth_event_t ev_timeout; static pth_key_t ev_key_select = PTH_KEY_INIT; static pth_key_t ev_key_timeout = PTH_KEY_INIT; fd_set rspare, wspare, espare; fd_set *rtmp, *wtmp, *etmp; int selected; int rc; pth_implicit_init(); pth_debug2("pth_select_ev: called from thread \"%s\"", pth_current->name); /* POSIX.1-2001/SUSv3 compliance */ if (nfd < 0 || nfd > FD_SETSIZE) return pth_error(-1, EINVAL); if (timeout != NULL) { if ( timeout->tv_sec < 0 || timeout->tv_usec < 0 || timeout->tv_usec >= 1000000 /* a full second */) return pth_error(-1, EINVAL); if (timeout->tv_sec > 31*24*60*60) timeout->tv_sec = 31*24*60*60; } /* first deal with the special situation of a plain microsecond delay */ if (nfd == 0 && rfds == NULL && wfds == NULL && efds == NULL && timeout != NULL) { if (timeout->tv_sec == 0 && timeout->tv_usec <= 10000 /* 1/100 second */) { /* very small delays are acceptable to be performed directly */ while ( pth_sc(select)(0, NULL, NULL, NULL, timeout) < 0 && errno == EINTR) ; } else { /* larger delays have to go through the scheduler */ ev = pth_event(PTH_EVENT_TIME|PTH_MODE_STATIC, &ev_key_timeout, pth_timeout(timeout->tv_sec, timeout->tv_usec)); if (ev_extra != NULL) pth_event_concat(ev, ev_extra, NULL); pth_wait(ev); if (ev_extra != NULL) { pth_event_isolate(ev); if (pth_event_status(ev) != PTH_STATUS_OCCURRED) return pth_error(-1, EINTR); } } /* POSIX.1-2001/SUSv3 compliance */ if (rfds != NULL) FD_ZERO(rfds); if (wfds != NULL) FD_ZERO(wfds); if (efds != NULL) FD_ZERO(efds); return 0; } /* now directly poll filedescriptor sets to avoid unnecessary (and resource consuming because of context switches, etc) event handling through the scheduler. We've to be carefully here, because not all platforms guaranty us that the sets are unmodified if an error or timeout occurred. */ delay.tv_sec = 0; delay.tv_usec = 0; rtmp = NULL; if (rfds != NULL) { memcpy(&rspare, rfds, sizeof(fd_set)); rtmp = &rspare; } wtmp = NULL; if (wfds != NULL) { memcpy(&wspare, wfds, sizeof(fd_set)); wtmp = &wspare; } etmp = NULL; if (efds != NULL) { memcpy(&espare, efds, sizeof(fd_set)); etmp = &espare; } while ((rc = pth_sc(select)(nfd, rtmp, wtmp, etmp, &delay)) < 0 && errno == EINTR) ; if (rc < 0) /* pass-through immediate error */ return pth_error(-1, errno); else if ( rc > 0 || ( rc == 0 && timeout != NULL && pth_time_cmp(timeout, PTH_TIME_ZERO) == 0)) { /* pass-through immediate success */ if (rfds != NULL) memcpy(rfds, &rspare, sizeof(fd_set)); if (wfds != NULL) memcpy(wfds, &wspare, sizeof(fd_set)); if (efds != NULL) memcpy(efds, &espare, sizeof(fd_set)); return rc; } /* suspend current thread until one filedescriptor is ready or the timeout occurred */ rc = -1; ev = ev_select = pth_event(PTH_EVENT_SELECT|PTH_MODE_STATIC, &ev_key_select, &rc, nfd, rfds, wfds, efds); ev_timeout = NULL; if (timeout != NULL) { ev_timeout = pth_event(PTH_EVENT_TIME|PTH_MODE_STATIC, &ev_key_timeout, pth_timeout(timeout->tv_sec, timeout->tv_usec)); pth_event_concat(ev, ev_timeout, NULL); } if (ev_extra != NULL) pth_event_concat(ev, ev_extra, NULL); pth_wait(ev); if (ev_extra != NULL) pth_event_isolate(ev_extra); if (timeout != NULL) pth_event_isolate(ev_timeout); /* select return code semantics */ if (pth_event_status(ev_select) == PTH_STATUS_FAILED) return pth_error(-1, EBADF); selected = FALSE; if (pth_event_status(ev_select) == PTH_STATUS_OCCURRED) selected = TRUE; if ( timeout != NULL && pth_event_status(ev_timeout) == PTH_STATUS_OCCURRED) { selected = TRUE; /* POSIX.1-2001/SUSv3 compliance */ if (rfds != NULL) FD_ZERO(rfds); if (wfds != NULL) FD_ZERO(wfds); if (efds != NULL) FD_ZERO(efds); rc = 0; } if (ev_extra != NULL && !selected) return pth_error(-1, EINTR); return rc; } /* Pth variant of pth_pselect(2) */ int pth_pselect(int nfds, fd_set *rfds, fd_set *wfds, fd_set *efds, const struct timespec *ts, const sigset_t *mask) { sigset_t omask; struct timeval tv; struct timeval *tvp; int rv; /* convert timeout */ if (ts != NULL) { tv.tv_sec = ts->tv_sec; tv.tv_usec = ts->tv_nsec / 1000; tvp = &tv; } else tvp = NULL; /* optionally set signal mask */ if (mask != NULL) if (pth_sc(sigprocmask)(SIG_SETMASK, mask, &omask) < 0) return pth_error(-1, errno); rv = pth_select(nfds, rfds, wfds, efds, tvp); /* optionally set signal mask */ if (mask != NULL) pth_shield { pth_sc(sigprocmask)(SIG_SETMASK, &omask, NULL); } return rv; } /* Pth variant of poll(2) */ int pth_poll(struct pollfd *pfd, nfds_t nfd, int timeout) { return pth_poll_ev(pfd, nfd, timeout, NULL); } /* Pth variant of poll(2) with extra events: NOTICE: THIS HAS TO BE BASED ON pth_select(2) BECAUSE INTERNALLY THE SCHEDULER IS ONLY select(2) BASED!! */ int pth_poll_ev(struct pollfd *pfd, nfds_t nfd, int timeout, pth_event_t ev_extra) { fd_set rfds, wfds, efds, xfds; struct timeval tv, *ptv; int maxfd, rc, n; unsigned int i; char data[64]; pth_implicit_init(); pth_debug2("pth_poll_ev: called from thread \"%s\"", pth_current->name); /* argument sanity checks */ if (pfd == NULL) return pth_error(-1, EFAULT); if (nfd < 0 || nfd > FD_SETSIZE) return pth_error(-1, EINVAL); /* convert timeout number into a timeval structure */ ptv = &tv; if (timeout == 0) { /* return immediately */ ptv->tv_sec = 0; ptv->tv_usec = 0; } else if (timeout == INFTIM /* (-1) */) { /* wait forever */ ptv = NULL; } else if (timeout > 0) { /* return after timeout */ ptv->tv_sec = (timeout / 1000); ptv->tv_usec = (timeout % 1000) * 1000; } else return pth_error(-1, EINVAL); /* create fd sets and determine max fd */ maxfd = -1; FD_ZERO(&rfds); FD_ZERO(&wfds); FD_ZERO(&efds); FD_ZERO(&xfds); for (i = 0; i < nfd; i++) { /* convert into fd_sets but remember that BSD select(2) says "the only exceptional condition detectable is out-of-band data received on a socket", hence we push POLLWRBAND events onto wfds instead of efds. Additionally, remember invalid filedescriptors in an extra fd_set xfds. */ if (!pth_util_fd_valid(pfd[i].fd)) { FD_SET(pfd[i].fd, &xfds); continue; } if (pfd[i].events & (POLLIN|POLLRDNORM)) FD_SET(pfd[i].fd, &rfds); if (pfd[i].events & (POLLOUT|POLLWRNORM|POLLWRBAND)) FD_SET(pfd[i].fd, &wfds); if (pfd[i].events & (POLLPRI|POLLRDBAND)) FD_SET(pfd[i].fd, &efds); if ( pfd[i].fd >= maxfd && (pfd[i].events & (POLLIN|POLLOUT|POLLPRI| POLLRDNORM|POLLRDBAND| POLLWRNORM|POLLWRBAND))) maxfd = pfd[i].fd; } /* examine fd sets with pth_select(3) */ rc = -1; if (maxfd != -1) { rc = pth_select_ev(maxfd+1, &rfds, &wfds, &efds, ptv, ev_extra); if (rc < 0) return pth_error(-1, errno); else if (rc == 0) return 0; } /* POSIX.1-2001/SUSv3 compliant result establishment */ n = 0; for (i = 0; i < nfd; i++) { pfd[i].revents = 0; if (FD_ISSET(pfd[i].fd, &xfds)) { if (pfd[i].fd >= 0) { pfd[i].revents |= POLLNVAL; n++; } continue; } if (maxfd == -1) continue; if (FD_ISSET(pfd[i].fd, &rfds)) { if (pfd[i].events & POLLIN) pfd[i].revents |= POLLIN; if (pfd[i].events & POLLRDNORM) pfd[i].revents |= POLLRDNORM; n++; /* support for POLLHUP */ if ( recv(pfd[i].fd, data, sizeof(data), MSG_PEEK) == -1 && ( errno == ESHUTDOWN || errno == ECONNRESET || errno == ECONNABORTED || errno == ENETRESET )) { pfd[i].revents &= ~(POLLIN); pfd[i].revents &= ~(POLLRDNORM); pfd[i].revents |= POLLHUP; } } else if (FD_ISSET(pfd[i].fd, &wfds)) { if (pfd[i].events & POLLOUT) pfd[i].revents |= POLLOUT; if (pfd[i].events & POLLWRNORM) pfd[i].revents |= POLLWRNORM; if (pfd[i].events & POLLWRBAND) pfd[i].revents |= POLLWRBAND; n++; } else if (FD_ISSET(pfd[i].fd, &efds)) { if (pfd[i].events & POLLPRI) pfd[i].revents |= POLLPRI; if (pfd[i].events & POLLRDBAND) pfd[i].revents |= POLLRDBAND; n++; } } return n; } /* Pth variant of connect(2) */ int pth_connect(int s, const struct sockaddr *addr, socklen_t addrlen) { return pth_connect_ev(s, addr, addrlen, NULL); } /* Pth variant of connect(2) with extra events */ int pth_connect_ev(int s, const struct sockaddr *addr, socklen_t addrlen, pth_event_t ev_extra) { pth_event_t ev; static pth_key_t ev_key = PTH_KEY_INIT; int rv, err; socklen_t errlen; int fdmode; pth_implicit_init(); pth_debug2("pth_connect_ev: enter from thread \"%s\"", pth_current->name); /* POSIX compliance */ if (!pth_util_fd_valid(s)) return pth_error(-1, EBADF); /* force filedescriptor into non-blocking mode */ if ((fdmode = pth_fdmode(s, PTH_FDMODE_NONBLOCK)) == PTH_FDMODE_ERROR) return pth_error(-1, EBADF); /* try to connect */ while ( (rv = pth_sc(connect)(s, (struct sockaddr *)addr, addrlen)) == -1 && errno == EINTR) ; /* restore filedescriptor mode */ pth_shield { pth_fdmode(s, fdmode); } /* if it is still on progress wait until socket is really writeable */ if (rv == -1 && errno == EINPROGRESS && fdmode != PTH_FDMODE_NONBLOCK) { if ((ev = pth_event(PTH_EVENT_FD|PTH_UNTIL_FD_WRITEABLE|PTH_MODE_STATIC, &ev_key, s)) == NULL) return pth_error(-1, errno); if (ev_extra != NULL) pth_event_concat(ev, ev_extra, NULL); pth_wait(ev); if (ev_extra != NULL) { pth_event_isolate(ev); if (pth_event_status(ev) != PTH_STATUS_OCCURRED) return pth_error(-1, EINTR); } errlen = sizeof(err); if (getsockopt(s, SOL_SOCKET, SO_ERROR, (void *)&err, &errlen) == -1) return -1; if (err == 0) return 0; return pth_error(rv, err); } pth_debug2("pth_connect_ev: leave to thread \"%s\"", pth_current->name); return rv; } /* Pth variant of accept(2) */ int pth_accept(int s, struct sockaddr *addr, socklen_t *addrlen) { return pth_accept_ev(s, addr, addrlen, NULL); } /* Pth variant of accept(2) with extra events */ int pth_accept_ev(int s, struct sockaddr *addr, socklen_t *addrlen, pth_event_t ev_extra) { pth_event_t ev; static pth_key_t ev_key = PTH_KEY_INIT; int fdmode; int rv; pth_implicit_init(); pth_debug2("pth_accept_ev: enter from thread \"%s\"", pth_current->name); /* POSIX compliance */ if (!pth_util_fd_valid(s)) return pth_error(-1, EBADF); /* force filedescriptor into non-blocking mode */ if ((fdmode = pth_fdmode(s, PTH_FDMODE_NONBLOCK)) == PTH_FDMODE_ERROR) return pth_error(-1, EBADF); /* poll socket via accept */ ev = NULL; while ((rv = pth_sc(accept)(s, addr, addrlen)) == -1 && (errno == EAGAIN || errno == EWOULDBLOCK) && fdmode != PTH_FDMODE_NONBLOCK) { /* do lazy event allocation */ if (ev == NULL) { if ((ev = pth_event(PTH_EVENT_FD|PTH_UNTIL_FD_READABLE|PTH_MODE_STATIC, &ev_key, s)) == NULL) return pth_error(-1, errno); if (ev_extra != NULL) pth_event_concat(ev, ev_extra, NULL); } /* wait until accept has a chance */ pth_wait(ev); /* check for the extra events */ if (ev_extra != NULL) { pth_event_isolate(ev); if (pth_event_status(ev) != PTH_STATUS_OCCURRED) { pth_fdmode(s, fdmode); return pth_error(-1, EINTR); } } } /* restore filedescriptor mode */ pth_shield { pth_fdmode(s, fdmode); if (rv != -1) pth_fdmode(rv, fdmode); } pth_debug2("pth_accept_ev: leave to thread \"%s\"", pth_current->name); return rv; } /* Pth variant of read(2) */ ssize_t pth_read(int fd, void *buf, size_t nbytes) { return pth_read_ev(fd, buf, nbytes, NULL); } /* Pth variant of read(2) with extra event(s) */ ssize_t pth_read_ev(int fd, void *buf, size_t nbytes, pth_event_t ev_extra) { struct timeval delay; pth_event_t ev; static pth_key_t ev_key = PTH_KEY_INIT; fd_set fds; int fdmode; int n; pth_implicit_init(); pth_debug2("pth_read_ev: enter from thread \"%s\"", pth_current->name); /* POSIX compliance */ if (nbytes == 0) return 0; if (!pth_util_fd_valid(fd)) return pth_error(-1, EBADF); /* check mode of filedescriptor */ if ((fdmode = pth_fdmode(fd, PTH_FDMODE_POLL)) == PTH_FDMODE_ERROR) return pth_error(-1, EBADF); /* poll filedescriptor if not already in non-blocking operation */ if (fdmode == PTH_FDMODE_BLOCK) { /* now directly poll filedescriptor for readability to avoid unneccessary (and resource consuming because of context switches, etc) event handling through the scheduler */ FD_ZERO(&fds); FD_SET(fd, &fds); delay.tv_sec = 0; delay.tv_usec = 0; while ((n = pth_sc(select)(fd+1, &fds, NULL, NULL, &delay)) < 0 && errno == EINTR) ; if (n < 0 && (errno == EINVAL || errno == EBADF)) return pth_error(-1, errno); /* if filedescriptor is still not readable, let thread sleep until it is or the extra event occurs */ if (n == 0) { ev = pth_event(PTH_EVENT_FD|PTH_UNTIL_FD_READABLE|PTH_MODE_STATIC, &ev_key, fd); if (ev_extra != NULL) pth_event_concat(ev, ev_extra, NULL); n = pth_wait(ev); if (ev_extra != NULL) { pth_event_isolate(ev); if (pth_event_status(ev) != PTH_STATUS_OCCURRED) return pth_error(-1, EINTR); } } } /* Now perform the actual read. We're now guarrantied to not block, either because we were already in non-blocking mode or we determined above by polling that the next read(2) call will not block. But keep in mind, that only 1 next read(2) call is guarrantied to not block (except for the EINTR situation). */ while ((n = pth_sc(read)(fd, buf, nbytes)) < 0 && errno == EINTR) ; pth_debug2("pth_read_ev: leave to thread \"%s\"", pth_current->name); return n; } /* Pth variant of write(2) */ ssize_t pth_write(int fd, const void *buf, size_t nbytes) { return pth_write_ev(fd, buf, nbytes, NULL); } /* Pth variant of write(2) with extra event(s) */ ssize_t pth_write_ev(int fd, const void *buf, size_t nbytes, pth_event_t ev_extra) { struct timeval delay; pth_event_t ev; static pth_key_t ev_key = PTH_KEY_INIT; fd_set fds; int fdmode; ssize_t rv; ssize_t s; int n; pth_implicit_init(); pth_debug2("pth_write_ev: enter from thread \"%s\"", pth_current->name); /* POSIX compliance */ if (nbytes == 0) return 0; if (!pth_util_fd_valid(fd)) return pth_error(-1, EBADF); /* force filedescriptor into non-blocking mode */ if ((fdmode = pth_fdmode(fd, PTH_FDMODE_NONBLOCK)) == PTH_FDMODE_ERROR) return pth_error(-1, EBADF); /* poll filedescriptor if not already in non-blocking operation */ if (fdmode != PTH_FDMODE_NONBLOCK) { /* now directly poll filedescriptor for writeability to avoid unneccessary (and resource consuming because of context switches, etc) event handling through the scheduler */ FD_ZERO(&fds); FD_SET(fd, &fds); delay.tv_sec = 0; delay.tv_usec = 0; while ((n = pth_sc(select)(fd+1, NULL, &fds, NULL, &delay)) < 0 && errno == EINTR) ; if (n < 0 && (errno == EINVAL || errno == EBADF)) return pth_error(-1, errno); rv = 0; for (;;) { /* if filedescriptor is still not writeable, let thread sleep until it is or event occurs */ if (n < 1) { ev = pth_event(PTH_EVENT_FD|PTH_UNTIL_FD_WRITEABLE|PTH_MODE_STATIC, &ev_key, fd); if (ev_extra != NULL) pth_event_concat(ev, ev_extra, NULL); pth_wait(ev); if (ev_extra != NULL) { pth_event_isolate(ev); if (pth_event_status(ev) != PTH_STATUS_OCCURRED) { pth_fdmode(fd, fdmode); return pth_error(-1, EINTR); } } } /* now perform the actual write operation */ while ((s = pth_sc(write)(fd, buf, nbytes)) < 0 && errno == EINTR) ; if (s > 0) rv += s; /* although we're physically now in non-blocking mode, iterate unless all data is written or an error occurs, because we've to mimic the usual blocking I/O behaviour of write(2). */ if (s > 0 && s < (ssize_t)nbytes) { nbytes -= s; buf = (void *)((char *)buf + s); n = 0; continue; } /* pass error to caller, but not for partial writes (rv > 0) */ if (s < 0 && rv == 0) rv = -1; /* stop looping */ break; } } else { /* just perform the actual write operation */ while ((rv = pth_sc(write)(fd, buf, nbytes)) < 0 && errno == EINTR) ; } /* restore filedescriptor mode */ pth_shield { pth_fdmode(fd, fdmode); } pth_debug2("pth_write_ev: leave to thread \"%s\"", pth_current->name); return rv; } /* Pth variant of readv(2) */ ssize_t pth_readv(int fd, const struct iovec *iov, int iovcnt) { return pth_readv_ev(fd, iov, iovcnt, NULL); } /* Pth variant of readv(2) with extra event(s) */ ssize_t pth_readv_ev(int fd, const struct iovec *iov, int iovcnt, pth_event_t ev_extra) { struct timeval delay; pth_event_t ev; static pth_key_t ev_key = PTH_KEY_INIT; fd_set fds; int fdmode; int n; pth_implicit_init(); pth_debug2("pth_readv_ev: enter from thread \"%s\"", pth_current->name); /* POSIX compliance */ if (iovcnt <= 0 || iovcnt > UIO_MAXIOV) return pth_error(-1, EINVAL); if (!pth_util_fd_valid(fd)) return pth_error(-1, EBADF); /* check mode of filedescriptor */ if ((fdmode = pth_fdmode(fd, PTH_FDMODE_POLL)) == PTH_FDMODE_ERROR) return pth_error(-1, EBADF); /* poll filedescriptor if not already in non-blocking operation */ if (fdmode == PTH_FDMODE_BLOCK) { /* first directly poll filedescriptor for readability to avoid unneccessary (and resource consuming because of context switches, etc) event handling through the scheduler */ FD_ZERO(&fds); FD_SET(fd, &fds); delay.tv_sec = 0; delay.tv_usec = 0; while ((n = pth_sc(select)(fd+1, &fds, NULL, NULL, &delay)) < 0 && errno == EINTR) ; /* if filedescriptor is still not readable, let thread sleep until it is or event occurs */ if (n < 1) { ev = pth_event(PTH_EVENT_FD|PTH_UNTIL_FD_READABLE|PTH_MODE_STATIC, &ev_key, fd); if (ev_extra != NULL) pth_event_concat(ev, ev_extra, NULL); n = pth_wait(ev); if (ev_extra != NULL) { pth_event_isolate(ev); if (pth_event_status(ev) != PTH_STATUS_OCCURRED) return pth_error(-1, EINTR); } } } /* Now perform the actual read. We're now guarrantied to not block, either because we were already in non-blocking mode or we determined above by polling that the next read(2) call will not block. But keep in mind, that only 1 next read(2) call is guarrantied to not block (except for the EINTR situation). */ #if PTH_FAKE_RWV while ((n = pth_readv_faked(fd, iov, iovcnt)) < 0 && errno == EINTR) ; #elif PTH_WATT_RWV while ((n = pth_sc(readv_s)(fd, iov, iovcnt)) < 0 && errno == EINTR) ; #else while ((n = pth_sc(readv)(fd, iov, iovcnt)) < 0 && errno == EINTR) ; #endif pth_debug2("pth_readv_ev: leave to thread \"%s\"", pth_current->name); return n; } #if PTH_FAKE_RWV /* A faked version of readv(2) */ intern ssize_t pth_readv_faked(int fd, const struct iovec *iov, int iovcnt) { char *buffer; size_t bytes, copy, rv; int i; /* determine total number of bytes to read */ bytes = 0; for (i = 0; i < iovcnt; i++) { if (iov[i].iov_len <= 0) return pth_error((ssize_t)(-1), EINVAL); bytes += iov[i].iov_len; } if (bytes <= 0) return pth_error((ssize_t)(-1), EINVAL); /* allocate a temporary buffer */ if ((buffer = (char *)malloc(bytes)) == NULL) return (ssize_t)(-1); /* read data into temporary buffer (caller guarrantied us to not block) */ rv = pth_sc(read)(fd, buffer, bytes); /* scatter read data into callers vector */ if (rv > 0) { bytes = rv; for (i = 0; i < iovcnt; i++) { copy = pth_util_min(iov[i].iov_len, bytes); memcpy(iov[i].iov_base, buffer, copy); buffer += copy; bytes -= copy; if (bytes <= 0) break; } } /* remove the temporary buffer */ pth_shield { free(buffer); } /* return number of read bytes */ return(rv); } #endif /* Pth variant of writev(2) */ ssize_t pth_writev(int fd, const struct iovec *iov, int iovcnt) { return pth_writev_ev(fd, iov, iovcnt, NULL); } /* Pth variant of writev(2) with extra event(s) */ ssize_t pth_writev_ev(int fd, const struct iovec *iov, int iovcnt, pth_event_t ev_extra) { struct timeval delay; pth_event_t ev; static pth_key_t ev_key = PTH_KEY_INIT; fd_set fds; int fdmode; struct iovec *liov; int liovcnt; size_t nbytes; ssize_t rv; ssize_t s; int n; struct iovec tiov_stack[32]; struct iovec *tiov; int tiovcnt; pth_implicit_init(); pth_debug2("pth_writev_ev: enter from thread \"%s\"", pth_current->name); /* POSIX compliance */ if (iovcnt <= 0 || iovcnt > UIO_MAXIOV) return pth_error(-1, EINVAL); if (!pth_util_fd_valid(fd)) return pth_error(-1, EBADF); /* force filedescriptor into non-blocking mode */ if ((fdmode = pth_fdmode(fd, PTH_FDMODE_NONBLOCK)) == PTH_FDMODE_ERROR) return pth_error(-1, EBADF); /* poll filedescriptor if not already in non-blocking operation */ if (fdmode != PTH_FDMODE_NONBLOCK) { /* provide temporary iovec structure */ if (iovcnt > sizeof(tiov_stack)) { tiovcnt = (sizeof(struct iovec) * UIO_MAXIOV); if ((tiov = (struct iovec *)malloc(tiovcnt)) == NULL) return pth_error(-1, errno); } else { tiovcnt = sizeof(tiov_stack); tiov = tiov_stack; } /* init return value and number of bytes to write */ rv = 0; nbytes = pth_writev_iov_bytes(iov, iovcnt); /* init local iovec structure */ liov = NULL; liovcnt = 0; pth_writev_iov_advance(iov, iovcnt, 0, &liov, &liovcnt, tiov, tiovcnt); /* first directly poll filedescriptor for writeability to avoid unneccessary (and resource consuming because of context switches, etc) event handling through the scheduler */ FD_ZERO(&fds); FD_SET(fd, &fds); delay.tv_sec = 0; delay.tv_usec = 0; while ((n = pth_sc(select)(fd+1, NULL, &fds, NULL, &delay)) < 0 && errno == EINTR) ; for (;;) { /* if filedescriptor is still not writeable, let thread sleep until it is or event occurs */ if (n < 1) { ev = pth_event(PTH_EVENT_FD|PTH_UNTIL_FD_WRITEABLE|PTH_MODE_STATIC, &ev_key, fd); if (ev_extra != NULL) pth_event_concat(ev, ev_extra, NULL); pth_wait(ev); if (ev_extra != NULL) { pth_event_isolate(ev); if (pth_event_status(ev) != PTH_STATUS_OCCURRED) { pth_fdmode(fd, fdmode); if (iovcnt > sizeof(tiov_stack)) free(tiov); return pth_error(-1, EINTR); } } } /* now perform the actual write operation */ #if PTH_FAKE_RWV while ((s = pth_writev_faked(fd, liov, liovcnt)) < 0 && errno == EINTR) ; #elif PTH_WATT_RWV while ((s = pth_sc(writev_s)(fd, liov, liovcnt)) < 0 && errno == EINTR) ; #else while ((s = pth_sc(writev)(fd, liov, liovcnt)) < 0 && errno == EINTR) ; #endif if (s > 0) rv += s; /* although we're physically now in non-blocking mode, iterate unless all data is written or an error occurs, because we've to mimic the usual blocking I/O behaviour of writev(2) */ if (s > 0 && s < (ssize_t)nbytes) { nbytes -= s; pth_writev_iov_advance(iov, iovcnt, s, &liov, &liovcnt, tiov, tiovcnt); n = 0; continue; } /* pass error to caller, but not for partial writes (rv > 0) */ if (s < 0 && rv == 0) rv = -1; /* stop looping */ break; } /* cleanup */ if (iovcnt > sizeof(tiov_stack)) free(tiov); } else { /* just perform the actual write operation */ #if PTH_FAKE_RWV while ((rv = pth_writev_faked(fd, iov, iovcnt)) < 0 && errno == EINTR) ; #elif PTH_WATT_RWV while ((rv = pth_sc(writev_s)(fd, iov, iovcnt)) < 0 && errno == EINTR) ; #else while ((rv = pth_sc(writev)(fd, iov, iovcnt)) < 0 && errno == EINTR) ; #endif } /* restore filedescriptor mode */ pth_shield { pth_fdmode(fd, fdmode); } pth_debug2("pth_writev_ev: leave to thread \"%s\"", pth_current->name); return rv; } /* calculate number of bytes in a struct iovec */ intern ssize_t pth_writev_iov_bytes(const struct iovec *iov, int iovcnt) { ssize_t bytes; int i; bytes = 0; for (i = 0; i < iovcnt; i++) { if (iov[i].iov_len <= 0) continue; bytes += iov[i].iov_len; } return bytes; } /* advance the virtual pointer of a struct iov */ intern void pth_writev_iov_advance(const struct iovec *riov, int riovcnt, size_t advance, struct iovec **liov, int *liovcnt, struct iovec *tiov, int tiovcnt) { int i; if (*liov == NULL && *liovcnt == 0) { /* initialize with real (const) structure on first step */ *liov = (struct iovec *)riov; *liovcnt = riovcnt; } if (advance > 0) { if (*liov == riov && *liovcnt == riovcnt) { /* reinitialize with a copy to be able to adjust it */ *liov = &tiov[0]; for (i = 0; i < riovcnt; i++) { tiov[i].iov_base = riov[i].iov_base; tiov[i].iov_len = riov[i].iov_len; } } /* advance the virtual pointer */ while (*liovcnt > 0 && advance > 0) { if ((*liov)->iov_len > advance) { (*liov)->iov_base = (char *)((*liov)->iov_base) + advance; (*liov)->iov_len -= advance; break; } else { advance -= (*liov)->iov_len; (*liovcnt)--; (*liov)++; } } } return; } #if PTH_FAKE_RWV /* A faked version of writev(2) */ intern ssize_t pth_writev_faked(int fd, const struct iovec *iov, int iovcnt) { char *buffer, *cp; size_t bytes, to_copy, copy, rv; int i; /* determine total number of bytes to write */ bytes = 0; for (i = 0; i < iovcnt; i++) { if (iov[i].iov_len <= 0) return pth_error((ssize_t)(-1), EINVAL); bytes += iov[i].iov_len; } if (bytes <= 0) return pth_error((ssize_t)(-1), EINVAL); /* allocate a temporary buffer to hold the data */ if ((buffer = (char *)malloc(bytes)) == NULL) return (ssize_t)(-1); /* concatenate the data from callers vector into buffer */ to_copy = bytes; cp = buffer; for (i = 0; i < iovcnt; i++) { copy = pth_util_min(iov[i].iov_len, to_copy); memcpy(cp, iov[i].iov_base, copy); to_copy -= copy; if (to_copy <= 0) break; } /* write continuous chunck of data (caller guarrantied us to not block) */ rv = pth_sc(write)(fd, buffer, bytes); /* remove the temporary buffer */ pth_shield { free(buffer); } return(rv); } #endif /* Pth variant of POSIX pread(3) */ ssize_t pth_pread(int fd, void *buf, size_t nbytes, off_t offset) { static pth_mutex_t mutex = PTH_MUTEX_INIT; off_t old_offset; ssize_t rc; /* protect us: pth_read can yield! */ if (!pth_mutex_acquire(&mutex, FALSE, NULL)) return (-1); /* remember current offset */ if ((old_offset = lseek(fd, 0, SEEK_CUR)) == (off_t)(-1)) { pth_mutex_release(&mutex); return (-1); } /* seek to requested offset */ if (lseek(fd, offset, SEEK_SET) == (off_t)(-1)) { pth_mutex_release(&mutex); return (-1); } /* perform the read operation */ rc = pth_read(fd, buf, nbytes); /* restore the old offset situation */ pth_shield { lseek(fd, old_offset, SEEK_SET); } /* unprotect and return result of read */ pth_mutex_release(&mutex); return rc; } /* Pth variant of POSIX pwrite(3) */ ssize_t pth_pwrite(int fd, const void *buf, size_t nbytes, off_t offset) { static pth_mutex_t mutex = PTH_MUTEX_INIT; off_t old_offset; ssize_t rc; /* protect us: pth_write can yield! */ if (!pth_mutex_acquire(&mutex, FALSE, NULL)) return (-1); /* remember current offset */ if ((old_offset = lseek(fd, 0, SEEK_CUR)) == (off_t)(-1)) { pth_mutex_release(&mutex); return (-1); } /* seek to requested offset */ if (lseek(fd, offset, SEEK_SET) == (off_t)(-1)) { pth_mutex_release(&mutex); return (-1); } /* perform the write operation */ rc = pth_write(fd, buf, nbytes); /* restore the old offset situation */ pth_shield { lseek(fd, old_offset, SEEK_SET); } /* unprotect and return result of write */ pth_mutex_release(&mutex); return rc; } /* Pth variant of SUSv2 recv(2) */ ssize_t pth_recv(int s, void *buf, size_t len, int flags) { return pth_recv_ev(s, buf, len, flags, NULL); } /* Pth variant of SUSv2 recv(2) with extra event(s) */ ssize_t pth_recv_ev(int s, void *buf, size_t len, int flags, pth_event_t ev) { return pth_recvfrom_ev(s, buf, len, flags, NULL, 0, ev); } /* Pth variant of SUSv2 recvfrom(2) */ ssize_t pth_recvfrom(int s, void *buf, size_t len, int flags, struct sockaddr *from, socklen_t *fromlen) { return pth_recvfrom_ev(s, buf, len, flags, from, fromlen, NULL); } /* Pth variant of SUSv2 recvfrom(2) with extra event(s) */ ssize_t pth_recvfrom_ev(int fd, void *buf, size_t nbytes, int flags, struct sockaddr *from, socklen_t *fromlen, pth_event_t ev_extra) { struct timeval delay; pth_event_t ev; static pth_key_t ev_key = PTH_KEY_INIT; fd_set fds; int fdmode; int n; pth_implicit_init(); pth_debug2("pth_recvfrom_ev: enter from thread \"%s\"", pth_current->name); /* POSIX compliance */ if (nbytes == 0) return 0; if (!pth_util_fd_valid(fd)) return pth_error(-1, EBADF); /* check mode of filedescriptor */ if ((fdmode = pth_fdmode(fd, PTH_FDMODE_POLL)) == PTH_FDMODE_ERROR) return pth_error(-1, EBADF); /* poll filedescriptor if not already in non-blocking operation */ if (fdmode == PTH_FDMODE_BLOCK) { /* now directly poll filedescriptor for readability to avoid unneccessary (and resource consuming because of context switches, etc) event handling through the scheduler */ if (!pth_util_fd_valid(fd)) return pth_error(-1, EBADF); FD_ZERO(&fds); FD_SET(fd, &fds); delay.tv_sec = 0; delay.tv_usec = 0; while ((n = pth_sc(select)(fd+1, &fds, NULL, NULL, &delay)) < 0 && errno == EINTR) ; if (n < 0 && (errno == EINVAL || errno == EBADF)) return pth_error(-1, errno); /* if filedescriptor is still not readable, let thread sleep until it is or the extra event occurs */ if (n == 0) { ev = pth_event(PTH_EVENT_FD|PTH_UNTIL_FD_READABLE|PTH_MODE_STATIC, &ev_key, fd); if (ev_extra != NULL) pth_event_concat(ev, ev_extra, NULL); n = pth_wait(ev); if (ev_extra != NULL) { pth_event_isolate(ev); if (pth_event_status(ev) != PTH_STATUS_OCCURRED) return pth_error(-1, EINTR); } } } /* now perform the actual read. We're now guarrantied to not block, either because we were already in non-blocking mode or we determined above by polling that the next recvfrom(2) call will not block. But keep in mind, that only 1 next recvfrom(2) call is guarrantied to not block (except for the EINTR situation). */ while ((n = pth_sc(recvfrom)(fd, buf, nbytes, flags, from, fromlen)) < 0 && errno == EINTR) ; pth_debug2("pth_recvfrom_ev: leave to thread \"%s\"", pth_current->name); return n; } /* Pth variant of SUSv2 send(2) */ ssize_t pth_send(int s, const void *buf, size_t len, int flags) { return pth_send_ev(s, buf, len, flags, NULL); } /* Pth variant of SUSv2 send(2) with extra event(s) */ ssize_t pth_send_ev(int s, const void *buf, size_t len, int flags, pth_event_t ev) { return pth_sendto_ev(s, buf, len, flags, NULL, 0, ev); } /* Pth variant of SUSv2 sendto(2) */ ssize_t pth_sendto(int s, const void *buf, size_t len, int flags, const struct sockaddr *to, socklen_t tolen) { return pth_sendto_ev(s, buf, len, flags, to, tolen, NULL); } /* Pth variant of SUSv2 sendto(2) with extra event(s) */ ssize_t pth_sendto_ev(int fd, const void *buf, size_t nbytes, int flags, const struct sockaddr *to, socklen_t tolen, pth_event_t ev_extra) { struct timeval delay; pth_event_t ev; static pth_key_t ev_key = PTH_KEY_INIT; fd_set fds; int fdmode; ssize_t rv; ssize_t s; int n; pth_implicit_init(); pth_debug2("pth_sendto_ev: enter from thread \"%s\"", pth_current->name); /* POSIX compliance */ if (nbytes == 0) return 0; if (!pth_util_fd_valid(fd)) return pth_error(-1, EBADF); /* force filedescriptor into non-blocking mode */ if ((fdmode = pth_fdmode(fd, PTH_FDMODE_NONBLOCK)) == PTH_FDMODE_ERROR) return pth_error(-1, EBADF); /* poll filedescriptor if not already in non-blocking operation */ if (fdmode != PTH_FDMODE_NONBLOCK) { /* now directly poll filedescriptor for writeability to avoid unneccessary (and resource consuming because of context switches, etc) event handling through the scheduler */ if (!pth_util_fd_valid(fd)) { pth_fdmode(fd, fdmode); return pth_error(-1, EBADF); } FD_ZERO(&fds); FD_SET(fd, &fds); delay.tv_sec = 0; delay.tv_usec = 0; while ((n = pth_sc(select)(fd+1, NULL, &fds, NULL, &delay)) < 0 && errno == EINTR) ; if (n < 0 && (errno == EINVAL || errno == EBADF)) return pth_error(-1, errno); rv = 0; for (;;) { /* if filedescriptor is still not writeable, let thread sleep until it is or event occurs */ if (n == 0) { ev = pth_event(PTH_EVENT_FD|PTH_UNTIL_FD_WRITEABLE|PTH_MODE_STATIC, &ev_key, fd); if (ev_extra != NULL) pth_event_concat(ev, ev_extra, NULL); pth_wait(ev); if (ev_extra != NULL) { pth_event_isolate(ev); if (pth_event_status(ev) != PTH_STATUS_OCCURRED) { pth_fdmode(fd, fdmode); return pth_error(-1, EINTR); } } } /* now perform the actual send operation */ while ((s = pth_sc(sendto)(fd, buf, nbytes, flags, to, tolen)) < 0 && errno == EINTR) ; if (s > 0) rv += s; /* although we're physically now in non-blocking mode, iterate unless all data is written or an error occurs, because we've to mimic the usual blocking I/O behaviour of write(2). */ if (s > 0 && s < (ssize_t)nbytes) { nbytes -= s; buf = (void *)((char *)buf + s); n = 0; continue; } /* pass error to caller, but not for partial writes (rv > 0) */ if (s < 0 && rv == 0) rv = -1; /* stop looping */ break; } } else { /* just perform the actual send operation */ while ((rv = pth_sc(sendto)(fd, buf, nbytes, flags, to, tolen)) < 0 && errno == EINTR) ; } /* restore filedescriptor mode */ pth_shield { pth_fdmode(fd, fdmode); } pth_debug2("pth_sendto_ev: leave to thread \"%s\"", pth_current->name); return rv; }