NAME¶
pth - GNU Portable Threads
VERSION¶
GNU Pth 2.0.7 (08-Jun-2006)
SYNOPSIS¶
- Global Library Management
- pth_init, pth_kill, pth_ctrl, pth_version.
- Thread Attribute Handling
- pth_attr_of, pth_attr_new, pth_attr_init, pth_attr_set,
pth_attr_get, pth_attr_destroy.
- Thread Control
- pth_spawn, pth_once, pth_self, pth_suspend, pth_resume,
pth_yield, pth_nap, pth_wait, pth_cancel, pth_abort, pth_raise, pth_join,
pth_exit.
- Utilities
- pth_fdmode, pth_time, pth_timeout, pth_sfiodisc.
- Cancellation Management
- pth_cancel_point, pth_cancel_state.
- Event Handling
- pth_event, pth_event_typeof, pth_event_extract,
pth_event_concat, pth_event_isolate, pth_event_walk, pth_event_status,
pth_event_free.
- Key-Based Storage
- pth_key_create, pth_key_delete, pth_key_setdata,
pth_key_getdata.
- Message Port Communication
- pth_msgport_create, pth_msgport_destroy, pth_msgport_find,
pth_msgport_pending, pth_msgport_put, pth_msgport_get,
pth_msgport_reply.
- Thread Cleanups
- pth_cleanup_push, pth_cleanup_pop.
- Process Forking
- pth_atfork_push, pth_atfork_pop, pth_fork.
- Synchronization
- pth_mutex_init, pth_mutex_acquire, pth_mutex_release,
pth_rwlock_init, pth_rwlock_acquire, pth_rwlock_release, pth_cond_init,
pth_cond_await, pth_cond_notify, pth_barrier_init, pth_barrier_reach.
- User-Space Context
- pth_uctx_create, pth_uctx_make, pth_uctx_switch,
pth_uctx_destroy.
- Generalized POSIX Replacement API
- pth_sigwait_ev, pth_accept_ev, pth_connect_ev,
pth_select_ev, pth_poll_ev, pth_read_ev, pth_readv_ev, pth_write_ev,
pth_writev_ev, pth_recv_ev, pth_recvfrom_ev, pth_send_ev,
pth_sendto_ev.
- Standard POSIX Replacement API
- pth_nanosleep, pth_usleep, pth_sleep, pth_waitpid,
pth_system, pth_sigmask, pth_sigwait, pth_accept, pth_connect, pth_select,
pth_pselect, pth_poll, pth_read, pth_readv, pth_write, pth_writev,
pth_pread, pth_pwrite, pth_recv, pth_recvfrom, pth_send, pth_sendto.
DESCRIPTION¶
____ _ _
⎪ _ \⎪ ⎪_⎪ ⎪__
⎪ ⎪_) ⎪ __⎪ '_ \ ``Only those who attempt
⎪ __/⎪ ⎪_⎪ ⎪ ⎪ ⎪ the absurd can achieve
⎪_⎪ \__⎪_⎪ ⎪_⎪ the impossible.''
Pth is a very portable POSIX/ANSI-C based library for Unix platforms
which provides non-preemptive priority-based scheduling for multiple threads
of execution (aka `multithreading') inside event-driven applications. All
threads run in the same address space of the application process, but each
thread has its own individual program counter, run-time stack, signal mask and
"errno" variable.
The thread scheduling itself is done in a cooperative way, i.e., the threads are
managed and dispatched by a priority- and event-driven non-preemptive
scheduler. The intention is that this way both better portability and run-time
performance is achieved than with preemptive scheduling. The event facility
allows threads to wait until various types of internal and external events
occur, including pending I/O on file descriptors, asynchronous signals,
elapsed timers, pending I/O on message ports, thread and process termination,
and even results of customized callback functions.
Pth also provides an optional emulation API for POSIX.1c threads
(`Pthreads') which can be used for backward compatibility to existing
multithreaded applications. See
Pth's
pthread(3) manual page for
details.
Threading Background
When programming event-driven applications, usually servers, lots of regular
jobs and one-shot requests have to be processed in parallel. To efficiently
simulate this parallel processing on uniprocessor machines, we use
`multitasking' -- that is, we have the application ask the operating system to
spawn multiple instances of itself. On Unix, typically the kernel implements
multitasking in a preemptive and priority-based way through heavy-weight
processes spawned with
fork(2). These processes usually do
not
share a common address space. Instead they are clearly separated from each
other, and are created by direct cloning a process address space (although
modern kernels use memory segment mapping and copy-on-write semantics to avoid
unnecessary copying of physical memory).
The drawbacks are obvious: Sharing data between the processes is complicated,
and can usually only be done efficiently through shared memory (but which
itself is not very portable). Synchronization is complicated because of the
preemptive nature of the Unix scheduler (one has to use
atomic locks,
etc). The machine's resources can be exhausted very quickly when the server
application has to serve too many long-running requests (heavy-weight
processes cost memory). And when each request spawns a sub-process to handle
it, the server performance and responsiveness is horrible (heavy-weight
processes cost time to spawn). Finally, the server application doesn't scale
very well with the load because of these resource problems. In practice, lots
of tricks are usually used to overcome these problems - ranging from
pre-forked sub-process pools to semi-serialized processing, etc.
One of the most elegant ways to solve these resource- and data-sharing problems
is to have multiple
light-weight threads of execution inside a single
(heavy-weight) process, i.e., to use
multithreading. Those
threads usually improve responsiveness and performance of the
application, often improve and simplify the internal program structure, and
most important, require less system resources than heavy-weight processes.
Threads are neither the optimal run-time facility for all types of
applications, nor can all applications benefit from them. But at least
event-driven server applications usually benefit greatly from using threads.
The World of Threading
Even though lots of documents exists which describe and define the world of
threading, to understand
Pth, you need only basic knowledge about
threading. The following definitions of thread-related terms should at least
help you understand thread programming enough to allow you to use
Pth.
- o process vs. thread
- A process on Unix systems consists of at least the
following fundamental ingredients: virtual memory table, program
code, program counter, heap memory, stack
memory, stack pointer, file descriptor set,
signal table. On every process switch, the kernel saves and
restores these ingredients for the individual processes. On the other
hand, a thread consists of only a private program counter, stack memory,
stack pointer and signal table. All other ingredients, in particular the
virtual memory, it shares with the other threads of the same process.
- o kernel-space vs. user-space
threading
- Threads on a Unix platform traditionally can be implemented
either inside kernel-space or user-space. When threads are implemented by
the kernel, the thread context switches are performed by the kernel
without the application's knowledge. Similarly, when threads are
implemented in user-space, the thread context switches are performed by an
application library, without the kernel's knowledge. There also are hybrid
threading approaches where, typically, a user-space library binds one or
more user-space threads to one or more kernel-space threads (there usually
called light-weight processes - or in short LWPs).
User-space threads are usually more portable and can perform faster and
cheaper context switches (for instance via swapcontext(2) or
setjmp(3)/ longjmp(3)) than kernel based threads. On the
other hand, kernel-space threads can take advantage of multiprocessor
machines and don't have any inherent I/O blocking problems. Kernel-space
threads are usually scheduled in preemptive way side-by-side with the
underlying processes. User-space threads on the other hand use either
preemptive or non-preemptive scheduling.
- o preemptive vs. non-preemptive thread
scheduling
- In preemptive scheduling, the scheduler lets a thread
execute until a blocking situation occurs (usually a function call which
would block) or the assigned timeslice elapses. Then it detracts control
from the thread without a chance for the thread to object. This is usually
realized by interrupting the thread through a hardware interrupt signal
(for kernel-space threads) or a software interrupt signal (for user-space
threads), like "SIGALRM" or "SIGVTALRM". In
non-preemptive scheduling, once a thread received control from the
scheduler it keeps it until either a blocking situation occurs (again a
function call which would block and instead switches back to the
scheduler) or the thread explicitly yields control back to the scheduler
in a cooperative way.
- o concurrency vs. parallelism
- Concurrency exists when at least two threads are in
progress at the same time. Parallelism arises when at least two
threads are executing simultaneously. Real parallelism can be only
achieved on multiprocessor machines, of course. But one also usually
speaks of parallelism or high concurrency in the context of
preemptive thread scheduling and of low concurrency in the context
of non-preemptive thread scheduling.
- o responsiveness
- The responsiveness of a system can be described by the user
visible delay until the system responses to an external request. When this
delay is small enough and the user doesn't recognize a noticeable delay,
the responsiveness of the system is considered good. When the user
recognizes or is even annoyed by the delay, the responsiveness of the
system is considered bad.
- o reentrant, thread-safe and
asynchronous-safe functions
- A reentrant function is one that behaves correctly if it is
called simultaneously by several threads and then also executes
simultaneously. Functions that access global state, such as memory or
files, of course, need to be carefully designed in order to be reentrant.
Two traditional approaches to solve these problems are caller-supplied
states and thread-specific data.
Thread-safety is the avoidance of data races, i.e., situations in
which data is set to either correct or incorrect value depending upon the
(unpredictable) order in which multiple threads access and modify the
data. So a function is thread-safe when it still behaves semantically
correct when called simultaneously by several threads (it is not required
that the functions also execute simultaneously). The traditional approach
to achieve thread-safety is to wrap a function body with an internal
mutual exclusion lock (aka `mutex'). As you should recognize, reentrant is
a stronger attribute than thread-safe, because it is harder to achieve and
results especially in no run-time contention between threads. So, a
reentrant function is always thread-safe, but not vice versa.
Additionally there is a related attribute for functions named
asynchronous-safe, which comes into play in conjunction with signal
handlers. This is very related to the problem of reentrant functions. An
asynchronous-safe function is one that can be called safe and without
side-effects from within a signal handler context. Usually very few
functions are of this type, because an application is very restricted in
what it can perform from within a signal handler (especially what system
functions it is allowed to call). The reason mainly is, because only a few
system functions are officially declared by POSIX as guaranteed to be
asynchronous-safe. Asynchronous-safe functions usually have to be already
reentrant.
User-Space Threads
User-space threads can be implemented in various way. The two traditional
approaches are:
- 1.
- Matrix-based explicit dispatching between small units of
execution:
Here the global procedures of the application are split into small execution
units (each is required to not run for more than a few milliseconds) and
those units are implemented by separate functions. Then a global matrix is
defined which describes the execution (and perhaps even dependency) order
of these functions. The main server procedure then just dispatches between
these units by calling one function after each other controlled by this
matrix. The threads are created by more than one jump-trail through this
matrix and by switching between these jump-trails controlled by
corresponding occurred events.
This approach gives the best possible performance, because one can fine-tune
the threads of execution by adjusting the matrix, and the scheduling is
done explicitly by the application itself. It is also very portable,
because the matrix is just an ordinary data structure, and functions are a
standard feature of ANSI C.
The disadvantage of this approach is that it is complicated to write large
applications with this approach, because in those applications one quickly
gets hundreds(!) of execution units and the control flow inside such an
application is very hard to understand (because it is interrupted by
function borders and one always has to remember the global dispatching
matrix to follow it). Additionally, all threads operate on the same
execution stack. Although this saves memory, it is often nasty, because
one cannot switch between threads in the middle of a function. Thus the
scheduling borders are the function borders.
- 2.
- Context-based implicit scheduling between threads of
execution:
Here the idea is that one programs the application as with forked processes,
i.e., one spawns a thread of execution and this runs from the begin to the
end without an interrupted control flow. But the control flow can be still
interrupted - even in the middle of a function. Actually in a preemptive
way, similar to what the kernel does for the heavy-weight processes, i.e.,
every few milliseconds the user-space scheduler switches between the
threads of execution. But the thread itself doesn't recognize this and
usually (except for synchronization issues) doesn't have to care about
this.
The advantage of this approach is that it's very easy to program, because
the control flow and context of a thread directly follows a procedure
without forced interrupts through function borders. Additionally, the
programming is very similar to a traditional and well understood
fork(2) based approach.
The disadvantage is that although the general performance is increased,
compared to using approaches based on heavy-weight processes, it is
decreased compared to the matrix-approach above. Because the implicit
preemptive scheduling does usually a lot more context switches (every
user-space context switch costs some overhead even when it is a lot
cheaper than a kernel-level context switch) than the explicit
cooperative/non-preemptive scheduling. Finally, there is no really
portable POSIX/ANSI-C based way to implement user-space preemptive
threading. Either the platform already has threads, or one has to hope
that some semi-portable package exists for it. And even those
semi-portable packages usually have to deal with assembler code and other
nasty internals and are not easy to port to forthcoming platforms.
So, in short: the matrix-dispatching approach is portable and fast, but nasty to
program. The thread scheduling approach is easy to program, but suffers from
synchronization and portability problems caused by its preemptive nature.
The Compromise of Pth
But why not combine the good aspects of both approaches while avoiding their bad
aspects? That's the goal of
Pth.
Pth implements easy-to-program
threads of execution, but avoids the problems of preemptive scheduling by
using non-preemptive scheduling instead.
This sounds like, and is, a useful approach. Nevertheless, one has to keep the
implications of non-preemptive thread scheduling in mind when working with
Pth. The following list summarizes a few essential points:
- o
- Pth provides maximum portability, but NOT the fanciest
features.
This is, because it uses a nifty and portable POSIX/ANSI-C approach for
thread creation (and this way doesn't require any platform dependent
assembler hacks) and schedules the threads in non-preemptive way (which
doesn't require unportable facilities like "SIGVTALRM"). On the
other hand, this way not all fancy threading features can be implemented.
Nevertheless the available facilities are enough to provide a robust and
full-featured threading system.
- o
- Pth increases the responsiveness and concurrency of an
event-driven application, but NOT the concurrency of
number-crunching applications.
The reason is the non-preemptive scheduling. Number-crunching applications
usually require preemptive scheduling to achieve concurrency because of
their long CPU bursts. For them, non-preemptive scheduling (even together
with explicit yielding) provides only the old concept of `coroutines'. On
the other hand, event driven applications benefit greatly from
non-preemptive scheduling. They have only short CPU bursts and lots of
events to wait on, and this way run faster under non-preemptive scheduling
because no unnecessary context switching occurs, as it is the case for
preemptive scheduling. That's why Pth is mainly intended for server
type applications, although there is no technical restriction.
- o
- Pth requires thread-safe functions, but NOT reentrant
functions.
This nice fact exists again because of the nature of non-preemptive
scheduling, where a function isn't interrupted and this way cannot be
reentered before it returned. This is a great portability benefit, because
thread-safety can be achieved more easily than reentrance possibility.
Especially this means that under Pth more existing third-party
libraries can be used without side-effects than it's the case for other
threading systems.
- o
- Pth doesn't require any kernel support, but can NOT
benefit from multiprocessor machines.
This means that Pth runs on almost all Unix kernels, because the
kernel does not need to be aware of the Pth threads (because they
are implemented entirely in user-space). On the other hand, it cannot
benefit from the existence of multiprocessors, because for this, kernel
support would be needed. In practice, this is no problem, because
multiprocessor systems are rare, and portability is almost more important
than highest concurrency.
The life cycle of a thread
To understand the
Pth Application Programming Interface (API), it helps
to first understand the life cycle of a thread in the
Pth threading
system. It can be illustrated with the following directed graph:
NEW
⎪
V
+---> READY ---+
⎪ ^ ⎪
⎪ ⎪ V
WAITING <--+-- RUNNING
⎪
: V
SUSPENDED DEAD
When a new thread is created, it is moved into the
NEW queue of the
scheduler. On the next dispatching for this thread, the scheduler picks it up
from there and moves it to the
READY queue. This is a queue containing
all threads which want to perform a CPU burst. There they are queued in
priority order. On each dispatching step, the scheduler always removes the
thread with the highest priority only. It then increases the priority of all
remaining threads by 1, to prevent them from `starving'.
The thread which was removed from the
READY queue is the new
RUNNING thread (there is always just one
RUNNING thread, of
course). The
RUNNING thread is assigned execution control. After this
thread yields execution (either explicitly by yielding execution or implicitly
by calling a function which would block) there are three possibilities: Either
it has terminated, then it is moved to the
DEAD queue, or it has events
on which it wants to wait, then it is moved into the
WAITING queue.
Else it is assumed it wants to perform more CPU bursts and immediately enters
the
READY queue again.
Before the next thread is taken out of the
READY queue, the
WAITING queue is checked for pending events. If one or more events
occurred, the threads that are waiting on them are immediately moved to the
READY queue.
The purpose of the
NEW queue has to do with the fact that in
Pth a
thread never directly switches to another thread. A thread always yields
execution to the scheduler and the scheduler dispatches to the next thread. So
a freshly spawned thread has to be kept somewhere until the scheduler gets a
chance to pick it up for scheduling. That is what the
NEW queue is for.
The purpose of the
DEAD queue is to support thread joining. When a thread
is marked to be unjoinable, it is directly kicked out of the system after it
terminated. But when it is joinable, it enters the
DEAD queue. There it
remains until another thread joins it.
Finally, there is a special separated queue named
SUSPENDED, to where
threads can be manually moved from the
NEW,
READY or
WAITING queues by the application. The purpose of this special queue is
to temporarily absorb suspended threads until they are again resumed by the
application. Suspended threads do not cost scheduling or event handling
resources, because they are temporarily completely out of the scheduler's
scope. If a thread is resumed, it is moved back to the queue from where it
originally came and this way again enters the schedulers scope.
APPLICATION PROGRAMMING INTERFACE (API)¶
In the following the
Pth Application Programming Interface (API)
is discussed in detail. With the knowledge given above, it should now be easy
to understand how to program threads with this API. In good Unix tradition,
Pth functions use special return values ("NULL" in pointer
context, "FALSE" in boolean context and "-1" in integer
context) to indicate an error condition and set (or pass through) the
"errno" system variable to pass more details about the error to the
caller.
Global Library Management
The following functions act on the library as a whole. They are used to
initialize and shutdown the scheduler and fetch information from it.
- int pth_init(void);
- This initializes the Pth library. It has to be the
first Pth API function call in an application, and is mandatory.
It's usually done at the begin of the main() function of the
application. This implicitly spawns the internal scheduler thread and
transforms the single execution unit of the current process into a thread
(the `main' thread). It returns "TRUE" on success and
"FALSE" on error.
- int pth_kill(void);
- This kills the Pth library. It should be the last
Pth API function call in an application, but is not really
required. It's usually done at the end of the main function of the
application. At least, it has to be called from within the main thread. It
implicitly kills all threads and transforms back the calling thread into
the single execution unit of the underlying process. The usual way to
terminate a Pth application is either a simple
`"pth_exit(0);"' in the main thread (which waits for all other
threads to terminate, kills the threading system and then terminates the
process) or a `"pth_kill(); exit(0)"' (which immediately kills
the threading system and terminates the process). The pth_kill()
return immediately with a return code of "FALSE" if it is not
called from within the main thread. Else it kills the threading system and
returns "TRUE".
- long pth_ctrl(unsigned long query, ...);
- This is a generalized query/control function for the
Pth library. The argument query is a bitmask formed out of
one or more "PTH_CTRL_" XXXX queries. Currently the
following queries are supported:
- "PTH_CTRL_GETTHREADS"
- This returns the total number of threads currently in
existence. This query actually is formed out of the combination of queries
for threads in a particular state, i.e., the
"PTH_CTRL_GETTHREADS" query is equal to the OR-combination of
all the following specialized queries:
"PTH_CTRL_GETTHREADS_NEW" for the number of threads in the new
queue (threads created via pth_spawn(3) but still not scheduled
once), "PTH_CTRL_GETTHREADS_READY" for the number of threads in
the ready queue (threads who want to do CPU bursts),
"PTH_CTRL_GETTHREADS_RUNNING" for the number of running threads
(always just one thread!), "PTH_CTRL_GETTHREADS_WAITING" for the
number of threads in the waiting queue (threads waiting for events),
"PTH_CTRL_GETTHREADS_SUSPENDED" for the number of threads in the
suspended queue (threads waiting to be resumed) and
"PTH_CTRL_GETTHREADS_DEAD" for the number of threads in the new
queue (terminated threads waiting for a join).
- "PTH_CTRL_GETAVLOAD"
- This requires a second argument of type `"float
*"' (pointer to a floating point variable). It stores a floating
point value describing the exponential averaged load of the scheduler in
this variable. The load is a function from the number of threads in the
ready queue of the schedulers dispatching unit. So a load around 1.0 means
there is only one ready thread (the standard situation when the
application has no high load). A higher load value means there a more
threads ready who want to do CPU bursts. The average load value updates
once per second only. The return value for this query is always 0.
- "PTH_CTRL_GETPRIO"
- This requires a second argument of type `"pth_t"'
which identifies a thread. It returns the priority (ranging from
"PTH_PRIO_MIN" to "PTH_PRIO_MAX") of the given
thread.
- "PTH_CTRL_GETNAME"
- This requires a second argument of type `"pth_t"'
which identifies a thread. It returns the name of the given thread, i.e.,
the return value of pth_ctrl(3) should be casted to a `"char
*"'.
- "PTH_CTRL_DUMPSTATE"
- This requires a second argument of type `"FILE
*"' to which a summary of the internal Pth library state is
written to. The main information which is currently written out is the
current state of the thread pool.
- "PTH_CTRL_FAVOURNEW"
- This requires a second argument of type `"int"'
which specified whether the GNU Pth scheduler favours new threads
on startup, i.e., whether they are moved from the new queue to the top
(argument is "TRUE") or middle (argument is "FALSE")
of the ready queue. The default is to favour new threads to make sure they
do not starve already at startup, although this slightly violates the
strict priority based scheduling.
The function returns "-1" on error.
- long pth_version(void);
- This function returns a hex-value
`0xVRRT LL' which describes the current
Pth library version. V is the version, RR the
revisions, LL the level and T the type of the level
(alphalevel=0, betalevel=1, patchlevel=2, etc). For instance Pth
version 1.0b1 is encoded as 0x100101. The reason for this unusual mapping
is that this way the version number is steadily increasing. The
same value is also available under compile time as
"PTH_VERSION".
Thread Attribute Handling
Attribute objects are used in
Pth for two things: First
stand-alone/unbound attribute objects are used to store attributes for to be
spawned threads. Bounded attribute objects are used to modify attributes of
already existing threads. The following attribute fields exists in attribute
objects:
- "PTH_ATTR_PRIO" (read-write)
["int"]
- Thread Priority between "PTH_PRIO_MIN" and
"PTH_PRIO_MAX". The default is "PTH_PRIO_STD".
- "PTH_ATTR_NAME" (read-write) ["char
*"]
- Name of thread (up to 40 characters are stored only),
mainly for debugging purposes.
- "PTH_ATTR_DISPATCHES" (read-write)
["int"]
- In bounded attribute objects, this field is incremented
every time the context is switched to the associated thread.
- "PTH_ATTR_JOINABLE" (read-write>
["int"]
- The thread detachment type, "TRUE" indicates a
joinable thread, "FALSE" indicates a detached thread. When a
thread is detached, after termination it is immediately kicked out of the
system instead of inserted into the dead queue.
- "PTH_ATTR_CANCEL_STATE" (read-write)
["unsigned int"]
- The thread cancellation state, i.e., a combination of
"PTH_CANCEL_ENABLE" or "PTH_CANCEL_DISABLE" and
"PTH_CANCEL_DEFERRED" or
"PTH_CANCEL_ASYNCHRONOUS".
- "PTH_ATTR_STACK_SIZE" (read-write)
["unsigned int"]
- The thread stack size in bytes. Use lower values than 64 KB
with great care!
- "PTH_ATTR_STACK_ADDR" (read-write) ["char
*"]
- A pointer to the lower address of a chunk of
malloc(3)'ed memory for the stack.
- "PTH_ATTR_TIME_SPAWN" (read-only)
["pth_time_t"]
- The time when the thread was spawned. This can be queried
only when the attribute object is bound to a thread.
- "PTH_ATTR_TIME_LAST" (read-only)
["pth_time_t"]
- The time when the thread was last dispatched. This can be
queried only when the attribute object is bound to a thread.
- "PTH_ATTR_TIME_RAN" (read-only)
["pth_time_t"]
- The total time the thread was running. This can be queried
only when the attribute object is bound to a thread.
- "PTH_ATTR_START_FUNC" (read-only) ["void
*(*)(void *)"]
- The thread start function. This can be queried only when
the attribute object is bound to a thread.
- "PTH_ATTR_START_ARG" (read-only) ["void
*"]
- The thread start argument. This can be queried only when
the attribute object is bound to a thread.
- "PTH_ATTR_STATE" (read-only)
["pth_state_t"]
- The scheduling state of the thread, i.e., either
"PTH_STATE_NEW", "PTH_STATE_READY",
"PTH_STATE_WAITING", or "PTH_STATE_DEAD" This can be
queried only when the attribute object is bound to a thread.
- "PTH_ATTR_EVENTS" (read-only)
["pth_event_t"]
- The event ring the thread is waiting for. This can be
queried only when the attribute object is bound to a thread.
- "PTH_ATTR_BOUND" (read-only)
["int"]
- Whether the attribute object is bound ("TRUE") to
a thread or not ("FALSE").
The following API functions can be used to handle the attribute objects:
- pth_attr_t pth_attr_of(pth_t tid);
- This returns a new attribute object bound to thread
tid. Any queries on this object directly fetch attributes from
tid. And attribute modifications directly change tid. Use
such attribute objects to modify existing threads.
- pth_attr_t pth_attr_new(void);
- This returns a new unbound attribute object. An
implicit pth_attr_init() is done on it. Any queries on this object
just fetch stored attributes from it. And attribute modifications just
change the stored attributes. Use such attribute objects to pre-configure
attributes for to be spawned threads.
- int pth_attr_init(pth_attr_t attr);
- This initializes an attribute object attr to the
default values: "PTH_ATTR_PRIO" := "PTH_PRIO_STD",
"PTH_ATTR_NAME" := `"unknown"',
"PTH_ATTR_DISPATCHES" := 0, "PTH_ATTR_JOINABLE" :=
"TRUE", "PTH_ATTR_CANCELSTATE" :=
"PTH_CANCEL_DEFAULT", "PTH_ATTR_STACK_SIZE" := 64*1024
and "PTH_ATTR_STACK_ADDR" := "NULL". All other
"PTH_ATTR_*" attributes are read-only attributes and don't
receive default values in attr, because they exists only for
bounded attribute objects.
- int pth_attr_set(pth_attr_t attr, int
field, ...);
- This sets the attribute field field in attr
to a value specified as an additional argument on the variable argument
list. The following attribute fields and argument pairs can be
used:
PTH_ATTR_PRIO int
PTH_ATTR_NAME char *
PTH_ATTR_DISPATCHES int
PTH_ATTR_JOINABLE int
PTH_ATTR_CANCEL_STATE unsigned int
PTH_ATTR_STACK_SIZE unsigned int
PTH_ATTR_STACK_ADDR char *
- int pth_attr_get(pth_attr_t attr, int
field, ...);
- This retrieves the attribute field field in
attr and stores its value in the variable specified through a
pointer in an additional argument on the variable argument list. The
following fields and argument pairs can be used:
PTH_ATTR_PRIO int *
PTH_ATTR_NAME char **
PTH_ATTR_DISPATCHES int *
PTH_ATTR_JOINABLE int *
PTH_ATTR_CANCEL_STATE unsigned int *
PTH_ATTR_STACK_SIZE unsigned int *
PTH_ATTR_STACK_ADDR char **
PTH_ATTR_TIME_SPAWN pth_time_t *
PTH_ATTR_TIME_LAST pth_time_t *
PTH_ATTR_TIME_RAN pth_time_t *
PTH_ATTR_START_FUNC void *(**)(void *)
PTH_ATTR_START_ARG void **
PTH_ATTR_STATE pth_state_t *
PTH_ATTR_EVENTS pth_event_t *
PTH_ATTR_BOUND int *
- int pth_attr_destroy(pth_attr_t attr);
- This destroys a attribute object attr. After this
attr is no longer a valid attribute object.
Thread Control
The following functions control the threading itself and make up the main API of
the
Pth library.
- pth_t pth_spawn(pth_attr_t attr, void
*(*entry)(void *), void * arg);
- This spawns a new thread with the attributes given in
attr (or "PTH_ATTR_DEFAULT" for default attributes -
which means that thread priority, joinability and cancel state are
inherited from the current thread) with the starting point at routine
entry; the dispatch count is not inherited from the current thread
if attr is not specified - rather, it is initialized to zero. This
entry routine is called as `pth_exit( entry(arg))' inside
the new thread unit, i.e., entry's return value is fed to an
implicit pth_exit(3). So the thread can also exit by just
returning. Nevertheless the thread can also exit explicitly at any time by
calling pth_exit(3). But keep in mind that calling the POSIX
function exit(3) still terminates the complete process and not just
the current thread.
There is no Pth-internal limit on the number of threads one can
spawn, except the limit implied by the available virtual memory.
Pth internally keeps track of thread in dynamic data structures.
The function returns "NULL" on error.
- int pth_once(pth_once_t *ctrlvar, void
(*func)(void *), void * arg);
- This is a convenience function which uses a control
variable of type "pth_once_t" to make sure a constructor
function func is called only once as ` func(arg)' in
the system. In other words: Only the first call to pth_once(3) by
any thread in the system succeeds. The variable referenced via
ctrlvar should be declared as `"pth_once_t"
variable-name = "PTH_ONCE_INIT";' before calling this
function.
- pth_t pth_self(void);
- This just returns the unique thread handle of the currently
running thread. This handle itself has to be treated as an opaque entity
by the application. It's usually used as an argument to other functions
who require an argument of type "pth_t".
- int pth_suspend(pth_t tid);
- This suspends a thread tid until it is manually
resumed again via pth_resume(3). For this, the thread is moved to
the SUSPENDED queue and this way is completely out of the
scheduler's event handling and thread dispatching scope. Suspending the
current thread is not allowed. The function returns "TRUE" on
success and "FALSE" on errors.
- int pth_resume(pth_t tid);
- This function resumes a previously suspended thread
tid, i.e. tid has to stay on the SUSPENDED queue. The
thread is moved to the NEW, READY or WAITING queue
(dependent on what its state was when the pth_suspend(3) call were
made) and this way again enters the event handling and thread dispatching
scope of the scheduler. The function returns "TRUE" on success
and "FALSE" on errors.
- int pth_raise(pth_t tid, int sig)
- This function raises a signal for delivery to thread
tid only. When one just raises a signal via raise(3) or
kill(2), its delivered to an arbitrary thread which has this signal
not blocked. With pth_raise(3) one can send a signal to a thread
and its guarantees that only this thread gets the signal delivered. But
keep in mind that nevertheless the signals action is still
configured process-wide. When sig is 0 plain thread checking
is performed, i.e., `"pth_raise(tid, 0)"' returns
"TRUE" when thread tid still exists in the PTH
system but doesn't send any signal to it.
- int pth_yield(pth_t tid);
- This explicitly yields back the execution control to the
scheduler thread. Usually the execution is implicitly transferred back to
the scheduler when a thread waits for an event. But when a thread has to
do larger CPU bursts, it can be reasonable to interrupt it explicitly by
doing a few pth_yield(3) calls to give other threads a chance to
execute, too. This obviously is the cooperating part of Pth. A
thread has not to yield execution, of course. But when you want to
program a server application with good response times the threads should
be cooperative, i.e., when they should split their CPU bursts into smaller
units with this call.
Usually one specifies tid as "NULL" to indicate to the
scheduler that it can freely decide which thread to dispatch next. But if
one wants to indicate to the scheduler that a particular thread should be
favored on the next dispatching step, one can specify this thread
explicitly. This allows the usage of the old concept of coroutines
where a thread/routine switches to a particular cooperating thread. If
tid is not "NULL" and points to a new or
ready thread, it is guaranteed that this thread receives execution
control on the next dispatching step. If tid is in a different
state (that is, not in "PTH_STATE_NEW" or
"PTH_STATE_READY") an error is reported.
The function usually returns "TRUE" for success and only
"FALSE" (with "errno" set to "EINVAL") if
tid specified an invalid or still not new or ready thread.
- int pth_nap(pth_time_t naptime);
- This functions suspends the execution of the current thread
until naptime is elapsed. naptime is of type
"pth_time_t" and this way has theoretically a resolution of one
microsecond. In practice you should neither rely on this nor that the
thread is awakened exactly after naptime has elapsed. It's only
guarantees that the thread will sleep at least naptime. But because
of the non-preemptive nature of Pth it can last longer (when
another thread kept the CPU for a long time). Additionally the resolution
is dependent of the implementation of timers by the operating system and
these usually have only a resolution of 10 microseconds or larger. But
usually this isn't important for an application unless it tries to use
this facility for real time tasks.
- int pth_wait(pth_event_t ev);
- This is the link between the scheduler and the event
facility (see below for the various pth_event_xxx() functions).
It's modeled like select(2), i.e., one gives this function one or
more events (in the event ring specified by ev) on which the
current thread wants to wait. The scheduler awakes the thread when one ore
more of them occurred or failed after tagging them as such. The ev
argument is a pointer to an event ring which isn't changed except
for the tagging. pth_wait(3) returns the number of occurred or
failed events and the application can use pth_event_status(3) to
test which events occurred or failed.
- int pth_cancel(pth_t tid);
- This cancels a thread tid. How the cancellation is
done depends on the cancellation state of tid which the thread can
configure itself. When its state is "PTH_CANCEL_DISABLE" a
cancellation request is just made pending. When it is
"PTH_CANCEL_ENABLE" it depends on the cancellation type what is
performed. When its "PTH_CANCEL_DEFERRED" again the cancellation
request is just made pending. But when its
"PTH_CANCEL_ASYNCHRONOUS" the thread is immediately canceled
before pth_cancel(3) returns. The effect of a thread cancellation
is equal to implicitly forcing the thread to call
`"pth_exit(PTH_CANCELED)"' at one of his cancellation points. In
Pth thread enter a cancellation point either explicitly via
pth_cancel_point(3) or implicitly by waiting for an event.
- int pth_abort(pth_t tid);
- This is the cruel way to cancel a thread tid. When
it's already dead and waits to be joined it just joins it (via
`"pth_join(" tid", NULL)"') and this way kicks
it out of the system. Else it forces the thread to be not joinable and to
allow asynchronous cancellation and then cancels it via
`"pth_cancel(" tid")"'.
- int pth_join(pth_t tid, void
**value);
- This joins the current thread with the thread specified via
tid. It first suspends the current thread until the tid
thread has terminated. Then it is awakened and stores the value of
tid's pth_exit(3) call into *value (if value
and not "NULL") and returns to the caller. A thread can be
joined only when it has the attribute "PTH_ATTR_JOINABLE" set to
"TRUE" (the default). A thread can only be joined once, i.e.,
after the pth_join(3) call the thread tid is completely
removed from the system.
- void pth_exit(void *value);
- This terminates the current thread. Whether it's
immediately removed from the system or inserted into the dead queue of the
scheduler depends on its join type which was specified at spawning time.
If it has the attribute "PTH_ATTR_JOINABLE" set to
"FALSE", it's immediately removed and value is ignored.
Else the thread is inserted into the dead queue and value
remembered for a subsequent pth_join(3) call by another thread.
Utilities
Utility functions.
- int pth_fdmode(int fd, int mode);
- This switches the non-blocking mode flag on file descriptor
fd. The argument mode can be "PTH_FDMODE_BLOCK"
for switching fd into blocking I/O mode,
"PTH_FDMODE_NONBLOCK" for switching fd into non-blocking
I/O mode or "PTH_FDMODE_POLL" for just polling the current mode.
The current mode is returned (either "PTH_FDMODE_BLOCK" or
"PTH_FDMODE_NONBLOCK") or "PTH_FDMODE_ERROR" on error.
Keep in mind that since Pth 1.1 there is no longer a requirement to
manually switch a file descriptor into non-blocking mode in order to use
it. This is automatically done temporarily inside Pth. Instead when
you now switch a file descriptor explicitly into non-blocking mode,
pth_read(3) or pth_write(3) will never block the current
thread.
- pth_time_t pth_time(long sec, long
usec);
- This is a constructor for a "pth_time_t"
structure which is a convenient function to avoid temporary structure
values. It returns a pth_time_t structure which holds the absolute
time value specified by sec and usec.
- pth_time_t pth_timeout(long sec, long
usec);
- This is a constructor for a "pth_time_t"
structure which is a convenient function to avoid temporary structure
values. It returns a pth_time_t structure which holds the absolute
time value calculated by adding sec and usec to the current
time.
- Sfdisc_t *pth_sfiodisc(void);
- This functions is always available, but only reasonably
usable when Pth was built with Sfio support
("--with-sfio" option) and "PTH_EXT_SFIO" is then
defined by "pth.h". It is useful for applications which want to
use the comprehensive Sfio I/O library with the Pth
threading library. Then this function can be used to get an Sfio
discipline structure ("Sfdisc_t") which can be pushed onto
Sfio streams ("Sfio_t") in order to let this stream use
pth_read(3)/pth_write(2) instead of
read(2)/write(2). The benefit is that this way I/O on the
Sfio stream does only block the current thread instead of the whole
process. The application has to free(3) the "Sfdisc_t"
structure when it is no longer needed. The Sfio package can be found at
http://www.research.att.com/sw/tools/sfio/.
Cancellation Management
Pth supports POSIX style thread cancellation via
pth_cancel(3) and
the following two related functions:
- void pth_cancel_state(int newstate, int
*oldstate);
- This manages the cancellation state of the current thread.
When oldstate is not "NULL" the function stores the old
cancellation state under the variable pointed to by oldstate. When
newstate is not 0 it sets the new cancellation state.
oldstate is created before newstate is set. A state is a
combination of "PTH_CANCEL_ENABLE" or
"PTH_CANCEL_DISABLE" and "PTH_CANCEL_DEFERRED" or
"PTH_CANCEL_ASYNCHRONOUS".
"PTH_CANCEL_ENABLE⎪PTH_CANCEL_DEFERRED" (or
"PTH_CANCEL_DEFAULT") is the default state where cancellation is
possible but only at cancellation points. Use
"PTH_CANCEL_DISABLE" to complete disable cancellation for a
thread and "PTH_CANCEL_ASYNCHRONOUS" for allowing asynchronous
cancellations, i.e., cancellations which can happen at any time.
- void pth_cancel_point(void);
- This explicitly enter a cancellation point. When the
current cancellation state is "PTH_CANCEL_DISABLE" or no
cancellation request is pending, this has no side-effect and returns
immediately. Else it calls `"pth_exit(PTH_CANCELED)"'.
Event Handling
Pth has a very flexible event facility which is linked into the scheduler
through the
pth_wait(3) function. The following functions provide the
handling of event rings.
- pth_event_t pth_event(unsigned long spec,
...);
- This creates a new event ring consisting of a single
initial event. The type of the generated event is specified by
spec. The following types are available:
- "PTH_EVENT_FD"
- This is a file descriptor event. One or more of
"PTH_UNTIL_FD_READABLE", "PTH_UNTIL_FD_WRITEABLE" or
"PTH_UNTIL_FD_EXCEPTION" have to be OR-ed into spec to
specify on which state of the file descriptor you want to wait. The file
descriptor itself has to be given as an additional argument. Example:
`"pth_event(PTH_EVENT_FD⎪PTH_UNTIL_FD_READABLE,
fd)"'.
- "PTH_EVENT_SELECT"
- This is a multiple file descriptor event modeled directly
after the select(2) call (actually it is also used to implement
pth_select(3) internally). It's a convenient way to wait for a
large set of file descriptors at once and at each file descriptor for a
different type of state. Additionally as a nice side-effect one receives
the number of file descriptors which causes the event to be occurred
(using BSD semantics, i.e., when a file descriptor occurred in two sets
it's counted twice). The arguments correspond directly to the
select(2) function arguments except that there is no timeout
argument (because timeouts already can be handled via
"PTH_EVENT_TIME" events).
Example: `"pth_event(PTH_EVENT_SELECT, &rc, nfd, rfds, wfds,
efds)"' where "rc" has to be of type `"int *"',
"nfd" has to be of type `"int"' and "rfds",
"wfds" and "efds" have to be of type `"fd_set
*"' (see select(2)). The number of occurred file descriptors
are stored in "rc".
- "PTH_EVENT_SIGS"
- This is a signal set event. The two additional arguments
have to be a pointer to a signal set (type `"sigset_t *"') and a
pointer to a signal number variable (type `"int *"'). This event
waits until one of the signals in the signal set occurred. As a result the
occurred signal number is stored in the second additional argument. Keep
in mind that the Pth scheduler doesn't block signals automatically.
So when you want to wait for a signal with this event you've to block it
via sigprocmask(2) or it will be delivered without your notice.
Example: `"sigemptyset(&set); sigaddset(&set, SIGINT);
pth_event(PTH_EVENT_SIG, &set, &sig);"'.
- "PTH_EVENT_TIME"
- This is a time point event. The additional argument has to
be of type "pth_time_t" (usually on-the-fly generated via
pth_time(3)). This events waits until the specified time point has
elapsed. Keep in mind that the value is an absolute time point and not an
offset. When you want to wait for a specified amount of time, you've to
add the current time to the offset (usually on-the-fly achieved via
pth_timeout(3)). Example: `"pth_event(PTH_EVENT_TIME,
pth_timeout(2,0))"'.
- "PTH_EVENT_MSG"
- This is a message port event. The additional argument has
to be of type "pth_msgport_t". This events waits until one or
more messages were received on the specified message port. Example:
`"pth_event(PTH_EVENT_MSG, mp)"'.
- "PTH_EVENT_TID"
- This is a thread event. The additional argument has to be
of type "pth_t". One of "PTH_UNTIL_TID_NEW",
"PTH_UNTIL_TID_READY", "PTH_UNTIL_TID_WAITING" or
"PTH_UNTIL_TID_DEAD" has to be OR-ed into spec to specify
on which state of the thread you want to wait. Example:
`"pth_event(PTH_EVENT_TID⎪PTH_UNTIL_TID_DEAD, tid)"'.
- "PTH_EVENT_FUNC"
- This is a custom callback function event. Three additional
arguments have to be given with the following types: `"int (*)(void
*)"', `"void *"' and `"pth_time_t"'. The first is
a function pointer to a check function and the second argument is a
user-supplied context value which is passed to this function. The
scheduler calls this function on a regular basis (on his own scheduler
stack, so be very careful!) and the thread is kept sleeping while the
function returns "FALSE". Once it returned "TRUE" the
thread will be awakened. The check interval is defined by the third
argument, i.e., the check function is polled again not until this amount
of time elapsed. Example: `"pth_event(PTH_EVENT_FUNC, func, arg,
pth_time(0,500000))"'.
- unsigned long pth_event_typeof(pth_event_t
ev);
- This returns the type of event ev. It's a
combination of the describing "PTH_EVENT_XX" and
"PTH_UNTIL_XX" value. This is especially useful to know which
arguments have to be supplied to the pth_event_extract(3)
function.
- int pth_event_extract(pth_event_t ev,
...);
- When pth_event(3) is treated like sprintf(3),
then this function is sscanf(3), i.e., it is the inverse operation
of pth_event(3). This means that it can be used to extract the
ingredients of an event. The ingredients are stored into variables which
are given as pointers on the variable argument list. Which pointers have
to be present depends on the event type and has to be determined by the
caller before via pth_event_typeof(3).
To make it clear, when you constructed ev via `"ev =
pth_event(PTH_EVENT_FD, fd);"' you have to extract it via
`"pth_event_extract(ev, &fd)"', etc. For multiple arguments
of an event the order of the pointer arguments is the same as for
pth_event(3). But always keep in mind that you have to always
supply pointers to variables and these variables have to be
of the same type as the argument of pth_event(3) required.
- pth_event_t pth_event_concat(pth_event_t ev,
...);
- This concatenates one or more additional event rings to the
event ring ev and returns ev. The end of the argument list
has to be marked with a "NULL" argument. Use this function to
create real events rings out of the single-event rings created by
pth_event(3).
- pth_event_t pth_event_isolate(pth_event_t
ev);
- This isolates the event ev from possibly appended
events in the event ring. When in ev only one event exists, this
returns "NULL". When remaining events exists, they form a new
event ring which is returned.
- pth_event_t pth_event_walk(pth_event_t ev,
int direction);
- This walks to the next (when direction is
"PTH_WALK_NEXT") or previews (when direction is
"PTH_WALK_PREV") event in the event ring ev and returns
this new reached event. Additionally "PTH_UNTIL_OCCURRED" can be
OR-ed into direction to walk to the next/previous occurred event in
the ring ev.
- pth_status_t pth_event_status(pth_event_t
ev);
- This returns the status of event ev. This is a fast
operation because only a tag on ev is checked which was either set
or still not set by the scheduler. In other words: This doesn't check the
event itself, it just checks the last knowledge of the scheduler. The
possible returned status codes are: "PTH_STATUS_PENDING" (event
is still pending), "PTH_STATUS_OCCURRED" (event successfully
occurred), "PTH_STATUS_FAILED" (event failed).
- int pth_event_free(pth_event_t ev, int
mode);
- This deallocates the event ev (when mode is
"PTH_FREE_THIS") or all events appended to the event ring under
ev (when mode is "PTH_FREE_ALL").
Key-Based Storage
The following functions provide thread-local storage through unique keys similar
to the POSIX
Pthread API. Use this for thread specific global data.
- int pth_key_create(pth_key_t *key, void
(*func)(void *));
- This created a new unique key and stores it in key.
Additionally func can specify a destructor function which is called
on the current threads termination with the key.
- int pth_key_delete(pth_key_t key);
- This explicitly destroys a key key.
- int pth_key_setdata(pth_key_t key, const void
* value);
- This stores value under key.
- void *pth_key_getdata(pth_key_t key);
- This retrieves the value under key.
Message Port Communication
The following functions provide message ports which can be used for efficient
and flexible inter-thread communication.
- pth_msgport_t pth_msgport_create(const char
*name);
- This returns a pointer to a new message port. If name
name is not "NULL", the name can be used by other
threads via pth_msgport_find(3) to find the message port in case
they do not know directly the pointer to the message port.
- void pth_msgport_destroy(pth_msgport_t
mp);
- This destroys a message port mp. Before all pending
messages on it are replied to their origin message port.
- pth_msgport_t pth_msgport_find(const char
*name);
- This finds a message port in the system by name and
returns the pointer to it.
- int pth_msgport_pending(pth_msgport_t
mp);
- This returns the number of pending messages on message port
mp.
- int pth_msgport_put(pth_msgport_t mp,
pth_message_t * m);
- This puts (or sends) a message m to message port
mp.
- pth_message_t *pth_msgport_get(pth_msgport_t
mp);
- This gets (or receives) the top message from message port
mp. Incoming messages are always kept in a queue, so there can be
more pending messages, of course.
- int pth_msgport_reply(pth_message_t *m);
- This replies a message m to the message port of the
sender.
Thread Cleanups
Per-thread cleanup functions.
- int pth_cleanup_push(void (*handler)(void *),
void * arg);
- This pushes the routine handler onto the stack of
cleanup routines for the current thread. These routines are called in LIFO
order when the thread terminates.
- int pth_cleanup_pop(int execute);
- This pops the top-most routine from the stack of cleanup
routines for the current thread. When execute is "TRUE"
the routine is additionally called.
Process Forking
The following functions provide some special support for process forking
situations inside the threading environment.
- int pth_atfork_push(void (*prepare)(void *),
void (*)(void * parent), void (*)(void *child), void
*arg);
- This function declares forking handlers to be called before
and after pth_fork(3), in the context of the thread that called
pth_fork(3). The prepare handler is called before
fork(2) processing commences. The parent handler is called
after fork(2) processing completes in the parent process. The
child handler is called after fork(2) processing completed
in the child process. If no handling is desired at one or more of these
three points, the corresponding handler can be given as "NULL".
Each handler is called with arg as the argument.
The order of calls to pth_atfork_push(3) is significant. The
parent and child handlers are called in the order in which
they were established by calls to pth_atfork_push(3), i.e., FIFO.
The prepare fork handlers are called in the opposite order, i.e.,
LIFO.
- int pth_atfork_pop(void);
- This removes the top-most handlers on the forking handler
stack which were established with the last pth_atfork_push(3) call.
It returns "FALSE" when no more handlers couldn't be removed
from the stack.
- pid_t pth_fork(void);
- This is a variant of fork(2) with the difference
that the current thread only is forked into a separate process, i.e., in
the parent process nothing changes while in the child process all threads
are gone except for the scheduler and the calling thread. When you really
want to duplicate all threads in the current process you should use
fork(2) directly. But this is usually not reasonable. Additionally
this function takes care of forking handlers as established by
pth_fork_push(3).
Synchronization
The following functions provide synchronization support via mutual exclusion
locks (
mutex), read-write locks (
rwlock), condition variables
(
cond) and barriers (
barrier). Keep in mind that in a
non-preemptive threading system like
Pth this might sound unnecessary
at the first look, because a thread isn't interrupted by the system. Actually
when you have a critical code section which doesn't contain any
pth_xxx() functions, you don't need any mutex to protect it, of course.
But when your critical code section contains any
pth_xxx() function the
chance is high that these temporarily switch to the scheduler. And this way
other threads can make progress and enter your critical code section, too.
This is especially true for critical code sections which implicitly or
explicitly use the event mechanism.
- int pth_mutex_init(pth_mutex_t *mutex);
- This dynamically initializes a mutex variable of type
`"pth_mutex_t"'. Alternatively one can also use static
initialization via `"pth_mutex_t mutex = PTH_MUTEX_INIT"'.
- int pth_mutex_acquire(pth_mutex_t *mutex, int
try, pth_event_t ev);
- This acquires a mutex mutex. If the mutex is already
locked by another thread, the current threads execution is suspended until
the mutex is unlocked again or additionally the extra events in ev
occurred (when ev is not "NULL"). Recursive locking is
explicitly supported, i.e., a thread is allowed to acquire a mutex more
than once before its released. But it then also has be released the same
number of times until the mutex is again lockable by others. When
try is "TRUE" this function never suspends execution.
Instead it returns "FALSE" with "errno" set to
"EBUSY".
- int pth_mutex_release(pth_mutex_t
*mutex);
- This decrements the recursion locking count on mutex
and when it is zero it releases the mutex mutex.
- int pth_rwlock_init(pth_rwlock_t
*rwlock);
- This dynamically initializes a read-write lock variable of
type `"pth_rwlock_t"'. Alternatively one can also use static
initialization via `"pth_rwlock_t rwlock =
PTH_RWLOCK_INIT"'.
- int pth_rwlock_acquire(pth_rwlock_t *rwlock,
int op, int try, pth_event_t ev);
- This acquires a read-only (when op is
"PTH_RWLOCK_RD") or a read-write (when op is
"PTH_RWLOCK_RW") lock rwlock. When the lock is only
locked by other threads in read-only mode, the lock succeeds. But when one
thread holds a read-write lock, all locking attempts suspend the current
thread until this lock is released again. Additionally in ev events
can be given to let the locking timeout, etc. When try is
"TRUE" this function never suspends execution. Instead it
returns "FALSE" with "errno" set to
"EBUSY".
- int pth_rwlock_release(pth_rwlock_t
*rwlock);
- This releases a previously acquired (read-only or
read-write) lock.
- int pth_cond_init(pth_cond_t *cond);
- This dynamically initializes a condition variable variable
of type `"pth_cond_t"'. Alternatively one can also use static
initialization via `"pth_cond_t cond = PTH_COND_INIT"'.
- int pth_cond_await(pth_cond_t *cond,
pth_mutex_t * mutex, pth_event_t ev);
- This awaits a condition situation. The caller has to follow
the semantics of the POSIX condition variables: mutex has to be
acquired before this function is called. The execution of the current
thread is then suspended either until the events in ev occurred
(when ev is not "NULL") or cond was notified by
another thread via pth_cond_notify(3). While the thread is waiting,
mutex is released. Before it returns mutex is
reacquired.
- int pth_cond_notify(pth_cond_t *cond, int
broadcast);
- This notified one or all threads which are waiting on
cond. When broadcast is "TRUE" all thread are
notified, else only a single (unspecified) one.
- int pth_barrier_init(pth_barrier_t *barrier,
int threshold);
- This dynamically initializes a barrier variable of type
`"pth_barrier_t"'. Alternatively one can also use static
initialization via `"pth_barrier_t barrier = PTH_BARRIER_INIT("
threadhold")"'.
- int pth_barrier_reach(pth_barrier_t
*barrier);
- This function reaches a barrier barrier. If this is
the last thread (as specified by threshold on init of
barrier) all threads are awakened. Else the current thread is
suspended until the last thread reached the barrier and this way awakes
all threads. The function returns (beside "FALSE" on error) the
value "TRUE" for any thread which neither reached the barrier as
the first nor the last thread; "PTH_BARRIER_HEADLIGHT" for the
thread which reached the barrier as the first thread and
"PTH_BARRIER_TAILLIGHT" for the thread which reached the barrier
as the last thread.
User-Space Context
The following functions provide a stand-alone sub-API for user-space context
switching. It internally is based on the same underlying machine context
switching mechanism the threads in
GNU Pth are based on. Hence these
functions you can use for implementing your own simple user-space threads. The
"pth_uctx_t" context is somewhat modeled after POSIX
ucontext(3).
The time required to create (via
pth_uctx_make(3)) a user-space context
can range from just a few microseconds up to a more dramatical time (depending
on the machine context switching method which is available on the platform).
On the other hand, the raw performance in switching the user-space contexts is
always very good (nearly independent of the used machine context switching
method). For instance, on an Intel Pentium-III CPU with 800Mhz running under
FreeBSD 4 one usually achieves about 260,000 user-space context switches (via
pth_uctx_switch(3)) per second.
- int pth_uctx_create(pth_uctx_t *uctx);
- This function creates a user-space context and stores it
into uctx. There is still no underlying user-space context
configured. You still have to do this with pth_uctx_make(3). On
success, this function returns "TRUE", else
"FALSE".
- int pth_uctx_make(pth_uctx_t uctx, char
*sk_addr, size_t sk_size, const sigset_t *sigmask, void
(* start_func)(void *), void *start_arg, pth_uctx_t
uctx_after);
- This function makes a new user-space context in uctx
which will operate on the run-time stack sk_addr (which is of
maximum size sk_size), with the signals in sigmask blocked
(if sigmask is not "NULL") and starting to execute with
the call start_func(start_arg). If sk_addr is
"NULL", a stack is dynamically allocated. The stack size
sk_size has to be at least 16384 (16KB). If the start function
start_func returns and uctx_after is not "NULL",
an implicit user-space context switch to this context is performed. Else
(if uctx_after is "NULL") the process is terminated with
exit(3). This function is somewhat modeled after POSIX
makecontext(3). On success, this function returns "TRUE",
else "FALSE".
- int pth_uctx_switch(pth_uctx_t uctx_from,
pth_uctx_t uctx_to);
- This function saves the current user-space context in
uctx_from for later restoring by another call to
pth_uctx_switch(3) and restores the new user-space context from
uctx_to, which previously had to be set with either a previous call
to pth_uctx_switch(3) or initially by pth_uctx_make(3). This
function is somewhat modeled after POSIX swapcontext(3). If
uctx_from or uctx_to are "NULL" or if
uctx_to contains no valid user-space context, "FALSE" is
returned instead of "TRUE". These are the only errors
possible.
- int pth_uctx_destroy(pth_uctx_t uctx);
- This function destroys the user-space context in
uctx. The run-time stack associated with the user-space context is
deallocated only if it was not given by the application (see
sk_addr of pth_uctx_create(3)). If uctx is
"NULL", "FALSE" is returned instead of
"TRUE". This is the only error possible.
Generalized POSIX Replacement API
The following functions are generalized replacements functions for the POSIX
API, i.e., they are similar to the functions under `
Standard POSIX
Replacement API' but all have an additional event argument which can be
used for timeouts, etc.
- int pth_sigwait_ev(const sigset_t *set, int
*sig, pth_event_t ev);
- This is equal to pth_sigwait(3) (see below), but has
an additional event argument ev. When pth_sigwait(3)
suspends the current threads execution it usually only uses the signal
event on set to awake. With this function any number of extra
events can be used to awake the current thread (remember that ev
actually is an event ring).
- int pth_connect_ev(int s, const struct
sockaddr * addr, socklen_t addrlen, pth_event_t
ev);
- This is equal to pth_connect(3) (see below), but has
an additional event argument ev. When pth_connect(3)
suspends the current threads execution it usually only uses the I/O event
on s to awake. With this function any number of extra events can be
used to awake the current thread (remember that ev actually is an
event ring).
- int pth_accept_ev(int s, struct sockaddr
*addr, socklen_t * addrlen, pth_event_t ev);
- This is equal to pth_accept(3) (see below), but has
an additional event argument ev. When pth_accept(3) suspends
the current threads execution it usually only uses the I/O event on
s to awake. With this function any number of extra events can be
used to awake the current thread (remember that ev actually is an
event ring).
- int pth_select_ev(int nfd, fd_set
*rfds, fd_set * wfds, fd_set *efds, struct timeval
*timeout, pth_event_t ev);
- This is equal to pth_select(3) (see below), but has
an additional event argument ev. When pth_select(3) suspends
the current threads execution it usually only uses the I/O event on
rfds, wfds and efds to awake. With this function any
number of extra events can be used to awake the current thread (remember
that ev actually is an event ring).
- int pth_poll_ev(struct pollfd *fds, unsigned
int nfd, int timeout, pth_event_t ev);
- This is equal to pth_poll(3) (see below), but has an
additional event argument ev. When pth_poll(3) suspends the
current threads execution it usually only uses the I/O event on fds
to awake. With this function any number of extra events can be used to
awake the current thread (remember that ev actually is an event
ring).
- ssize_t pth_read_ev(int fd, void *buf,
size_t nbytes, pth_event_t ev);
- This is equal to pth_read(3) (see below), but has an
additional event argument ev. When pth_read(3) suspends the
current threads execution it usually only uses the I/O event on fd
to awake. With this function any number of extra events can be used to
awake the current thread (remember that ev actually is an event
ring).
- ssize_t pth_readv_ev(int fd, const struct
iovec * iovec, int iovcnt, pth_event_t ev);
- This is equal to pth_readv(3) (see below), but has
an additional event argument ev. When pth_readv(3) suspends
the current threads execution it usually only uses the I/O event on
fd to awake. With this function any number of extra events can be
used to awake the current thread (remember that ev actually is an
event ring).
- ssize_t pth_write_ev(int fd, const void
*buf, size_t nbytes, pth_event_t ev);
- This is equal to pth_write(3) (see below), but has
an additional event argument ev. When pth_write(3) suspends
the current threads execution it usually only uses the I/O event on
fd to awake. With this function any number of extra events can be
used to awake the current thread (remember that ev actually is an
event ring).
- ssize_t pth_writev_ev(int fd, const struct
iovec * iovec, int iovcnt, pth_event_t ev);
- This is equal to pth_writev(3) (see below), but has
an additional event argument ev. When pth_writev(3) suspends
the current threads execution it usually only uses the I/O event on
fd to awake. With this function any number of extra events can be
used to awake the current thread (remember that ev actually is an
event ring).
- ssize_t pth_recv_ev(int fd, void *buf,
size_t nbytes, int flags, pth_event_t ev);
- This is equal to pth_recv(3) (see below), but has an
additional event argument ev. When pth_recv(3) suspends the
current threads execution it usually only uses the I/O event on fd
to awake. With this function any number of extra events can be used to
awake the current thread (remember that ev actually is an event
ring).
- ssize_t pth_recvfrom_ev(int fd, void
*buf, size_t nbytes, int flags, struct sockaddr
*from, socklen_t * fromlen, pth_event_t ev);
- This is equal to pth_recvfrom(3) (see below), but
has an additional event argument ev. When pth_recvfrom(3)
suspends the current threads execution it usually only uses the I/O event
on fd to awake. With this function any number of extra events can
be used to awake the current thread (remember that ev actually is
an event ring).
- ssize_t pth_send_ev(int fd, const void
*buf, size_t nbytes, int flags, pth_event_t
ev);
- This is equal to pth_send(3) (see below), but has an
additional event argument ev. When pth_send(3) suspends the
current threads execution it usually only uses the I/O event on fd
to awake. With this function any number of extra events can be used to
awake the current thread (remember that ev actually is an event
ring).
- ssize_t pth_sendto_ev(int fd, const void
*buf, size_t nbytes, int flags, const struct sockaddr
*to, socklen_t tolen, pth_event_t ev);
- This is equal to pth_sendto(3) (see below), but has
an additional event argument ev. When pth_sendto(3) suspends
the current threads execution it usually only uses the I/O event on
fd to awake. With this function any number of extra events can be
used to awake the current thread (remember that ev actually is an
event ring).
Standard POSIX Replacement API
The following functions are standard replacements functions for the POSIX API.
The difference is mainly that they suspend the current thread only instead of
the whole process in case the file descriptors will block.
- int pth_nanosleep(const struct timespec
*rqtp, struct timespec * rmtp);
- This is a variant of the POSIX nanosleep(3)
function. It suspends the current threads execution until the amount of
time in rqtp elapsed. The thread is guaranteed to not wake up
before this time, but because of the non-preemptive scheduling nature of
Pth, it can be awakened later, of course. If rmtp is not
"NULL", the "timespec" structure it references is
updated to contain the unslept amount (the request time minus the time
actually slept time). The difference between nanosleep(3) and
pth_nanosleep(3) is that that pth_nanosleep(3) suspends only
the execution of the current thread and not the whole process.
- int pth_usleep(unsigned int usec);
- This is a variant of the 4.3BSD usleep(3) function.
It suspends the current threads execution until usec microseconds
(= usec*1/1000000 sec) elapsed. The thread is guaranteed to not
wake up before this time, but because of the non-preemptive scheduling
nature of Pth, it can be awakened later, of course. The difference
between usleep(3) and pth_usleep(3) is that that
pth_usleep(3) suspends only the execution of the current thread and
not the whole process.
- unsigned int pth_sleep(unsigned int
sec);
- This is a variant of the POSIX sleep(3) function. It
suspends the current threads execution until sec seconds elapsed.
The thread is guaranteed to not wake up before this time, but because of
the non-preemptive scheduling nature of Pth, it can be awakened
later, of course. The difference between sleep(3) and
pth_sleep(3) is that pth_sleep(3) suspends only the
execution of the current thread and not the whole process.
- pid_t pth_waitpid(pid_t pid, int
*status, int options);
- This is a variant of the POSIX waitpid(2) function.
It suspends the current threads execution until status information
is available for a terminated child process pid. The difference
between waitpid(2) and pth_waitpid(3) is that
pth_waitpid(3) suspends only the execution of the current thread
and not the whole process. For more details about the arguments and return
code semantics see waitpid(2).
- int pth_system(const char *cmd);
- This is a variant of the POSIX system(3) function.
It executes the shell command cmd with Bourne Shell
("sh") and suspends the current threads execution until this
command terminates. The difference between system(3) and
pth_system(3) is that pth_system(3) suspends only the
execution of the current thread and not the whole process. For more
details about the arguments and return code semantics see
system(3).
- int pth_sigmask(int how, const sigset_t
*set, sigset_t * oset)
- This is the Pth thread-related equivalent of POSIX
sigprocmask(2) respectively pthread_sigmask(3). The
arguments how, set and oset directly relate to
sigprocmask(2), because Pth internally just uses
sigprocmask(2) here. So alternatively you can also directly call
sigprocmask(2), but for consistency reasons you should use this
function pth_sigmask(3).
- int pth_sigwait(const sigset_t *set, int
*sig);
- This is a variant of the POSIX.1c sigwait(3)
function. It suspends the current threads execution until a signal in
set occurred and stores the signal number in sig. The
important point is that the signal is not delivered to a signal handler.
Instead it's caught by the scheduler only in order to awake the
pth_sigwait() call. The trick and noticeable point here is that
this way you get an asynchronous aware application that is written
completely synchronously. When you think about the problem of
asynchronous safe functions you should recognize that this is a
great benefit.
- int pth_connect(int s, const struct sockaddr
* addr, socklen_t addrlen);
- This is a variant of the 4.2BSD connect(2) function.
It establishes a connection on a socket s to target specified in
addr and addrlen. The difference between connect(2)
and pth_connect(3) is that pth_connect(3) suspends only the
execution of the current thread and not the whole process. For more
details about the arguments and return code semantics see
connect(2).
- int pth_accept(int s, struct sockaddr
*addr, socklen_t * addrlen);
- This is a variant of the 4.2BSD accept(2) function.
It accepts a connection on a socket by extracting the first connection
request on the queue of pending connections, creating a new socket with
the same properties of s and allocates a new file descriptor for
the socket (which is returned). The difference between accept(2)
and pth_accept(3) is that pth_accept(3) suspends only the
execution of the current thread and not the whole process. For more
details about the arguments and return code semantics see
accept(2).
- int pth_select(int nfd, fd_set *rfds,
fd_set * wfds, fd_set *efds, struct timeval
*timeout);
- This is a variant of the 4.2BSD select(2) function.
It examines the I/O descriptor sets whose addresses are passed in
rfds, wfds, and efds to see if some of their
descriptors are ready for reading, are ready for writing, or have an
exceptional condition pending, respectively. For more details about the
arguments and return code semantics see select(2).
- int pth_pselect(int nfd, fd_set *rfds,
fd_set * wfds, fd_set *efds, const struct timespec
*timeout, const sigset_t * sigmask);
- This is a variant of the POSIX pselect(2) function,
which in turn is a stronger variant of 4.2BSD select(2). The
difference is that the higher-resolution "struct timespec" is
passed instead of the lower-resolution "struct timeval" and that
a signal mask is specified which is temporarily set while waiting for
input. For more details about the arguments and return code semantics see
pselect(2) and select(2).
- int pth_poll(struct pollfd *fds, unsigned int
nfd, int timeout);
- This is a variant of the SysV poll(2) function. It
examines the I/O descriptors which are passed in the array fds to
see if some of them are ready for reading, are ready for writing, or have
an exceptional condition pending, respectively. For more details about the
arguments and return code semantics see poll(2).
- ssize_t pth_read(int fd, void *buf,
size_t nbytes);
- This is a variant of the POSIX read(2) function. It
reads up to nbytes bytes into buf from file descriptor
fd. The difference between read(2) and pth_read(2) is
that pth_read(2) suspends execution of the current thread until the
file descriptor is ready for reading. For more details about the arguments
and return code semantics see read(2).
- ssize_t pth_readv(int fd, const struct iovec
* iovec, int iovcnt);
- This is a variant of the POSIX readv(2) function. It
reads data from file descriptor fd into the first iovcnt
rows of the iov vector. The difference between readv(2) and
pth_readv(2) is that pth_readv(2) suspends execution of the
current thread until the file descriptor is ready for reading. For more
details about the arguments and return code semantics see
readv(2).
- ssize_t pth_write(int fd, const void
*buf, size_t nbytes);
- This is a variant of the POSIX write(2) function. It
writes nbytes bytes from buf to file descriptor fd.
The difference between write(2) and pth_write(2) is that
pth_write(2) suspends execution of the current thread until the
file descriptor is ready for writing. For more details about the arguments
and return code semantics see write(2).
- ssize_t pth_writev(int fd, const struct iovec
* iovec, int iovcnt);
- This is a variant of the POSIX writev(2) function.
It writes data to file descriptor fd from the first iovcnt
rows of the iov vector. The difference between writev(2) and
pth_writev(2) is that pth_writev(2) suspends execution of
the current thread until the file descriptor is ready for reading. For
more details about the arguments and return code semantics see
writev(2).
- ssize_t pth_pread(int fd, void *buf,
size_t nbytes, off_t offset);
- This is a variant of the POSIX pread(3) function. It
performs the same action as a regular read(2), except that it reads
from a given position in the file without changing the file pointer. The
first three arguments are the same as for pth_read(3) with the
addition of a fourth argument offset for the desired position
inside the file.
- ssize_t pth_pwrite(int fd, const void
*buf, size_t nbytes, off_t offset);
- This is a variant of the POSIX pwrite(3) function.
It performs the same action as a regular write(2), except that it
writes to a given position in the file without changing the file pointer.
The first three arguments are the same as for pth_write(3) with the
addition of a fourth argument offset for the desired position
inside the file.
- ssize_t pth_recv(int fd, void *buf,
size_t nbytes, int flags);
- This is a variant of the SUSv2 recv(2) function and
equal to ``pth_recvfrom(fd, buf, nbytes, flags, NULL, 0)''.
- ssize_t pth_recvfrom(int fd, void
*buf, size_t nbytes, int flags, struct sockaddr
*from, socklen_t * fromlen);
- This is a variant of the SUSv2 recvfrom(2) function.
It reads up to nbytes bytes into buf from file descriptor
fd while using flags and from/fromlen. The
difference between recvfrom(2) and pth_recvfrom(2) is that
pth_recvfrom(2) suspends execution of the current thread until the
file descriptor is ready for reading. For more details about the arguments
and return code semantics see recvfrom(2).
- ssize_t pth_send(int fd, const void
*buf, size_t nbytes, int flags);
- This is a variant of the SUSv2 send(2) function and
equal to ``pth_sendto(fd, buf, nbytes, flags, NULL, 0)''.
- ssize_t pth_sendto(int fd, const void
*buf, size_t nbytes, int flags, const struct sockaddr
*to, socklen_t tolen);
- This is a variant of the SUSv2 sendto(2) function.
It writes nbytes bytes from buf to file descriptor fd
while using flags and to/tolen. The difference
between sendto(2) and pth_sendto(2) is that
pth_sendto(2) suspends execution of the current thread until the
file descriptor is ready for writing. For more details about the arguments
and return code semantics see sendto(2).
EXAMPLE¶
The following example is a useless server which does nothing more than listening
on TCP port 12345 and displaying the current time to the socket when a
connection was established. For each incoming connection a thread is spawned.
Additionally, to see more multithreading, a useless ticker thread runs
simultaneously which outputs the current time to "stderr" every 5
seconds. The example contains
no error checking and is
only
intended to show you the look and feel of
Pth.
#include <stdio.h>
#include <stdlib.h>
#include <errno.h>
#include <sys/types.h>
#include <sys/socket.h>
#include <netinet/in.h>
#include <arpa/inet.h>
#include <signal.h>
#include <netdb.h>
#include <unistd.h>
#include "pth.h"
#define PORT 12345
/* the socket connection handler thread */
static void *handler(void *_arg)
{
int fd = (int)_arg;
time_t now;
char *ct;
now = time(NULL);
ct = ctime(&now);
pth_write(fd, ct, strlen(ct));
close(fd);
return NULL;
}
/* the stderr time ticker thread */
static void *ticker(void *_arg)
{
time_t now;
char *ct;
float load;
for (;;) {
pth_sleep(5);
now = time(NULL);
ct = ctime(&now);
ct[strlen(ct)-1] = '\0';
pth_ctrl(PTH_CTRL_GETAVLOAD, &load);
printf("ticker: time: %s, average load: %.2f\n", ct, load);
}
}
/* the main thread/procedure */
int main(int argc, char *argv[])
{
pth_attr_t attr;
struct sockaddr_in sar;
struct protoent *pe;
struct sockaddr_in peer_addr;
int peer_len;
int sa, sw;
int port;
pth_init();
signal(SIGPIPE, SIG_IGN);
attr = pth_attr_new();
pth_attr_set(attr, PTH_ATTR_NAME, "ticker");
pth_attr_set(attr, PTH_ATTR_STACK_SIZE, 64*1024);
pth_attr_set(attr, PTH_ATTR_JOINABLE, FALSE);
pth_spawn(attr, ticker, NULL);
pe = getprotobyname("tcp");
sa = socket(AF_INET, SOCK_STREAM, pe->p_proto);
sar.sin_family = AF_INET;
sar.sin_addr.s_addr = INADDR_ANY;
sar.sin_port = htons(PORT);
bind(sa, (struct sockaddr *)&sar, sizeof(struct sockaddr_in));
listen(sa, 10);
pth_attr_set(attr, PTH_ATTR_NAME, "handler");
for (;;) {
peer_len = sizeof(peer_addr);
sw = pth_accept(sa, (struct sockaddr *)&peer_addr, &peer_len);
pth_spawn(attr, handler, (void *)sw);
}
}
BUILD ENVIRONMENTS¶
In this section we will discuss the canonical ways to establish the build
environment for a
Pth based program. The possibilities supported by
Pth range from very simple environments to rather complex ones.
Manual Build Environment (Novice)
As a first example, assume we have the above test program staying in the source
file "foo.c". Then we can create a very simple build environment by
just adding the following "Makefile":
$ vi Makefile
⎪ CC = cc
⎪ CFLAGS = `pth-config --cflags`
⎪ LDFLAGS = `pth-config --ldflags`
⎪ LIBS = `pth-config --libs`
⎪
⎪ all: foo
⎪ foo: foo.o
⎪ $(CC) $(LDFLAGS) -o foo foo.o $(LIBS)
⎪ foo.o: foo.c
⎪ $(CC) $(CFLAGS) -c foo.c
⎪ clean:
⎪ rm -f foo foo.o
This imports the necessary compiler and linker flags on-the-fly from the
Pth installation via its "pth-config" program. This approach
is straight-forward and works fine for small projects.
Autoconf Build Environment (Advanced)
The previous approach is simple but inflexible. First, to speed up building, it
would be nice to not expand the compiler and linker flags every time the
compiler is started. Second, it would be useful to also be able to build
against uninstalled
Pth, that is, against a
Pth source tree
which was just configured and built, but not installed. Third, it would be
also useful to allow checking of the
Pth version to make sure it is at
least a minimum required version. And finally, it would be also great to make
sure
Pth works correctly by first performing some sanity compile and
run-time checks. All this can be done if we use GNU
autoconf and the
"AC_CHECK_PTH" macro provided by
Pth. For this, we establish
the following three files:
First we again need the "Makefile", but this time it contains
autoconf placeholders and additional cleanup targets. And we create it
under the name "Makefile.in", because it is now an input file for
autoconf:
$ vi Makefile.in
⎪ CC = @CC@
⎪ CFLAGS = @CFLAGS@
⎪ LDFLAGS = @LDFLAGS@
⎪ LIBS = @LIBS@
⎪
⎪ all: foo
⎪ foo: foo.o
⎪ $(CC) $(LDFLAGS) -o foo foo.o $(LIBS)
⎪ foo.o: foo.c
⎪ $(CC) $(CFLAGS) -c foo.c
⎪ clean:
⎪ rm -f foo foo.o
⎪ distclean:
⎪ rm -f foo foo.o
⎪ rm -f config.log config.status config.cache
⎪ rm -f Makefile
Because
autoconf generates additional files, we added a canonical
"distclean" target which cleans this up. Secondly, we wrote
"configure.ac", a (minimal)
autoconf script specification:
$ vi configure.ac
⎪ AC_INIT(Makefile.in)
⎪ AC_CHECK_PTH(1.3.0)
⎪ AC_OUTPUT(Makefile)
Then we let
autoconf's "aclocal" program generate for us an
"aclocal.m4" file containing
Pth's "AC_CHECK_PTH"
macro. Then we generate the final "configure" script out of this
"aclocal.m4" file and the "configure.ac" file:
$ aclocal --acdir=`pth-config --acdir`
$ autoconf
After these steps, the working directory should look similar to this:
$ ls -l
-rw-r--r-- 1 rse users 176 Nov 3 11:11 Makefile.in
-rw-r--r-- 1 rse users 15314 Nov 3 11:16 aclocal.m4
-rwxr-xr-x 1 rse users 52045 Nov 3 11:16 configure
-rw-r--r-- 1 rse users 63 Nov 3 11:11 configure.ac
-rw-r--r-- 1 rse users 4227 Nov 3 11:11 foo.c
If we now run "configure" we get a correct "Makefile" which
immediately can be used to build "foo" (assuming that
Pth is
already installed somewhere, so that "pth-config" is in $PATH):
$ ./configure
creating cache ./config.cache
checking for gcc... gcc
checking whether the C compiler (gcc ) works... yes
checking whether the C compiler (gcc ) is a cross-compiler... no
checking whether we are using GNU C... yes
checking whether gcc accepts -g... yes
checking how to run the C preprocessor... gcc -E
checking for GNU Pth... version 1.3.0, installed under /usr/local
updating cache ./config.cache
creating ./config.status
creating Makefile
rse@en1:/e/gnu/pth/ac
$ make
gcc -g -O2 -I/usr/local/include -c foo.c
gcc -L/usr/local/lib -o foo foo.o -lpth
If
Pth is installed in non-standard locations or "pth-config"
is not in $PATH, one just has to drop the "configure" script a note
about the location by running "configure" with the option
"--with-pth="
dir (where
dir is the argument which was
used with the "--prefix" option when
Pth was installed).
Autoconf Build Environment with Local Copy of Pth (Expert)
Finally let us assume the "foo" program stays under either a
GPL or
LGPL distribution license and we want to make it a
stand-alone package for easier distribution and installation. That is, we
don't want to oblige the end-user to install
Pth just to allow our
"foo" package to compile. For this, it is a convenient practice to
include the required libraries (here
Pth) into the source tree of the
package (here "foo").
Pth ships with all necessary support to
allow us to easily achieve this approach. Say, we want
Pth in a
subdirectory named "pth/" and this directory should be seamlessly
integrated into the configuration and build process of "foo".
First we again start with the "Makefile.in", but this time it is a
more advanced version which supports subdirectory movement:
$ vi Makefile.in
⎪ CC = @CC@
⎪ CFLAGS = @CFLAGS@
⎪ LDFLAGS = @LDFLAGS@
⎪ LIBS = @LIBS@
⎪
⎪ SUBDIRS = pth
⎪
⎪ all: subdirs_all foo
⎪
⎪ subdirs_all:
⎪ @$(MAKE) $(MFLAGS) subdirs TARGET=all
⎪ subdirs_clean:
⎪ @$(MAKE) $(MFLAGS) subdirs TARGET=clean
⎪ subdirs_distclean:
⎪ @$(MAKE) $(MFLAGS) subdirs TARGET=distclean
⎪ subdirs:
⎪ @for subdir in $(SUBDIRS); do \
⎪ echo "===> $$subdir ($(TARGET))"; \
⎪ (cd $$subdir; $(MAKE) $(MFLAGS) $(TARGET) ⎪⎪ exit 1) ⎪⎪ exit 1; \
⎪ echo "<=== $$subdir"; \
⎪ done
⎪
⎪ foo: foo.o
⎪ $(CC) $(LDFLAGS) -o foo foo.o $(LIBS)
⎪ foo.o: foo.c
⎪ $(CC) $(CFLAGS) -c foo.c
⎪
⎪ clean: subdirs_clean
⎪ rm -f foo foo.o
⎪ distclean: subdirs_distclean
⎪ rm -f foo foo.o
⎪ rm -f config.log config.status config.cache
⎪ rm -f Makefile
Then we create a slightly different
autoconf script
"configure.ac":
$ vi configure.ac
⎪ AC_INIT(Makefile.in)
⎪ AC_CONFIG_AUX_DIR(pth)
⎪ AC_CHECK_PTH(1.3.0, subdir:pth --disable-tests)
⎪ AC_CONFIG_SUBDIRS(pth)
⎪ AC_OUTPUT(Makefile)
Here we provided a default value for "foo"'s "--with-pth"
option as the second argument to "AC_CHECK_PTH" which indicates that
Pth can be found in the subdirectory named "pth/".
Additionally we specified that the "--disable-tests" option of
Pth should be passed to the "pth/" subdirectory, because we
need only to build the
Pth library itself. And we added a
"AC_CONFIG_SUBDIR" call which indicates to
autoconf that it
should configure the "pth/" subdirectory, too. The
"AC_CONFIG_AUX_DIR" directive was added just to make
autoconf
happy, because it wants to find a "install.sh" or "shtool"
script if "AC_CONFIG_SUBDIRS" is used.
Now we let
autoconf's "aclocal" program again generate for us
an "aclocal.m4" file with the contents of
Pth's
"AC_CHECK_PTH" macro. Finally we generate the "configure"
script out of this "aclocal.m4" file and the
"configure.ac" file.
$ aclocal --acdir=`pth-config --acdir`
$ autoconf
Now we have to create the "pth/" subdirectory itself. For this, we
extract the
Pth distribution to the "foo" source tree and
just rename it to "pth/":
$ gunzip <pth-X.Y.Z.tar.gz ⎪ tar xvf -
$ mv pth-X.Y.Z pth
Optionally to reduce the size of the "pth/" subdirectory, we can strip
down the
Pth sources to a minimum with the
striptease feature:
$ cd pth
$ ./configure
$ make striptease
$ cd ..
After this the source tree of "foo" should look similar to this:
$ ls -l
-rw-r--r-- 1 rse users 709 Nov 3 11:51 Makefile.in
-rw-r--r-- 1 rse users 16431 Nov 3 12:20 aclocal.m4
-rwxr-xr-x 1 rse users 57403 Nov 3 12:21 configure
-rw-r--r-- 1 rse users 129 Nov 3 12:21 configure.ac
-rw-r--r-- 1 rse users 4227 Nov 3 11:11 foo.c
drwxr-xr-x 2 rse users 3584 Nov 3 12:36 pth
$ ls -l pth/
-rw-rw-r-- 1 rse users 26344 Nov 1 20:12 COPYING
-rw-rw-r-- 1 rse users 2042 Nov 3 12:36 Makefile.in
-rw-rw-r-- 1 rse users 3967 Nov 1 19:48 README
-rw-rw-r-- 1 rse users 340 Nov 3 12:36 README.1st
-rw-rw-r-- 1 rse users 28719 Oct 31 17:06 config.guess
-rw-rw-r-- 1 rse users 24274 Aug 18 13:31 config.sub
-rwxrwxr-x 1 rse users 155141 Nov 3 12:36 configure
-rw-rw-r-- 1 rse users 162021 Nov 3 12:36 pth.c
-rw-rw-r-- 1 rse users 18687 Nov 2 15:19 pth.h.in
-rw-rw-r-- 1 rse users 5251 Oct 31 12:46 pth_acdef.h.in
-rw-rw-r-- 1 rse users 2120 Nov 1 11:27 pth_acmac.h.in
-rw-rw-r-- 1 rse users 2323 Nov 1 11:27 pth_p.h.in
-rw-rw-r-- 1 rse users 946 Nov 1 11:27 pth_vers.c
-rw-rw-r-- 1 rse users 26848 Nov 1 11:27 pthread.c
-rw-rw-r-- 1 rse users 18772 Nov 1 11:27 pthread.h.in
-rwxrwxr-x 1 rse users 26188 Nov 3 12:36 shtool
Now when we configure and build the "foo" package it looks similar to
this:
$ ./configure
creating cache ./config.cache
checking for gcc... gcc
checking whether the C compiler (gcc ) works... yes
checking whether the C compiler (gcc ) is a cross-compiler... no
checking whether we are using GNU C... yes
checking whether gcc accepts -g... yes
checking how to run the C preprocessor... gcc -E
checking for GNU Pth... version 1.3.0, local under pth
updating cache ./config.cache
creating ./config.status
creating Makefile
configuring in pth
running /bin/sh ./configure --enable-subdir --enable-batch
--disable-tests --cache-file=.././config.cache --srcdir=.
loading cache .././config.cache
checking for gcc... (cached) gcc
checking whether the C compiler (gcc ) works... yes
checking whether the C compiler (gcc ) is a cross-compiler... no
[...]
$ make
===> pth (all)
./shtool scpp -o pth_p.h -t pth_p.h.in -Dcpp -Cintern -M '==#==' pth.c
pth_vers.c
gcc -c -I. -O2 -pipe pth.c
gcc -c -I. -O2 -pipe pth_vers.c
ar rc libpth.a pth.o pth_vers.o
ranlib libpth.a
<=== pth
gcc -g -O2 -Ipth -c foo.c
gcc -Lpth -o foo foo.o -lpth
As you can see,
autoconf now automatically configures the local (stripped
down) copy of
Pth in the subdirectory "pth/" and the
"Makefile" automatically builds the subdirectory, too.
SYSTEM CALL WRAPPER FACILITY¶
Pth per default uses an explicit API, including the system calls. For
instance you've to explicitly use
pth_read(3) when you need a
thread-aware
read(3) and cannot expect that by just calling
read(3) only the current thread is blocked. Instead with the standard
read(3) call the whole process will be blocked. But because for some
applications (mainly those consisting of lots of third-party stuff) this can
be inconvenient. Here it's required that a call to
read(3) `magically'
means
pth_read(3). The problem here is that such magic
Pth
cannot provide per default because it's not really portable. Nevertheless
Pth provides a two step approach to solve this problem:
Soft System Call Mapping
This variant is available on all platforms and can
always be enabled by
building
Pth with "--enable-syscall-soft". This then triggers
some "#define"'s in the "pth.h" header which map for
instance
read(3) to
pth_read(3), etc. Currently the following
functions are mapped:
fork(2),
nanosleep(3),
usleep(3),
sleep(3),
sigwait(3),
waitpid(2),
system(3),
select(2),
poll(2),
connect(2),
accept(2),
read(2),
write(2),
recv(2),
send(2),
recvfrom(2),
sendto(2).
The drawback of this approach is just that really all source files of the
application where these function calls occur have to include
"pth.h", of course. And this also means that existing libraries,
including the vendor's
stdio, usually will still block the whole
process if one of its I/O functions block.
Hard System Call Mapping
This variant is available only on those platforms where the
syscall(2)
function exists and there it can be enabled by building
Pth with
"--enable-syscall-hard". This then builds wrapper functions (for
instances
read(3)) into the
Pth library which internally call
the real
Pth replacement functions (
pth_read(3)). Currently the
following functions are mapped:
fork(2),
nanosleep(3),
usleep(3),
sleep(3),
waitpid(2),
system(3),
select(2),
poll(2),
connect(2),
accept(2),
read(2),
write(2).
The drawback of this approach is that it depends on
syscall(2) interface
and prototype conflicts can occur while building the wrapper functions due to
different function signatures in the vendor C header files. But the advantage
of this mapping variant is that the source files of the application where
these function calls occur have not to include "pth.h" and that
existing libraries, including the vendor's
stdio, magically become
thread-aware (and then block only the current thread).
IMPLEMENTATION NOTES¶
Pth is very portable because it has only one part which perhaps has to be
ported to new platforms (the machine context initialization). But it is
written in a way which works on mostly all Unix platforms which support
makecontext(2) or at least
sigstack(2) or
sigaltstack(2)
[see "pth_mctx.c" for details]. Any other
Pth code is POSIX
and ANSI C based only.
The context switching is done via either SUSv2
makecontext(2) or POSIX
make[sig]
setjmp(3) and [sig]
longjmp(3). Here all CPU registers,
the program counter and the stack pointer are switched. Additionally the
Pth dispatcher switches also the global Unix "errno" variable
[see "pth_mctx.c" for details] and the signal mask (either
implicitly via
sigsetjmp(3) or in an emulated way via explicit
setprocmask(2) calls).
The
Pth event manager is mainly
select(2) and
gettimeofday(2) based, i.e., the current time is fetched via
gettimeofday(2) once per context switch for time calculations and all
I/O events are implemented via a single central
select(2) call [see
"pth_sched.c" for details].
The thread control block management is done via virtual priority queues without
any additional data structure overhead. For this, the queue linkage attributes
are part of the thread control blocks and the queues are actually implemented
as rings with a selected element as the entry point [see "pth_tcb.h"
and "pth_pqueue.c" for details].
Most time critical code sections (especially the dispatcher and event manager)
are speeded up by inline functions (implemented as ANSI C pre-processor
macros). Additionally any debugging code is
completely removed from the
source when not built with "-DPTH_DEBUG" (see Autoconf
"--enable-debug" option), i.e., not only stub functions remain [see
"pth_debug.c" for details].
RESTRICTIONS¶
Pth (intentionally) provides no replacements for non-thread-safe
functions (like
strtok(3) which uses a static internal buffer) or
synchronous system functions (like
gethostbyname(3) which doesn't
provide an asynchronous mode where it doesn't block). When you want to use
those functions in your server application together with threads, you've to
either link the application against special third-party libraries (or for
thread-safe/reentrant functions possibly against an existing
"libc_r" of the platform vendor). For an asynchronous DNS resolver
library use the GNU
adns package from Ian Jackson ( see
http://www.gnu.org/software/adns/adns.html ).
HISTORY¶
The
Pth library was designed and implemented between February and July
1999 by
Ralf S. Engelschall after evaluating numerous (mostly
preemptive) thread libraries and after intensive discussions with
Peter
Simons,
Martin Kraemer,
Lars Eilebrecht and
Ralph
Babel related to an experimental (matrix based) non-preemptive C++
scheduler class written by
Peter Simons.
Pth was then implemented in order to combine the
non-preemptive
approach of multithreading (which provides better portability and performance)
with an API similar to the popular one found in
Pthread libraries
(which provides easy programming).
So the essential idea of the non-preemptive approach was taken over from
Peter Simons scheduler. The priority based scheduling algorithm was
suggested by
Martin Kraemer. Some code inspiration also came from an
experimental threading library (
rsthreads) written by
Robert
S. Thau for an ancient internal test version of the Apache webserver.
The concept and API of message ports was borrowed from AmigaOS'
Exec
subsystem. The concept and idea for the flexible event mechanism came from
Paul Vixie's
eventlib (which can be found as a part of
BIND v8).
BUG REPORTS AND SUPPORT¶
If you think you have found a bug in
Pth, you should send a report as
complete as possible to
bug-pth@gnu.org. If you can, please try to fix
the problem and include a patch, made with '"diff -u3"', in your
report. Always, at least, include a reasonable amount of description in your
report to allow the author to deterministically reproduce the bug.
For further support you additionally can subscribe to the
pth-users@gnu.org mailing list by sending an Email to
pth-users-request@gnu.org with `"subscribe pth-users"' (or
`"subscribe pth-users"
address' if you want to subscribe from
a particular Email
address) in the body. Then you can discuss your
issues with other
Pth users by sending messages to
pth-users@gnu.org. Currently (as of August 2000) you can reach about
110 Pth users on this mailing list. Old postings you can find at
http://www.mail-archive.com/pth-users@gnu.org/.
SEE ALSO¶
Related Web Locations
`comp.programming.threads Newsgroup Archive',
http://www.deja.com/topics_if.xp?
search=topic&group=comp.programming.threads
`comp.programming.threads Frequently Asked Questions (F.A.Q.)',
http://www.lambdacs.com/newsgroup/FAQ.html
`
Multithreading - Definitions and Guidelines', Numeric Quest Inc 1998;
http://www.numeric-quest.com/lang/multi-frame.html
`
The Single UNIX Specification, Version 2 - Threads', The Open Group
1997;
http://www.opengroup.org/onlinepubs /007908799/xsh/threads.html
SMI Thread Resources, Sun Microsystems Inc;
http://www.sun.com/workshop/threads/
Bibliography on threads and multithreading, Torsten Amundsen;
http://liinwww.ira.uka.de/bibliography/Os/threads.html
Related Books
B. Nichols, D. Buttlar, J.P. Farrel: `
Pthreads Programming - A POSIX
Standard for Better Multiprocessing', O'Reilly 1996; ISBN 1-56592-115-1
B. Lewis, D. J. Berg: `
Multithreaded Programming with Pthreads', Sun
Microsystems Press, Prentice Hall 1998; ISBN 0-13-680729-1
B. Lewis, D. J. Berg: `
Threads Primer - A Guide To Multithreaded
Programming', Prentice Hall 1996; ISBN 0-13-443698-9
S. J. Norton, M. D. Dipasquale: `
Thread Time - The Multithreaded Programming
Guide', Prentice Hall 1997; ISBN 0-13-190067-6
D. R. Butenhof: `
Programming with POSIX Threads', Addison Wesley 1997;
ISBN 0-201-63392-2
Related Manpages
pth-config(1),
pthread(3).
getcontext(2),
setcontext(2),
makecontext(2),
swapcontext(2),
sigstack(2),
sigaltstack(2),
sigaction(2),
sigemptyset(2),
sigaddset(2),
sigprocmask(2),
sigsuspend(2),
sigsetjmp(3),
siglongjmp(3),
setjmp(3),
longjmp(3),
select(2),
gettimeofday(2).
AUTHOR¶
Ralf S. Engelschall
rse@engelschall.com
www.engelschall.com