One of the strong features of Unix-like operating systems, is their ability to run several processes simultaneously, and let them all share the CPU(s), memory, and other resources. Any none-trivial system developed on Unix systems will sooner or later resort to splitting its tasks into more than one process. True, many times threads will be preferred (thought these are candidates for a separate tutorial), but the methods used in both of them tend to be rather similar - how to start and stop processes, how to communicate with other processes, how to synchronize processes.
We'll try to learn here the various features that the system supplies us with in order to answer these questions. One should note that dealing with multi-process systems takes a slightly different approach than dealing with a single-process program - events happen to occur in parallel, debugging is more complicated, and there's always the risk of having a bug cause endless process creation that'll bring your system to a halt. In fact, when i took a course that tried teaching these subjects, at the week we had to deliver one of the exercises, the system hardly managed to survive the bunch of students running buggy programs that create new processes endlessly, or leaving background processes running into endless loop, and so on. OK, lets stop intimidating, and get to see how it's done.
Before we talk about processes, we need to understand exactly what a process is. If you know exactly what it is, and are familiar with the notion of 'Re-entrancy', you may skip to the next section...
You didn't skip? OK. Lets try to write a proper definition:
As you might understand from this definition, a process is not a program. several processes may be executing the same computer program at the same time, for the same user or for several different users. For example, there is normally one copy of the 'tcsh' shell on the system, but there may be many tcsh processes running - one for each interactive connection of a user to the system. It might be that many different processes will try to execute the same piece of code at the same time, perhaps trying to utilize the same resources, and we should be ready to accommodate such situations. This leads us to the concept of 'Re-entrancy'.
This re-entrancy might mean that two or more processes try to execute this piece of code at the same time. it might also mean that a single process tries to execute the same function several times simultaneously. How this may be possible? a simple example is a recursive function. The process starts executing it, and somewhere in the middle (before exiting the function), it calls the same function again. This means that the function should only use local variables to save its state information, for example.
Of-course, with a multi-process code, we don't have conflicts of variables, because normally the data section of each process is separate from that of other processes (so process A that runs program P and process B that runs the same program P, have distinct copies of the global variable 'i' of that program), but there might be other resources that would cause a piece of code to be non-reentrant. For example, if the program opens a file and writes some data to it, and two processes try to run the program at the same time, the contents of the file might be ruined in an unpredictable way. This is why a program must protect itself by using some kind of 'locking' mechanism, that will only allow one process at a time to open the file and write data into the file. An example of such a mechanism is the usage of semaphores, which will be discussed later on.
As you might (hopefully) already know, it is possible for a user to run a process in the system, suspend it (Ctrl-Z), and move it to the background (using the 'bg' command). If you're not familiar with this, you would do best to read the 'Job Control' section of the 'csh' manual page (or of 'bash', if that is the shell you normally use). However, we are interested in learning how to create new processes from within a C program.
fork()
System Call The fork()
system call is the basic
way to create a new process. It is also a very unique system call, since
it returns twice(!) to the caller. Sounds confusing? good. This confusion
stems from the attempt to define as few systems calls as possible, it seems.
OK, lets see:
fork()
fork()
call, so both parent and child process see the exact
same image. The only distinction is when the call returns. When it returns
in the parent process, its return value is the process ID (PID) of the child
process. When it returns inside the child process, its return value is '0'.
If for some reason this call failed (not enough memory, too many processes,
etc.), no new process is created, and the return value of the call is '-1'.
In case the process was created successfully, both child process and parent
process continue from the same place in the code where the fork()
call was used. To make things clearer, lets see an example of a code that uses this system call to create a child process that prints (you guessed it) "hello world" to the screen, and exits.
#include <unistd.h> /* defines fork(), and pid_t. */
#include <sys/wait.h> /* defines the wait() system call. */
/* storage place for the pid of the child process, and its exit status. */
pid_t child_pid;
int child_status;
/* lets fork off a child process... */
child_pid = fork();
/* check what the fork() call actually did */
switch (child_pid) {
case -1: /* fork() failed */
perror("fork"); /* print a system-defined error message */
exit(1);
case 0: /* fork() succeeded, we're inside the child process */
printf("hello world\n");
exit(0); /* here the CHILD process exits, not the parent. */
default: /* fork() succeeded, we're inside the parent process */
wait(&child_status); /* wait till the child process exits */
}
/* parent's process code may continue here... */
Notes:
perror()
function prints an error message based on the value of the errno variable,
to stderr. wait()
system call waits until
any child process exits, and stores its exit status in the variable
supplied. There are a set of macros to check this status, that will
be explained in the next section. Note: fork()
copies also a memory
area known as the 'U Area' (or User Area). This area contains, amongst other
things, the file descriptor table of the process. This means that after returning
from the fork()
call, the child process inherits all files that
were open in the parent process. If one of them reads from such an open file,
the read/write pointer is advanced for both of them. On the other hand, files
opened after the call to fork()
are not shared by both processes.
Further more, if one process closes a shared file, it is still kept open in
the other process.
Once we have created a child process, there are two possibilities. Either the parent process exits before the child, or the child exits before the parent. Now, Unix's semantics regarding parent-child process relations state something like this:
When the parent process is not properly coded, the child remains in the zombie state forever. Such processes can be noticed by running the 'ps' command (shows the process list), and seeing processes having the string "<defunct>" as their command name.
wait()
System Call The simple way of a process to acknowledge the
death of a child process is by using the wait()
system call.
As we mentioned earlier, When wait()
is called, the process
is suspended until one of its child processes exits, and then the call returns
with the exit status of the child process. If it has a zombie child process,
the call returns immediately, with the exit status of that process.
The problem with calling wait()
directly,
is that usually you want the parent process to do other things, while its
child process executes its code. Otherwise, you're not really enjoying multi-processes,
do you? That problem has a solution by using signals. When a child process
dies, a signal, SIGCHLD (or SIGCLD) is sent to its parent process. Thus,
using a proper signal handler, the parent will get an asynchronous notification,
and then when it'll call wait()
, the system assures that the
call will return immediately, since there is already a zombie child. Here
is an example of our "hello world" program, using a signal handler this time.
#include <stdio.h> /* basic I/O routines. */
#include <unistd.h> /* define fork(), etc. */
#include <sys/types.h> /* define pid_t, etc. */
#include <sys/wait.h> /* define wait(), etc. */
#include <signal.h> /* define signal(), etc. */
/* first, here is the code for the signal handler */
void catch_child(int sig_num)
{
/* when we get here, we know there's a zombie child waiting */
int child_status;
wait(&child_status);
printf("child exited.\n");
}
.
.
/* and somewhere in the main() function ... */
.
.
/* define the signal handler for the CHLD signal */
signal(SIGCHLD, catch_child);
/* and the child process forking code... */
{
int child_pid;
int i;
child_pid = fork();
switch (child_pid) {
case -1: /* fork() failed */
perror("fork");
exit(1);
case 0: /* inside child process */
printf("hello world\n");
sleep(5); /* sleep a little, so we'll have */
/* time to see what is going on */
exit(0);
default: /* inside parent process */
break;
}
/* parent process goes on, minding its own business... */
/* for example, some output... */
for (i=0; i<10; i++) {
printf("%d\n", i);
sleep(1); /* sleep for a second, so we'll have time to see the mix */
}
}
Lets examine the flow of this program a little:
fork()
to spawn a child
process. exit()
, a
CHLD signal is sent by the system to the parent. wait()
call in the parent causes
the child to be completely removed off the system. Once we got our processes to run, we suddenly realize that they cannot communicate. After all, often when we start one process from another, they are supposed to accomplish some related tasks. One of the mechanisms that allow related-processes to communicate is the pipe, or the anonymous pipe.
One of the mechanisms that allow related-processes to communicate is the pipe, or the anonymous pipe. A pipe is a one-way mechanism that allows two related processes (i.e. one is an ancestor of the other) to send a byte stream from one of them to the other one. Naturally, to use such a channel properly, one needs to form some kind of protocol in which data is sent over the pipe. Also, if we want a two-way communication, we'll need two pipes, and a lot of caution...
The system assures us of one thing: The order in which data is written to the pipe, is the same order as that in which data is read from the pipe. The system also assures that data won't get lost in the middle, unless one of the processes (the sender or the receiver) exits prematurely.
pipe()
System CallThis system call is used to create a read-write pipe that may later be used to communicate with a process we'll fork off. The call takes as an argument an array of 2 integers that will be used to save the two file descriptors used to access the pipe. The first to read from the pipe, and the second to write to the pipe. Here is how to use this function:
/* first, define an array to store the two file descriptors */
int pipes[2];
/* now, create the pipe */
int rc = pipe(pipes);
if (rc == -1) { /* pipe() failed */
perror("pipe");
exit(1);
}
If the call to pipe()
succeeded, a
pipe will be created, pipes[0] will contain the number of its read file descriptor,
and pipes[1] will contain the number of its write file descriptor.
Now that a pipe was created, it should be put to
some real use. To do this, we first call fork()
to create a
child process, and then use the fact that the memory image of the child process
is identical to the memory image of the parent process, so the pipes[] array
is still defined the same way in both of them, and thus they both have the
file descriptors of the pipe. Further more, since the file descriptor table
is also copied during the fork, the file descriptors are still valid inside
the child process.
Lets see an example of a two-process system in which one (the parent process) reads input from the user, and sends it to the other (the child), which then prints the data to the screen. The sending of the data is done using the pipe, and the protocol simply states that every byte passed via the pipe represents a single character typed by the user.
#include <stdio.h> /* standard I/O routines. */
#include <unistd.h> /* defines pipe(), amongst other things. */
/* this routine handles the work of the child process. */
void do_child(int data_pipe[]) {
int c; /* data received from the parent. */
int rc; /* return status of read(). */
/* first, close the un-needed write-part of the pipe. */
close(data_pipe[1]);
/* now enter a loop of reading data from the pipe, and printing it */
while ((rc = read(data_pipe[0], &c, 1)) > 0) {
putchar(c);
}
/* probably pipe was broken, or got EOF via the pipe. */
exit(0);
}
/* this routine handles the work of the parent process. */
void do_parent(int data_pipe[])
{
int c; /* data received from the user. */
int rc; /* return status of getchar(). */
/* first, close the un-needed read-part of the pipe. */
close(data_pipe[0]);
/* now enter a loop of read user input, and writing it to the pipe. */
while ((c = getchar()) > 0) {
/* write the character to the pipe. */
rc = write(data_pipe[1], &c, 1);
if (rc == -1) { /* write failed - notify the user and exit */
perror("Parent: write");
close(data_pipe[1]);
exit(1);
}
}
/* probably got EOF from the user. */
close(data_pipe[1]); /* close the pipe, to let the child know we're done. */
exit(0);
}
/* and the main function. */
int main(int argc, char* argv[])
{
int data_pipe[2]; /* an array to store the file descriptors of the pipe. */
int pid; /* pid of child process, or 0, as returned via fork. */
int rc; /* stores return values of various routines. */
/* first, create a pipe. */
rc = pipe(data_pipe);
if (rc == -1) {
perror("pipe");
exit(1);
}
/* now fork off a child process, and set their handling routines. */
pid = fork();
switch (pid) {
case -1: /* fork failed. */
perror("fork");
exit(1);
case 0: /* inside child process. */
do_child(data_pipe);
/* NOT REACHED */
default: /* inside parent process. */
do_parent(data_pipe);
/* NOT REACHED */
}
return 0; /* NOT REACHED */
}
As we can see, the child process closed the write-end of the pipe (since it only needs to read from the pipe), while the parent process closed the read-end of the pipe (since it only needs to write to the pipe). This closing of the un-needed file descriptor was done to free up a file descriptor entry from the file descriptors table of the process. It isn't necessary in a small program such as this, but since the file descriptors table is limited in size, we shouldn't waste unnecessary entries.
The complete source code for this example may be found in the file one-way-pipe.c.
In a more complex system, we'll soon discover that this one-way communications is too limiting. Thus, we'd want to be able to communication in both directions - from parent to child, and from child to parent. The good news is that all we need to do is open two pipes - one to be used in each direction. The bad news, however, is that using two pipes might cause us to get into a situation known as 'deadlock':
Such a situation might occur when two processes communicate via two pipes. Here are two scenarios that could led to such a deadlock:
write()
system call gets blocked until the buffer has some free space. The only
way to free space on the buffer, is by reading data from the pipe.write()
system call. Since no other process is reading
from any of the pipes, our two processes have just entered a deadlock.
Lets see an example of a (hopefully) deadlock-free program in which one process reads input from the user, writes it to the other process via a pipe. the second process translates each upper-case letter to a lower-case letter and sends the data back to the first process. Finally, the first process writes the data to standard output.
#include <stdio.h> /* standard I/O routines. */
#include <unistd.h> /* defines pipe(), amongst other things. */
#include <ctype.h> /* defines isascii(), toupper(), and other */
/* character manipulation routines. */
/* function executed by the user-interacting process. */
void user_handler(int input_pipe[], int output_pipe[])
{
int c; /* user input - must be 'int', to recognize EOF (= -1). */
char ch; /* the same - as a char. */
int rc; /* return values of functions. */
/* first, close unnecessary file descriptors */
close(input_pipe[1]); /* we don't need to write to this pipe. */
close(output_pipe[0]); /* we don't need to read from this pipe. */
/* loop: read input, send via one pipe, read via other */
/* pipe, and write to stdout. exit on EOF from user. */
while ((c = getchar()) > 0) {
/* note - when we 'read' and 'write', we must deal with a char, */
/* rather then an int, because an int is longer then a char, */
/* and writing only one byte from it, will lead to unexpected */
/* results, depending on how an int is stored on the system. */
ch = (char)c;
/* write to translator */
rc = write(output_pipe[1], &ch, 1);
if (rc == -1) { /* write failed - notify the user and exit. */
perror("user_handler: write");
close(input_pipe[0]);
close(output_pipe[1]);
exit(1);
}
/* read back from translator */
rc = read(input_pipe[0], &ch, 1);
c = (int)ch;
if (rc <= 0) { /* read failed - notify user and exit. */
perror("user_handler: read");
close(input_pipe[0]);
close(output_pipe[1]);
exit(1);
}
/* print translated character to stdout. */
putchar(c);
}
/* close pipes and exit. */
close(input_pipe[0]);
close(output_pipe[1]);
exit(0);
}
/* now comes the function executed by the translator process. */
void translator(int input_pipe[], int output_pipe[])
{
int c; /* user input - must be 'int', to recognize EOF (= -1). */
char ch; /* the same - as a char. */
int rc; /* return values of functions. */
/* first, close unnecessary file descriptors */
close(input_pipe[1]); /* we don't need to write to this pipe. */
close(output_pipe[0]); /* we don't need to read from this pipe. */
/* enter a loop of reading from the user_handler's pipe, translating */
/* the character, and writing back to the user handler. */
while (read(input_pipe[0], &ch, 1) > 0) {
/* translate any upper-case letter to lower-case. */
c = (int)ch;
if (isascii(c) && isupper(c))
c = tolower(c);
ch = (char)c;
/* write translated character back to user_handler. */
rc = write(output_pipe[1], &ch, 1);
if (rc == -1) { /* write failed - notify user and exit. */
perror("translator: write");
close(input_pipe[0]);
close(output_pipe[1]);
exit(1);
}
}
/* close pipes and exit. */
close(input_pipe[0]);
close(output_pipe[1]);
exit(0);
}
/* and finally, the main function: spawn off two processes, */
/* and let each of them execute its function. */
int main(int argc, char* argv[])
{
/* 2 arrays to contain file descriptors, for two pipes. */
int user_to_translator[2];
int translator_to_user[2];
int pid; /* pid of child process, or 0, as returned via fork. */
int rc; /* stores return values of various routines. */
/* first, create one pipe. */
rc = pipe(user_to_translator);
if (rc == -1) {
perror("main: pipe user_to_translator");
exit(1);
}
/* then, create another pipe. */
rc = pipe(translator_to_user);
if (rc == -1) {
perror("main: pipe translator_to_user");
exit(1);
}
/* now fork off a child process, and set their handling routines. */
pid = fork();
switch (pid) {
case -1: /* fork failed. */
perror("main: fork");
exit(1);
case 0: /* inside child process. */
translator(user_to_translator, translator_to_user); /* line 'A' */
/* NOT REACHED */
default: /* inside parent process. */
user_handler(translator_to_user, user_to_translator); /* line 'B' */
/* NOT REACHED */
}
return 0; /* NOT REACHED */
}
A few notes:
isascii()
is a function that checks if the given
character code is a valid ASCII code. isupper()
is a function that checks if a given character is an upper-case letter.
tolower()
is a function that translates an upper-case
letter to its equivalent lower-case letter. translator(user_to_translator, user_to_translator); /*
line 'A' */
user_handler(translator_to_user, translator_to_user); /*
line 'B' */
The complete source code for this example may be found in the file two-way-pipe.c.
One limitation of anonymous pipes is that only processes 'related' to the process that created the pipe (i.e. siblings of that process.) may communicate using them. If we want two un-related processes to communicate via pipes, we need to use named pipes.
A named pipe (also called a named FIFO, or just FIFO) is a pipe whose access point is a file kept on the file system. By opening this file for reading, a process gets access to the reading end of the pipe. By opening the file for writing, the process gets access to the writing end of the pipe. If a process opens the file for reading, it is blocked until another process opens the file for writing. The same goes the other way around.
mknod
Command A named pipe may be created either via the 'mknod'
(or its newer replacement, 'mkfifo'), or via the mknod()
system
call (or by the POSIX-compliant mkfifo()
function). To create
a named pipe with the file named 'prog_pipe', we can use the following command:
mknod prog_pipe p
We could also provide a full path to where we want the named pipe created.
If we then type 'ls -l prog_pipe', we will see something like this:
prw-rw-r-- 1 choo choo 0 Nov 7 01:59 prog_pipe
Opening a named pipe is done just like opening
any other file in the system, using the open()
system call,
or using the fopen()
standard C function. If the call succeeds,
we get a file descriptor (in the case of open()
, or a 'FILE'
pointer (in the case of fopen()
), which we may use either for
reading or for writing, depending on the parameters passed to open()
or to fopen()
.
Reading from a named pipe is very similar to reading from a file, and the same goes for writing to a named pipe. Yet there are several differences:
Thus, when writing a program that uses a named pipe, we must take these limitations into account. We could also turn the file descriptor via which we access the named pipe to a non-blocking mode. This, however, is out of the scope of our tutorial. For info about how to do that, and how to handle a non-blocking pipe, please refer to the manual pages of 'open(2)', fcntl(2), read(2) and write(2).
As an example to an obscure usage of named pipes,
we will borrow some idea from a program that allows one to count how many
times they have been "fingered" lately. As you might know, on many Unix systems,
there is a finger daemon, that accepts requests from users running the "finger"
program, with a possible user name, and tells them when this user last logged
on, as well as some other information. Amongst other thing, the finger daemon
also checks if the user has a file named '.plan' (that is dot followed by
"plan") in her home directory. If there is such a file, the finger daemon
opens it, and prints its contents to the client. For example, on my Linux
machine, fingering my account might show something like:
[choo@simey1 ~]$ finger choo
Login: choo Name: guy keren
Directory: /home/choo Shell: /bin/tcsh
On since Fri Nov 6 15:46 (IDT) on tty6
No mail.
Plan:
- Breed a new type of dogs.
- Water the plants during all seasons.
- Finish the next tutorial on time.
This feature of the finger daemon may be used to
create a program that tells the client how many times i was fingered. For
that to work, we first create a named pipe, where the '.plan' file resides:
mknod /home/choo/.plan p
If i now try to finger myself, the output will stop before showing the 'plan'
file. How so? this is because of the blocking nature of a named pipe. When
the finger daemon opens my '.plan' file, there is no write process, and thus
the finger daemon blocks. Thus, don't run this on a system where you expect
other users to finger you often.
The second part of the trick, is compiling the named-pipe-plan.c
program, and running it. note that it contains the full path to the '.plan'
file, so change that to the appropriate value for your account, before compiling
it. When you run the program, it gets into an endless loop of opening the
named pipe in writing mode, write a message to the named pipe, close it,
and sleep for a second. Look at the program's source code for more information.
A sample of its output looks like this:
[choo@simey1 ~]$ finger choo
Login: choo Name: guy keren
Directory: /home/choo Shell: /bin/tcsh
On since Fri Nov 6 15:46 (IDT) on tty6
No mail.
Plan:
I have been fingered 8 times today
Various sockets-based mechanisms may be used to communicate amongst processes. The underlying communications protocol may be TCP, UDP, IP, or any other protocol from the TCP/IP protocols family. There is also a socket of type 'Unix-domain', which uses some protocol internal to the operating system to communicate between processes all residing on a single machine. Unix-domain sockets are similar to named pipes in that the communicating processes use a file in the system to connect to establish a connection. For more information about programming with sockets, please refer to our tutorial about internetworking with Unix sockets.
Many variants of Unix these days support a set of inter-process communications methods, which are derived from Unix System V release 4, originating from AT&T Bell laboratories. These mechanisms include message queues (used for sending and receiving messages), shared memory (used to allow several processes share data in memory) and semaphores (used to co-ordinate access by several processes, to other resources). Each of these resource types is handled by the system, and unlike anonymous pipes, may out-live the process that created it. These resources also have some security support by the system, that allows one to specify which processes may access a given message queue, for example.
The fact that these resources are global to the system has two contradicting implications. On one hand, it means that if a process exits, the data it sent through a message queue, or placed in shared memory is still there, and can be collected by other processes. On the other hand, this also means that the programmer has to take care of freeing these resources, or they occupy system resources until the next reboot, or until being removed by hand.
I am going to make a statement here about these
communications mechanisms, that might annoy some readers: System V IPC mechanisms
are evil regarding their implementation, and should not be used unless there
is a very good reason. One of the problem with these mechanism, is that one
cannot use the select()
(or its replacement, poll()
)
with them, and thus a process waiting for a message to be placed in a message
queue, cannot be notified about messages coming via other resources (e.g.
other message queues, pipes or sockets). In my opinion, this limitation is
an oversight by the designers of these mechanisms. Had they used file descriptors
to denote IPC resources (like they are used for pipes, sockets and files)
life would be easier.
Another problem with System V IPC is their system-global nature. The total number of message queues that may live in the system, for example, is shared by all processes. Worse than that, the number of messages waiting in all messages queues is also limited globally. One process spewing many such messages will break all processes using message queues. The same goes for other such resources. There are various other limitations imposed by API (Application programming interface). For example, one may wait on a limited set of semaphores at the same time. If you want more than this, you have to split the waiting task, or re-design your application.
Having said that, there are still various applications where using system V IPC (we'll call it SysV IPC, for short) will save you a large amount of time. In these cases, you should go ahead and use these mechanism - just handle with care.
Before delving into the usage of the different System V IPC mechanisms, we will describe the security model used to limit access to these resources.
Each resource in SysV IPC may be either private or public. Private means that it may be accessed only by the process that created it, or by child processes of this process. Public means that it may be potentially accessed by any process in the system, except when access permission modes state otherwise.
SysV IPC resources may be protected using access
mode permissions, much like files and directories are protected by the Unix
system. Each such resource has an owning user and an owning group. Permission
modes define if and how processes belonging to different users in the system
may access this resource. Permissions may be set separately for the owning
user, for users from the owning group, and everyone else. permissions may
be set for reading the resource (e.g. reading messages from a message queue),
or writing to the resource (e.g. sending a message on a queue, changing the
value of a semaphore). A structure of type 'ipc_perm', which is defined as
follows:
struct ipc_perm
{
key_t key; /* key identifying the resource */
ushort uid; /* owner effective user ID and effective group ID */
ushort gid;
ushort cuid; /* creator effective user ID and effective group ID */
ushort cgid;
ushort mode; /* access modes */
ushort seq; /* sequence number */
};
These fields have the following meanings:
Part of the SysV IPC API allows us to modify the access permissions for the resources. We will encounter them when discussing the different IPC methods.
Since SysV IPC resources live outside the scope of a single process, there is a need to manage them somehow - delete resources that were left by irresponsible processes (or process crashes); check the number of existing resources of each type (especially to find if the system-global limit was reached), etc. Two utilities were created for handling these jobs: 'ipcs' - to check usage of SysV IPC resources, and 'ipcrm' - to remove such resources.
Running 'ipcs' will show us statistics separately for each of the three resource types (shared memory segments, semaphore arrays and message queues). For each resource type, the command will show us some statistics for each resource that exists in the system. It will show its identifier, owner, size of resources it occupies in the system, and permission flags. We may give 'ipcs' a flag to ask it to show only resources of one type ('-m' for shared Memory segments, -q for message Queues and '-s' for Semaphore arrays). We may also use 'ipcs' with the '-l' flag to see the system enforced limits on these resources, or the '-u' flag to show us usage summary. Refer to the manual page of 'ipcs' for more information.
The 'ipcrm' command accepts a resource type ('shm', 'msg' or 'sem') and a resource ID, and removes the given resource from the system. We need to have the proper permissions in order to delete a resource.
One of the problems with pipes is that it is up to you, as a programmer, to establish the protocol. Now, usually this protocol is based on sending separate messages. With a stream taken from a pipe, it means you have to somehow parse the bytes, and separate them to packets. Another problem is that data sent via pipes always arrives in a FIFO order. This means that before you can read any part of the stream, you have to consume all the bytes sent before the piece you're looking for, and thus you need to construct your own queuing mechanism on which you place the data you just skipped, to be read later. If that's what you're interested at, this is a good time to get acquainted with message queues.
A message queue is a queue onto which messages can be placed. A message is composed of a message type (which is a number), and message data. A message queue can be either private, or public. If it is private, it can be accessed only by its creating process or child processes of that creator. If it's public, it can be accessed by any process that knows the queue's key. Several processes may write messages onto a message queue, or read messages from the queue. Messages may be read by type, and thus not have to be read in a FIFO order as is the case with pipes.
msgget()
In order to use a message queue, it has to be created
first. The msgget()
system call is used to do just that. This
system call accepts two parameters - a queue key, and flags. The key may
be one of:
open()
system call, and will be explained later, and it also contains access permission
bits. The lowest 9 bits of the flags are used to define access permission
for the queue, much like similar 9 bits are used to control access to files.
the bits are separated into 3 groups - user, group and others. In each set,
the first bit refers to read permission, the second bit - to write permission,
and the third bit is ignored (no execute permission is relevant to message
queues).
#include <stdio.h> /* standard I/O routines. */
#include <sys/types.h> /* standard system data types. */
#include <sys/ipc.h> /* common system V IPC structures. */
#include <sys/msg.h> /* message-queue specific functions. */
/* create a private message queue, with access only to the owner. */
int queue_id = msgget(IPC_PRIVATE, 0600); /* <-- this is an octal number. */
if (queue_id == -1) {
perror("msgget");
exit(1);
}
A few notes about this code:
struct msgbuf
Before we go to writing messages to the queue or
reading messages from it, we need to see how a message looks. The system
defines a structure named 'msgbuf' for this purpose. Here is how it is defined:
struct msgbuf {
long mtype; /* message type, a positive number (cannot be zero). */
char mtext[1]; /* message body array. usually larger than one byte. */
};
sizeof(struct msgbuf)
. Lets see how we create an "hello
world" message:
/* first, define the message string */
char* msg_text = "hello world";
/* allocate a message with enough space for length of string and */
/* one extra byte for the terminating null character. */
struct msgbuf* msg =
(struct msgbuf*)malloc(sizeof(struct msgbuf) + strlen(msg_text));
/* set the message type. for example - set it to '1'. */
msg->mtype = 1;
/* finally, place the "hello world" string inside the message. */
strcpy(msg->mtext, msg_text);
Few notes:
strlen(msg_text)
more than the size of "struct
msgbuf
", and didn't need to allocate an extra place for the null
character, cause that's already contained in the msgbuf structure (the
1 byte of mtext there). memset()
, and not strcpy()
.
msgsnd()
Once we created the message queue, and a message
structure, we can place it on the message queue, using the msgsnd()
system call. This system call copies our message structure and places that
as the last message on the queue. It takes the following parameters:
int msqid
- id of message queue, as returned from the msgget()
call.
struct msgbuf* msg
- a pointer
to a properly initializes message structure, such as the one we prepared
in the previous section. int msgsz
- the size of
the data part (mtext) of the message, in bytes. int msgflg
- flags specifying
how to send the message. may be a logical "or" of the following:
IPC_NOWAIT
- if the message cannot be sent immediately, without blocking the process,
return '-1', and set errno
to EAGAIN
.
msgsnd()
like this:
int rc = msgsnd(queue_id, msg, strlen(msg_text)+1, 0);
if (rc == -1) {
perror("msgsnd");
exit(1);
}
msgsnd()
assumes
the data in the message to be an arbitrary sequence of bytes, so it cannot
know we've got the null character there too, unless we state that explicitly.
msgrcv()
We may use the system call msgrcv()
In order to read a message from a message queue. This system call accepts
the following list of parameters:
int msqid
- id of the queue, as returned from msgget()
. struct msgbuf* msg
- a pointer
to a pre-allocated msgbuf structure. It should generally be large enough
to contain a message with some arbitrary data (see more below). int msgsz
- size of largest
message text we wish to receive. Must NOT be larger than the amount
of space we allocated for the message text in 'msg'. int msgtyp
- Type of message
we wish to read. may be one of:
int msgflg
- a logical 'or'
combination of any of the following flags:
IPC_NOWAIT
- if there is no message on the queue matching what we want to
read, return '-1', and set errno
to ENOMSG
.
MSG_EXCEPT
- if the
message type parameter is a positive integer, then return the first
message whose type is NOT equal to the given integer. MSG_NOERROR
- If a message
with a text part larger than 'msgsz' matches what we want to read,
then truncate the text when copying the message to our msgbuf structure.
If this flag is not set and the message text is too large, the
system call returns '-1', and errno
is set to E2BIG
.
/* prepare a message structure large enough to read our "hello world". */
struct msgbuf* recv_msg =
(struct msgbuf*)malloc(sizeof(struct msgbuf)+strlen("hello world"));
/* use msgrcv() to read the message. We agree to get any type, and thus */
/* use '0' in the message type parameter, and use no flags (0). */
int rc = msgrcv(queue_id, recv_msg, strlen("hello world")+1, 0, 0);
if (rc == -1) {
perror("msgrcv");
exit(1);
}
A few notes:
msgrcv()
call would have blocked our process until one of the following happens:
errno
would be set to EIDRM
). errno
would be set to EINTR
.
As an example of using non-private message queues, we will show a program, named "queue_sender", that creates a message queue, and then starts sending messages with different priorities onto the queue. A second program, named "queue_reader", may be run that reads the messages from the queue, and does something with them (in our example - just prints their contents to standard output). The "queue_reader" is given a number on its command line, which is the priority of messages that it should read. By running several copies of this program simultaneously, we can achieve a basic level of concurrency. Such a mechanism may be used by a system in which several clients may be sending requests of different types, that need to be handled differently. The complete source code may be found in the public-queue directory.
One of the problems when writing multi-process application is the need to synchronize various operations between the processes. Communicating requests using pipes, sockets and message queues is one way to do it. however, sometimes we need to synchronize operations amongst more than two processes, or to synchronize access to data resources that might be accessed by several processes in parallel. Semaphores are a means supplied with SysV IPC that allow us to synchronize such operations.
A semaphore is a resource that contains an integer value, and allows processes to synchronize by testing and setting this value in a single atomic operation. This means that the process that tests the value of a semaphore and sets it to a different value (based on the test), is guaranteed no other process will interfere with the operation in the middle.
Two types of operations can be carried on a semaphore: wait and signal. A set operation first checks if the semaphore's value equals some number. If it does, it decreases its value and returns. If it does not, the operation blocks the calling process until the semaphore's value reaches the desired value. A signal operation increments the value of the semaphore, possibly awakening one or more processes that are waiting on the semaphore. How this mechanism can be put to practical use will be explained soon.
A semaphore set is a structure that stores a group of semaphores together, and possibly allows the process to commit a transaction on part or all of the semaphores in the set together. In here, a transaction means that we are guaranteed that either all operations are done successfully, or none is done at all. Note that a semaphore set is not a general parallel programming concept, it's just an extra mechanism supplied by SysV IPC.
semget()
Creation of a semaphore set is done using the semget()
system call. Similarly to the creation of message queues, we supply some
ID for the set, and some flags (used to define access permission mode and
a few options). We also supply the number of semaphores we want to have in
the given set. This number is limited to SEMMSL
, as defined in
file /usr/include/sys/sem.h
. Lets see an example:
/* ID of the semaphore set. */
int sem_set_id_1;
int sem_set_id_2;
/* create a private semaphore set with one semaphore in it, */
/* with access only to the owner. */
sem_set_id_1 = semget(IPC_PRIVATE, 1, IPC_CREAT | 0600);
if (sem_set_id_1 == -1) {
perror("main: semget");
exit(1);
}
/* create a semaphore set with ID 250, three semaphores */
/* in the set, with access only to the owner. */
sem_set_id_2 = semget(250, 3, IPC_CREAT | 0600);
if (sem_set_id_2 == -1) {
perror("main: semget");
exit(1);
}
semctl()
After the semaphore set is created, we need to
initialize the value of the semaphores in the set. We do that using the semctl()
system call. Note that this system call has other uses, but they are not relevant
to our needs right now. Lets assume we want to set the values of the three
semaphores in our second set to values 3, 6 and 0, respectively. The ID of
the first semaphore in the set is '0', the ID of the second semaphore is
'1', and so on.
/* use this to store return values of system calls. */
int rc;
/* initialize the first semaphore in our set to '3'. */
rc = semctl(sem_set_id_2, 0, SETVAL, 3);
if (rc == -1) {
perror("main: semctl");
exit(1);
}
/* initialize the second semaphore in our set to '6'. */
rc = semctl(sem_set_id_2, 1, SETVAL, 6);
if (rc == -1) {
perror("main: semctl");
exit(1);
}
/* initialize the third semaphore in our set to '0'. */
rc = semctl(sem_set_id_2, 2, SETVAL, 0);
if (rc == -1) {
perror("main: semctl");
exit(1);
}
There are one comment to be made about the way
we used semctl()
here. According to the manual, the last parameter
for this system call should be a union of type union semun
.
However, since the SETVAL
(set value) operation only uses the
int val
part of the union, we simply passed an integer to the
function. The proper way to use this system call was to define a variable
of this union type, and set its value appropriately, like this:
/* use this variable to pass the value to the semctl() call */
union semun sem_val;
/* initialize the first semaphore in our set to '3'. */
sem_val.val = 0;
rc = semctl(sem_set_id_2, 2, SETVAL, sem_val);
if (rc == -1) {
perror("main: semctl");
exit(1);
}
semop()
Sometimes we have a resource that we want to allow
only one process at a time to manipulate. For example, we have a file that
we only want written into only by one process at a time, to avoid corrupting
its contents. Of-course, we could use various file locking mechanisms to
protect the file, but we will demonstrate the usage of semaphores for this
purpose as an example. Later on we will see the real usage of semaphores,
to protect access to shared memory segments. Anyway, here is a code snippest.
It assumes the semaphore in our set whose id is "sem_set_id" was initialized
to 1 initially:
/* this function updates the contents of the file with the given path name. */
void update_file(char* file_path, int number)
{
/* structure for semaphore operations. */
struct sembuf sem_op;
FILE* file;
/* wait on the semaphore, unless it's value is non-negative. */
sem_op.sem_num = 0;
sem_op.sem_op = -1; /* <-- Comment 1 */
sem_op.sem_flg = 0;
semop(sem_set_id, &sem_op, 1);
/* Comment 2 */
/* we "locked" the semaphore, and are assured exclusive access to file. */
/* manipulate the file in some way. for example, write a number into it. */
file = fopen(file_path, "w");
if (file) {
fprintf(file, "%d\n", number);
fclose(file);
}
/* finally, signal the semaphore - increase its value by one. */
sem_op.sem_num = 0;
sem_op.sem_op = 1; /* <-- Comment 3 */
sem_op.sem_flg = 0;
semop(sem_set_id, &sem_op, 1);
}
This code needs some explanations, especially regarding
the semantics of the semop()
calls.
semop()
to wait on
the semaphore. Supplying '-1' in sem_op.sem_op
means: If
the value of the semaphore is greater than or equal to '1', decrease
this value by one, and return to the caller. Otherwise (the value is
1 or less), block the calling process, until the value of the semaphore
becomes '1', at which point we return to the caller. semop()
assure us that when we return from this function, the value of the semaphore
is 0. Why? it couldn't be less, or else semop()
won't return.
It couldn't be more due to the way we later on signal the semaphore.
And why it cannot be more than '0'? read on to find out... Now, lets assume that any process that tries to access the file, does it only via a call to our "update_file" function. As you can see, when it goes through the function, it always decrements the value of the semaphore by 1, and then increases it by 1. Thus, the semaphore's value can never go above its initial value, which is '1'. Now lets check two scenarios:
semop()
call, the value of the semaphore is decremented to '0', and thus our
process is not blocked. We continue to execute the file update, and
with the second semop()
call, we raise the value of the
semaphore back to '1'. semop()
,
the value of the semaphore is '0', and when we call semop()
,
our process is blocked. When the other process signals the semaphore
with the second semop()
call, it increases the value of
the semaphore back to '0', and it wakes up the process blocked on the
semaphore, which is our process. We now get into executing the
file handling code, and finally we raise the semaphore's value back to '1'
with our second call to semop()
. We have the source code for a program demonstrating
the mutex concept, in the file named sem-mutex.c.
The program launches several processes (5, as defined by the NUM_PROCS macro),
each of which is executing the "update_file" function several times in a
row, and then exits. Try running the program, and scan its output. Each process
prints out its PID as it updates the file, so you can see what happens when.
Try to play with the DELAY macro (specifying how long a process waits between
two calls to "update_file") and see how it effects the order of the operations.
Check what happens if you replace the delay loop in the "do_child_loop" function,
with a call to sleep()
.
semop()
Using a semaphore as a mutex is not utilizing the full power of the semaphore. As we saw, a semaphore contains a counter, that may be used for more complex operations. Those operations often use a programming model called "producer-consumer". In this model, we have one or more processes that produce something, and one or more processes that consume that something. For example, one set of processes accept printing requests from clients and place them in a spool directory, and another set of processes take the files from the spool directory and actually print them using the printer.
To control such a printing system, we need the producers to maintain a count of the number of files waiting in the spool directory and incrementing it for every new file placed there. The consumers check this counter, and whenever it gets above zero, one of them grabs a file from the spool, and sends it to the printer. If there are no files in the spool (i.e. the counter value is zero), all consumer processes get blocked. The behavior of this counter sounds very familiar.... it is the exact same behavior of a counting semaphore.
Lets see how we can use a semaphore as a counter.
We still use the same two operations on the semaphore, namely "signal" and
"wait".
/* this variable will contain the semaphore set. */
int sem_set_id;
/* semaphore value, for semctl(). */
union semun sem_val;
/* structure for semaphore operations. */
struct sembuf sem_op;
/* first we create a semaphore set with a single semaphore, */
/* whose counter is initialized to '0'. */
sem_set_id = semget(IPC_PRIVATE, 1, 0600);
if (sem_set_id == -1) {
perror("semget");
exit(1);
}
sem_val.val = 0;
semctl(sem_set_id, 0, SETVAL, sem_val);
/* we now do some producing function, and then signal the */
/* semaphore, increasing its counter by one. */
.
.
sem_op.sem_num = 0;
sem_op.sem_op = 1;
sem_op.sem_flg = 0;
semop(sem_set_id, &sem_op, 1);
.
.
.
/* meanwhile, in a different process, we try to consume the */
/* resource protected (and counter) by the semaphore. */
/* we block on the semaphore, unless it's value is non-negative. */
sem_op.sem_num = 0;
sem_op.sem_op = -1;
sem_op.sem_flg = 0;
semop(sem_set_id, &sem_op, 1);
/* when we get here, it means that the semaphore's value is '0' */
/* or more, so there's something to consume. */
.
.
The full source code for a simple program that implements a producer-consumer system with two processes, is found in the file sem-producer-consumer.c.
As a complete example of using semaphores, we write a very simple print spool system. Two separate programs will be used. One runs as the printing command, and is found in the file tiny-lpr.c. It gets a file path on its command line, and copies this file into the spool area, increasing a global (on-private) semaphore by one. Another program runs as the printer daemon, and is found in the file tiny-lpd.c. It waits on the same global semaphore, and whenever its value is larger than one, it locates a file in the spool directory and sends it to the printer. In order to avoid race conditions when copying files into the directory and removing files from this directory, a second semaphore will be used as a mutex, to protect the spool directory. The complete tiny-spooler mini-project is found in the tiny-spool directory.
One problem might be that copying a file takes
a lot of time, and thus locking the spool directory for a long while. In
order to avoid that, 3 directories will be used. One serves as a temporary
place for tiny-lpr to copy files into. One will be used as the common spool
directory, and one will be used as a temporary directory into which tiny-lpd
will move the files before printing them. By putting all 3 directories on
the same disk, we assure that files can be moved between them using the rename()
system call, in one fast operation (regardless of the file size).
As we have seen, many methods were created in order
to let processes communicate. All this communications is done in order to
share data. The problem is that all these methods are sequential in nature.
What can we do in order to allow processes to share data in a random-access
manner?
Shared memory comes to the rescue. As you might know, on a Unix system, each
process has its own virtual address space, and the system makes sure no process
would access the memory area of another process. This means that if one process
corrupts its memory's contents, this does not directly affect any other process
in the system.
With shared memory, we declare a given section in the memory as one that will be used simultaneously by several processes. This means that the data found in this memory section (or memory segment) will be seen by several processes. This also means that several processes might try to alter this memory area at the same time, and thus some method should be used to synchronize their access to this memory area (did anyone say "apply mutual exclusion using a semaphore" ?).
In order to understand the concept of shared memory, we should first check how virtual memory is managed on the system.
In order to achieve virtual memory, the system divides memory into small pages each of the same size. For each process, a table mapping virtual memory pages into physical memory pages is kept. When the process is scheduled for running, its memory table is loaded by the operating system, and each memory access causes a mapping (by the CPU) to a physical memory page. If the virtual memory page is not found in memory, it is looked up in swap space, and loaded from there (this operation is also called 'page in').
When the process is started, it is being allocated a memory segment to hold the runtime stack, a memory segment to hold the programs code (the code segment), and a memory area for data (the data segment). Each such segment might be composed of many memory pages. When ever the process needs to allocate more memory, new pages are being allocated for it, to enlarge its data segment.
When a process is being forked off from another
process, the memory page table of the parent process is being copied to the
child process, but not the pages themselves. If the child process will try
to update any of these pages, then this page specifically will be copied,
and then only the copy of the child process will be modified. This behavior
is very efficient for processes that call fork()
and immediately
use the exec()
system call to replace the program it runs.
What we see from all of this is that all we need in order to support shared memory, is to some memory pages as shared, and to allow a way to identify them. This way, one process will create a shared memory segment, other processes will attach to them (by placing their physical address in the process's memory pages table). From now all these processes will access the same physical memory when accessing these pages, thus sharing this memory area.
A shared memory segment first needs to be allocated
(created), using the shmget()
system call. This call gets a
key for the segment (like the keys used in msgget()
and semget()
),
the desired segment size, and flags to denote access permissions and whether
to create this page if it does not exist yet. shmget()
returns
an identifier that can be later used to access the memory segment. Here is
how to use this call:
/* this variable is used to hold the returned segment identifier. */
int shm_id;
/* allocate a shared memory segment with size of 2048 bytes, */
/* accessible only to the current user. */
shm_id = shmget(100, 2048, IPC_CREAT | IPC_EXCL | 0600);
if (shm_id == -1) {
perror("shmget: ");
exit(1);
}
IPC_EXCL
in the flags to shmget()
. In that case,
the call will succeed only if the page did not exist before.
After we allocated a memory page, we need to add
it to the memory page table of the process. This is done using the shmat()
(shared-memory attach) system call. Assuming 'shm_id' contains an identifier
returned by a call to shmget()
, here is how to do this:
/* these variables are used to specify where the page is attached. */
char* shm_addr;
char* shm_addr_ro;
/* attach the given shared memory segment, at some free position */
/* that will be allocated by the system. */
shm_addr = shmat(shm_id, NULL, 0);
if (!shm_addr) { /* operation failed. */
perror("shmat: ");
exit(1);
}
/* attach the same shared memory segment again, this time in */
/* read-only mode. Any write operation to this page using this */
/* address will cause a segmentation violation (SIGSEGV) signal. */
shm_addr_ro = shmat(shm_id, NULL, SHM_RDONLY);
if (!shm_addr_ro) { /* operation failed. */
perror("shmat: ");
exit(1);
}
Placing data in a shared memory segment is done
by using the pointer returned by the shmat()
system call. Any
kind of data may be placed in a shared segment, except for pointers. The
reason for this is simple: pointers contain virtual addresses. Since the
same segment might be attached in a different virtual address in each process,
a pointer referring to one memory area in one process might refer to a different
memory area in another process. We can try to work around this problem by
attaching the shared segment in the same virtual address in all processes
(by supplying an address as the second parameter to shmat()
,
and adding the SHM_RND
flag to its third parameter), but this
might fail if the given virtual address is already in use by the process.
Here is an example of placing data in a shared
memory segment, and later on reading this data. We assume that 'shm_addr'
is a character pointer, containing an address returned by a call to shmat()
.
/* define a structure to be used in the given shared memory segment. */
struct country {
char name[30];
char capital_city[30];
char currency[30];
int population;
};
/* define a countries array variable. */
int* countries_num;
struct country* countries;
/* create a countries index on the shared memory segment. */
countries_num = (int*) shm_addr;
*countries_num = 0;
countries = (struct country*) ((void*)shm_addr+sizeof(int));
strcpy(countries[0].capital_city, "U.S.A");
strcpy(countries[0].capital_city, "Washington");
strcpy(countries[0].currency, "U.S. Dollar");
countries[0].population = 250000000;
(*countries_num)++;
strcpy(countries[1].capital_city, "Israel");
strcpy(countries[1].capital_city, "Jerusalem");
strcpy(countries[1].currency, "New Israeli Shekel");
countries[1].population = 6000000;
(*countries_num)++;
strcpy(countries[1].capital_city, "France");
strcpy(countries[1].capital_city, "Paris");
strcpy(countries[1].currency, "Frank");
countries[1].population = 60000000;
(*countries_num)++;
/* now, print out the countries data. */
for (i=0; i < (*countries_num); i++) {
printf("Country %d:\n", i+1);
printf(" name: %s:\n", countries[i].name);
printf(" capital city: %s:\n", countries[i].capital_city);
printf(" currency: %s:\n", countries[i].currency);
printf(" population: %d:\n", countries[i].population);
}
malloc()
. shmget()
,
there is no need to use malloc()
when placing data in that
segment. Instead, we do all memory management ourselves, by simple pointer
arithmetic operations. We also need to make sure the shared segment
was allocated enough memory to accommodate future growth of our data
- there are no means for enlarging the size of the segment once allocated
(unlike when using normal memory management - we can always move data
to a new memory location using the realloc()
function).
After we finished using a shared memory segment,
we should destroy it. It is safe to destroy it even if it is still in use
(i.e. attached by some process). In such a case, the segment will be destroyed
only after all processes detach it. Here is how to destroy a segment:
/* this structure is used by the shmctl()
system call. */
struct shmid_ds shm_desc;
/* destroy the shared memory segment. */
if (shmctl(shm_id, IPC_RMID, &shm_desc) == -1) {
perror("main: shmctl: ");
}
As a naive example of using shared memory, we collected the source code from the above sections into a file named shared-mem.c. It shows how a single process uses shared memory. Naturally, when two processes (or more) use a single shared memory segment, there may be race conditions, if one process tries to update this segment, while another is reading from it. To avoid this, we need to use some locking mechanism - SysV semaphores (used as mutexes) come to mind here. An example of two processes that access the same shared memory segment using a semaphore to synchronize their access, is found in the file shared-mem-with-semaphore.c.
ftok()
One of the problems with SysV IPC methods is the need to choose a unique identifier for our processes. How can we make sure that the identifier of a semaphore in our project won't collide with the identifier of a semaphore in some other program installed on the system?
To help with that, the ftok()
system
call was introduced. This system call accepts two parameters, a path to a
file and a character, and generates a more-or-less unique identifier. It
does that by finding the "i-node" number of the file (more or less the number
of the disk sector containing this file's information), combines it with
the second parameter, and thus generates an identifier, that can be later
fed to semget
, shmget()
or msgget()
.
Here is how to use ftok()
:
/* identifier returned by ftok()
*/
key_t set_key;
/* generate a "unique" key for our set, using the */
/* directory "/usr/local/lib/ourprojectdir". */
set_key = ftok("/usr/local/lib/ourprojectdir", 'a');
if (set_key == -1) {
perror("ftok: ");
exit(1);
}
/* now we can use 'set_key' to generate a set id, for example. */
sem_set_id = semget(set_key, 1, IPC_CREAT | 0600);
.
.
ftok
call with this file will generate a
different key. Thus, the file used should be a steady file, and not one that
is likely to be moved to a different disk or erased and re-created.
This document is copyright (c) 1998-2002 by guy keren.
The material in this document is provided AS IS, without any expressed or
implied warranty, or claim of fitness for a particular purpose. Neither the
author nor any contributers shell be liable for any damages incured directly
or indirectly by using the material contained in this document.
permission to copy this document (electronically or on paper, for personal
or organization internal use) or publish it on-line is hereby granted, provided
that the document is copied as-is, this copyright notice is preserved, and
a link to the original document is written in the document's body, or in
the page linking to the copy of this document.
Permission to make translations of this document is also granted, under
these terms - assuming the translation preserves the meaning of the text,
the copyright notice is preserved as-is, and a link to the original document
is written in the document's body, or in the page linking to the copy of
this document.
For any questions about the document and its license, please contact the author.