Unixabcd
Unixabcd
2. Uniqueness of UNIX
The features that made UNIX a hit from the start are:
Multitasking capability
Multi-user capability
Portability
UNIX programs
Library of application software
2.1. Multitasking
Many computers do just one thing at a time, as anyone who uses a PC or laptop can attest. Try logging onto our
company's network while opening the browser while opening a word processing program. Chances are the processor
will freeze for a few seconds while it sorts out the multiple instructions. UNIX, on the other hand, lets a computer do
several things at once, such as printing out one file while the user edits another file. This is a major feature for users,
since users don't have to wait for one application to end before starting another one.
2.2. Multi-user
The same design that permits multitasking permits multiple users to use the computer. The computer can take the
commands of a number of users -- determined by the design of the computer -- to run programs, access files, and
print documents at the same time. The computer can't tell the printer to print all the requests at once, but it does
prioritize the requests to keep everything orderly. It also lets several users access the same document by
compartmentalizing the document so that the changes of one user don't override the changes of another user.
Tools those aren’t necessary for the operation of UNIX but provide the user with additional capabilities, such
as typesetting capabilities and e-mail.
The heart of the operating system, the kernel controls the hardware and turns part of the system on and off at the
programmer’s command. If we ask the computer to list (ls) all the files in a directory, the kernel tells the computer to
read all the files in that directory from the disk and display them on our screen.
At this prompt, type in the username and press the enter/return/ key. Remember that UNIX is case sensitive (i.e.
Will, WILL and will are all different logins). We should then be prompted for the password:
Login : will-1
password :
Type the password in at the prompt and press the enter/return/ key. Note that the password will not be displayed
on the screen as We type it in.
If We mistype the username or password we will get an appropriate message from the computer and We will be
presented with the login: prompt again. Otherwise We should be presented with a shell prompt which looks
something like this:
$
To log out of a text-based UNIX shell, type "exit" at the shell prompt (or if that doesn't work try "logout"; if that
doesn't work press ctrl-d).
$ passwd
The system will prompt us for our old password, then for our new password. To eliminate any possible typing errors
we have made in our new password, it will ask We to reconfirm our new password.
Remember the following points when choosing our password:
Avoid characters which might not appear on all keyboards, e.g. '£'.
Make it at least 7 or 8 characters long and try to use a mix of letters, numbers and punctuation.
Note: One of the things we should do when we log in for the first time is to change our password.
4. File Types
There are four types of files in the Unix file system.
However, the difference is the cp does create a new file, while ln merely gives an alternate name to an old file. If we
make Y using ln, then Y is merely a new name for the same physical file X.
$ mkdir –p dir1/dir2
Example
$ uname –n
<hostname>
2. mv
mv (short for move) is a Unix command that moves a file from one place to another. The original file is deleted, and
the new file may have the same or a different name.
Syntax
$ mv [-f] [-i] oldname newname
-f mv will move the file(s) without prompting even if it is writing over an existing target. Note that this is
the default if the standard input is not a terminal.
-i Prompts before overwriting another file.
oldname The oldname of the file renaming.
newname The newname of the file renaming.
filename The name of the file we want to move directory - The directory of were we want the file to go.
Examples
$ mv myfile.txt newdirectory/
moves the file myfile.txt to the directory newdirectory.
$ mv myfile.txt ../
moves the file myfile.txt back one directory (if available).
3. date
The date command can be used to display or set the date. If a user has superuser privileges, he or she can set the
date by supplying a numeric string with the following command:
Fortunately there are options to manipulate the format. The format option is preceded by a + followed by any number
of field descriptors indicated by a % followed by a character to indicate which field is desired. The allowed field
descriptors are:
%n a newline
%t a tab
%D date as mm/dd/yy
%H hour (00-23)
%M minute (00-59)
%S second (00-59)
%T time as HH:MM:SS
Examples
$ date
Mon Jan 6 16:07:23 PST 1997
Examples
$ date -s "11/20/2003 12:48:00"
Set the date to the date and time shown.
Examples
To remove the file myfile.txt without prompting the user.
$ rm myfile.txt
A number of files can all be removed at the same time, Here we remove all the files with the lis extension.
$ rm *.lis
ultimateanswer.lis : ?y
lessultimateanswer.lis : ?n
moreultimateanswer.lis : …
Remove can also be used without asking for a confirmation. To do this uses the -force option. Here we remove all
the files with the lis extension, without asking for a confirmation.
$ rm -f *.lis
Beware when using the -f option we can easily remove all files from a directory by accident! To remove all files
without asking for a confirmation.
$ rm -f *
4.2. rmdir
Used to remove the directories form the system
Syntax
rmdir [-p] [-s][-r] directoryname
-p Allow users to remove the directory dirname and its parent directories which become empty. A
message is printed to standard error if all or part of the path could not be removed.
-s Suppress the message printed on the standard error when -p is in effect.
-r Delete nonempty directory
directory The name of the directory that we wish to delete.
Examples
removes the directory mydir
$ rmdir mydir
To deletes all directories in the current directory whose directory names begins with the characters "index".
$ rmdir index*
To remove a directory, even if files existed in that directory.
$ rm -r directory
To delete the directory named "new-novel". This directory, and all of it’s contents, are erased from the disk, including
any sub-directories and files.
$ rm -r new-novel
5. ls
The ls command lists the files in your current working directory. When we log onto your account on UNIX, your
current working directory is your home or personal directory. This is the directory in which we have personal disk
space to put files on or to create sub-directories under. The ls command also has options available. Options follow
the hyphen ( - ) sign. Two of the most useful options are a (return all files, even "hidden") and we (give long or full file
information). The ls command also accepts strings with the asterisk * used as a "wildcard" to tell UNIX to search for
all files that contain the specified sub-string.
Syntax
ls [-a] [-A] [-b] [-c] [-C] [-d] [-f] [-F] [-g] [-i] [-l] [-L] [-m] [-o] [-p]
[-q] [-r] [-R] [-s] [-t] [-u] [-x] [pathnames]
-a Shows us all files, even files that are hidden (these files begin with a dot.)
-A List all files including the hidden files. However, does not display the working directory (.) or the parent
directory (..).
-b Force printing of non-printable characters to be in octal \ ddd notation.
-c Use time of last modification of the i-node (file created, mode changed, and so forth) for sorting (-t) or
printing (-l or -n).
-C Multi-column output with entries sorted down the columns. Generally this is the default option.
-d If an argument is a directory it only lists its name not its contents.
-f Force each argument to be interpreted as a directory and list the name found in each slot. This option
turns off -l, -t, -s, and -r, and turns on -a; the order is the order in which entries appear in the directory.
-F Mark directories with a trailing slash (/), doors with a trailing greater-than sign (>), executable files with
a trailing asterisk (*), FIFOs with a trailing vertical bar (|), symbolic links with a trailing at-sign (@), and
AF_UNIX address family sockets with a trailing equals sign (=).
-g Same as -l except the owner is not printed.
-i For each file, print the i-node number in the first column of the report.
-l Shows us huge amounts of information (permissions, owners, size, and when last modified.)
-L If an argument is a symbolic link, list the file or directory the link references rather than the link itself.
-m Stream output format; files are listed across the page, separated by commas.
-n The same as -l, except that the owner's UID and group's GID numbers are printed, rather than the
associated character strings.
-o The same as -l, except that the group is not printed.
-p Displays a slash ( / ) in front of all directories.
-q Force printing of non-printable characters in file names as the character question mark (?).
-r Reverses the order of how the files are displayed.
-R Includes the contents of subdirectories.
-s Give size in blocks, including indirect blocks, for each entry.
Examples
$ ls -al *test*
-rw-r--r-- 1 hcsdar usg 592 Sep 1 1993 .test
drwx------ 2 hcsdar usg 512 Nov 11 16:21 dirtest
-rw-r--r-- 2 hcsdar usg 1097 Nov 2 1993 test
-rw------- 1 hcsdar usg 1097 Oct 19 15:54 test.bin
-rw------- 1 hcsdar usg 1216 Jul 15 1993 test.fil
What Does Coloumn1 tells us?
Column 1- tells us the type of file, what privileges it has and to whom these privileges are granted. There are three
types of privileges. Read and write privileges are easy to understand. The exec privilege is a little more difficult. We
can make a file "executable" by giving it exec privileges. This means that commands in the file will be executed when
we type the file name in at the UNIX prompt. It also means that when a directory which, to UNIX is a file like any other
file, can be "scanned" to see what files and sub-directories are in it. Privileges are granted to three levels of users:
1) The owner of the file. The owner is usually, but not always, the userid that created the file.
2) The group to which the owner belongs. At GSU, the group is usually, but not always designated as the first
three letters of the userid of the owner.
3) Everybody else who has an account on the UNIX machine where the file resides.
Column 2 - Number of links
Column 3 - Owner of the file. Normally the owner of the file is the user account that originally created it.
Column 4 - Group under which the file belongs. This is by default the group to which the account belongs or first
three letters of the userid. The group can be changed by the chgrp command.
Column 5 - Size of file (bytes).
Column 6 - Date of last update
Column 7 - Name of file
Examples
Rather than list the files contained in the /usr directory, this command lists information about the /usr directory itself
(without generating a listing of the contents of /usr). This is very useful when we want to check the permissions of the
$ ls -ld /usr
List the contents of your home directory by adding a tilde after the ls command.
$ ls ~
List the contents of your root directory.
$ ls /ls ../
List the contents of the parent directory.
$ ls */
List the contents of all sub directories.
6. finger
In Unix, finger is a program we can use to find information about computer users. It usually lists the login name, the
full name, and possibly other details about the user we are fingering. These details may include the office location
and phone number (if known), login time, idle time, time mail was last read, and the user's plan and project files. The
information listed varies, and we may not be able to get any information from some sites.
Syntax
finger [-b] [-f] [-h] [-i] [-l] [-m] [-p] [-q] [-s] [-w] [username]
-b Suppress printing the user's home directory and shell in a long format printout.
-f Suppress printing the header that is normally printed in a non-long format printout.
-h Suppress printing of the .project file in a long format printout.
-i Force "idle" output format, which is similar to short format except that only the login name, terminal, login time,
and idle time are printed.
-l Force long output format.
-m Match arguments only on user name (not first or last name).
-p Suppress printing of the .plan file in a long format printout.
-q Force quick output format, which is similar to short format except that only the login name, terminal, and login
time are printed.
-s Force short output format.
-w Suppress printing the full name in a short format printout.
Examples
We can find out someone's username and whether they are logged in with the finger command.
finger name_of_person[@remote_host]
"name" can be a first or last name, or a username.
$ finger wil-1
$ Login name: wil-1 In real life: Faculty1
how to change Finger
Most Unix systems have a chfn (change finger) command. It allows us to change the standard information that is
displayed when someone fingers your account
To change your finger information, on most systems, at the Unix shell prompt, enter chfn. We will be prompted to
enter values for each of the following fields:
Changing finger information for username
$ chfin
Name [your name]:
Location [XY 0436]:
Office Phone [555-1212]:
Home Phone [555-7352]:
7. who
The who command displays a list of users currently logged in to the local system. It displays each users login name,
the login device (TTY port), the login date and time. The command reads the binary file /var/admn/utmpx to obtain
this information and information about where the users logged in from
If a user logged in remotely the who command displays the remote host name or internet Protocol (IP) address in the
last column of the output.
Syntax
$ who [-a] [-b] [-d] [-H] [-l] [-m] [-nx] [-p] [-q] [-r] [-s] [-t] [-T] [-u]
[am i] [ file ]
-a Process /var/adm/utmp or the named file with -b, -d, -l, -p, -r, -t, -T, and -u options turned on.
-b Indicate the time and date of the last reboot.
-d Display all processes that have expired and not been respawned by init . The exit field appears for dead
processes and contains the termination and exit values (as returned by wait), of the dead process. This can
be useful in determining why a process terminated.
-H Output column headings above the regular output.
-l List only those lines on which the system is waiting for someone to login. The name field is LOGIN in such
cases. Other fields are the same as for user entries except that the state field does not exist.
-m Output only information about the current terminal.
-n x Take a numeric argument, x, which specifies the number of users to display per line. x must be at least 1. The
-n option may only be used with -q.
-p List any other process which is currently active and has been previously spawned by init . The name field is
the name of the program executed by init as found in /sbin/inittab. The state, line , and idle fields have no
meaning. The comment field shows the id field of the line from /sbin/inittab that spawned this process.
-q (quick who ) display only the names and the number of users currently logged on. When this option is used, all
other options are ignored.
-r Indicate the current run-level of the init process.
-s (default) List only the name, line, and time fields.
-t Indicate the last change to the system clock (using the date utility) by root. See su and date.
-T Same as the -s option, except that the state field is also written. state is one of the characters listed under the
/usr/bin/who version of this option. If the -u option is used with -T, the idle time is added to the end of the
previous format.
-u List only those users who are currently logged in. The name is the user's login name. The line is the name of
the line as found in the directory /dev. The time is the time that the user logged in. The idle column contains
the number of hours and minutes since activity last occurred on that particular line. A dot (.) indicates that the
terminal has seen activity in the last minute and is therefore ``current''. If more than twenty-four hours have
elapsed or the line has not been used since boot time, the entry is marked old. This field is useful when trying
to determine whether a person is working at the terminal or not. The pid is the process-ID of the user's shell.
The comment is the comment field associated with this line as found in /sbin/inittab. This can contain
information about where the terminal is located, the telephone number of the dataset, type of terminal if hard-
wired, and so forth.
am i In the "C" locale, limit the output to describing the invoking user, equivalent to the -m option. The am and i or I
must be separate arguments.
file Specify a path name of a file to substitute for the database of logged-on users that who uses by default.
Examples
$ who
The general format for output is:
name [state] line time [idle] [pid] [comment] [exit]
where:
name user's login name.
state capability of writing to the terminal.
line name of the line found in /dev.
time time since user's login.
idle time elapsed since the user's last activity.
pid user's process id.
comment comment line in inittab(4).
8. cal
Print a 12-month calendar (beginning with January) for the given year, or a one-month calendar of the given month
and year. month ranges from 1 to 12. year ranges from 1 to 9999. With no arguments, print a calendar for the current
month.
Before we can do the calendar program we must have a file named calendar at the root of your profile. Within that file
we may have something similar to:
Syntax
$ cal [options] [[month] year]
-j Display Julian dates (days numbered 1 to 365, starting from January 1).
-m Display Monday as the first day of the week.
-y Display entire year.
-V Display the source of the calendar file.
month Specifies the month for us want the calendar to be displayed. Must be the numeric representation of the
month. For example: January is 1 and December is 12.
year Specifies the year that we want to be displayed.
EXAMPLES
$ cal
$ cal -j
$ cal –m
$ cal –y
$ cal –y 1980
$ cal 12 2006
$ cal 2006 > year_file
Wildcards in UNIX
1. How to use UNIX Wildcards
Many computer operating systems provide ways to select certain files without typing complete filenames. For
example, we may wish to remove all files whose names end with "old". Unix allows us to use wildcards (more formally
known as metacharacters) to stand for one or more characters in a filename.
The two basic wildcard characters are ? and *. The wildcard ? matches any one character. The wildcard * matches
any grouping of zero or more characters. Some examples may help to clarify this. (Remember that Unix is case-
sensitive). Assume that your directory contains the following files:
Chap bite bin
bit Chap6 it
test.new abc
Lit site test.old
Big snit bin.old
ls [A-Z]it
Will find any file ending in "it" and beginning with a capital letter (in this case, the file Lit).
Wildcards can also be combined with [ ] sequences. To list any file beginning with a capital letter, we would use:
$ ls [A-Z]*
Chap1 Chap6 Lit
2. I/O Redirection
2.1 Standard File Descriptors
The Unix environment allows for each process to have access to three standard file descriptors by default. They are
0 standard input
1 standard output
2 standard error
It is the responsibility of the shell when executing a command to provide appropriate file descriptors to the process for
each of these standard files. Most Unix tools are developed to take their input from the standard input file and write
their output to the standard output file. Error messages that do not make up part of the expected output are usually
written to the standard error file.
Unless otherwise specified, the shell will usually pass it's own standard file descriptors down to the process that it
executes, allowing the output from any called tools to be included with the output of the script.
Through using I/O redirection, the developer can modify how the shell handles the file descriptors and usually either
replace one of the standard interactive file descriptors with a file on disk, or create a pipe to connect the output file
descriptor of one process to the input file descriptor of another process.
Redirection can also be used to perform redirection on file descriptors for a group of commands.
For the redirections, the standard file description redirection can be modified by placing the file descriptor identifier in
front of the redirection symbol.
For example 2> redirects standard error instead of the default standard input. 1> redirects standard output, which is
the default setting.
Example:
This would be the same as having a text file with the lines "bravo", "delta", "alpha" and "charlie" in it and redirecting it
using sort <input.txt but is simpler and cleaner (no problems with accidentally forgetting to include the input.txt file
when we distribute your script)
Path Names
1. Relative and Absolute Pathnames
1.1 Relative Pathnames
The use of the ".." notation allows us to navigate the directory tree structure. The ".." symbol means "parent
directory." Names with ".." in them are relative names because their meaning depends on where they are issued (the
present working directory). we can string together several ".." symbols, separated by the / symbol and other directory
names, to change directories. For example, if we are in portfolio and want to change to mary, we can do this with a cd
command followed by the relative pathname between portfolio and mary like this (first using pwd to show where we
are):
$ pwd
/users/john/portfolio
$cd ../../mary
$pwd
/users/mary
$
Directory or file references starting with .. are relative pathnames.
Directory or file references starting with a directory name are also relative pathnames. For example, if we are in the
users directory, the directory reference john/portfolio is a relative pathname:
$ pwd
/users
$cd john/portfolio
$pwd
/users/john/portfolio
$
2. Directory Abbreviations
Keyword Description
~ Your home (login) directory
~username Another user's home directory
. Working (current) directory
Basic Unix – Path Names Page 20 of 64
WK: 1 - Day: 5
Examples
cp foo bar
Copy a file named "foo" (in the current directory); name the copy "bar"
cp foo ~/Documents
Copy a file named "foo" (in the current directory) into your Documents directory
cp foo ~/Documents/bar
Copy a file named "a" (in the current directory) into your Documents directory and name the copy "bar"
cp *.jpg ~/Documents
Copy all files with names ending in ".jpg" into your Documents directory
cp -R Documents "Documents backup"
Copy an entire directory named "Documents"; name the copy "Documents backup". The quotes are needed because
of the space in the directory name.
sudo cp -Rp /Users "/Users backup"
Copy the entire /Users directory (including all of the user home folders inside it), preserving as much as possible of
the files' information (ownership, permissions, etc, but not resource forks) as cp knows how to; name the copy "Users
backup". Root access is required to use -p, so the example uses sudo to get root access temporarily.
Basic Unix – File Permissions Page 21 of 64
WK: 1 - Day: 6
File Permissions
1. Understanding file permissions on Unix: a brief tutorial
Every user on a Unix system has a unique username, and is a member of at least one group (the primary group for
that user). This group information is held in the password file (/etc/passwd). A user can also be a member of one or
more other groups. The auxiliary group information is held in the file /etc/group. Only the administrator can create
new groups or add/delete group members (one of the shortcomings of the system).
Every directory and file on the system has an owner, and also an associated group. It also has a set of permission
flags which specify separate read, write and execute permissions for the 'user' (owner), 'group', and 'other' (everyone
else with an account on the computer) The 'ls' command shows the permissions and group associated with files when
used with the ‘-l’ option. On some systems (e.g. Coos), the '-g' option is also needed to see the group information.
An example of the output produced by 'ls -l' is shown below.
drwx------ 2 richard staff 2048 Jan 2 1997 private
drwxrws--- 2 richard staff 2048 Jan 2 1997 admin
-rw-rw---- 2 richard staff 12040 Aug 20 1996 admin/userinfo
drwxr-xr-x 3 richard user 2048 May 13 09:27 public
Understanding how to read this output is useful to all unix users, but especially people using group access
permissions.
Field 1: a set of ten permission flags.
Field 2: link count (don't worry about this)
Field 3: owner of the file
Field 4: associated group for the file
Field 5: size in bytes
Field 6-8: date of last modification (format varies, but always 3 fields)
Field 9: name of file (possibly with path, depending on how ls was called)
The permission flags are read as follows (left to right)
position Meaning
Directory flag, 'd' if a directory, '-' if a normal file, something else occasionally may appear here for special
1
devices.
W File is writeable. On a directory, write access means we can add or delete files
File is executable (only for programs and shell scripts - not useful for data files). Execute permission on a
X
directory means we can list the files in that directory
S In the place where 'x' would normally go is called the set-UID or set-groupID flag.
Basic Unix – File Permissions Page 22 of 64
WK: 1 - Day: 6
The file can be changed or deleted by the If user or process can change directory contents
Write (w) corresponding user or group to which this set of somehow: create new or delete existing files in the
symbols applies directory or rename files.
On an executable program with set-UID or set-groupID, that program runs with the effective permissions of its owner
or group.
For a directory, the set-groupID flag means that all files created inside that directory will inherit the group of the
directory. Without this flag, a file takes on the primary group of the user creating the file. This property is important to
people trying to maintain a directory as group accessible. The subdirectories also inherit the set-groupID property.
Basic Unix – File Permissions Page 23 of 64
WK: 1 - Day: 6
1.1 Concepts
Unix commands alone are powerful, but when we combine them together, we can accomplish complex tasks with
ease. The way we combine Unix commands is through using pipes and filters.
1.3.1 head
head is a program on Unix and Unix-like systems used to display the first few lines of a text file or piped data. The
command-syntax is:
$ head [options] <file_name>
By default, head will print the first 10 lines of its input to the standard output. The number of lines printed may be
changed with a command line option. The following example shows the first 20 lines of filename:
$ head -20 filename
This displays the first 5 lines of all files starting with foo:
$ head -5 foo*
1.3.2 tail
tail is a program on Unix and Unix-like systems used to display the last few lines of a text file or piped data. The
command-syntax is:
$ tail [options] <file_name>
By default, tail will print the last 10 lines of its input to the standard output. With command line options the number of
lines printed and the printing units (lines, blocks or bytes) may be changed. The following example shows the last 20
lines of filename:
$ tail -20 filename
This example show all lines of filename after the first 2 lines:
$ tail +2 filename
File monitoring
tail has a special command line option -f (follow) that allows a file to be monitored. Instead of displaying the last few
lines and exiting, tail displays the lines and then monitors the file. As new lines are added to the file by another
process, tail updates the display. This is particularly useful for monitoring log files. The following command will display
the last 10 lines of messages and append new lines to the display as new lines are added to messages:
$ tail -f /var/adm/messages
To interrupt tail while it is monitoring, break-in with CTRL-C
1.3.3 more
more is better, isn't it? Better than what? Better than the cat command. cat dumps its arguments to stdout, which is
the terminal (unless we redirect it with > or >>). But what if we're working on your dissertation, and we'd like to read it
page by page, we'd use a command like:
$ more dissertation.txt
This will generate a nice page-by-page display of your masterpiece. Type
$ man more
at a command prompt, and check out the man page to get more details. Here we are only going to tell us the most
important features of more ( i.e. the features that we use). There are three important things we should know:
7. Typing q while examining a file quits more
8. Typing /SEARCHSTRING while examining a file searches for SEARCHSTRING
9. more is a great example of a filter
1.3.4 less
Opposite of the more command.
Both less and more display the contents of a file one screen at a time, waiting for us to press the Spacebar between
screens. This lets us read text without it scrolling quickly off your screen. The less utility is generally more flexible and
powerful than more, but more is available on all Unix systems while less may not be.
Basic Unix – Pipes and Filters Page 27 of 64
WK: 2 - Day: 1
The less command is a pager that allows us to move forward and backward (instead of only forward, as the more
pager behaves on some systems) when output is displayed one screen at a time. :
To read the contents of a file named textfile in the current directory, enter:
$ less textfile
The less utility is often used for reading the output of other commands. For example, to read the output of the ls
command one screen at a time, enter:
$ ls -la | less
In both examples, we could substitute more for less with similar results. To exit either less or more, press q . To exit
less after viewing the file, press q .
1.3.5 wc
In Unix, to get the line, word, or character count of a document, use the wc command. At the Unix shell prompt, enter:
wc filename Replace filename with the file or files for which we want information. For each file, wc will output three
numbers. The first is the line count, the second the word count, and the third is the character count. For example, if
we entered wc .login, the output would be something similar to the following: 38 135 847 .login To narrow the
focus of your query, we may use one or more of the following wc options:
Note: In some versions of wc, the -m option will not be available or -c will report characters. However, in most
cases, the values for -c and -m are equal.
Syntax:
To count the characters in a file. Here it counts the no of characters in the file abc.txt
$ Wc –c / abc.txt
For example, to find out how many bytes are in the .login file, we could enter:
$ wc -c .login
We may also pipe standard output into wc to determine the size of a stream. For example, to find out how many files
are in a directory, enter:
/bin/ls -l | wc -l
1.3.6 sort
sort is a standard Unix command line program that prints the lines of its input or concatenation of all files listed in it's
argument list in sorted order. The -r flag will reverse the sort order.
Example of sort in action:
$ cat phonebook
4 /bin/domainname
4 /bin/echo
4 /bin/hostname
4 /bin/pwd
...
24 /bin/ls
30 /bin/ps
44 /bin/ed
54 /bin/rmail
80 /bin/pax
102 /bin/sh
304 /bin/csh
The -n +1 option makes the program to sort according to numerical value, using the second column of data:
$ cat zipcode
Adam 12345
Bob 34567
Joe 56789
Sam 45678
Wendy 23456
Adam 12345
Wendy 23456
Bob 34567
Sam 45678
Joe 56789
The -r option just reverses the order of the sort:
$ cat zipcode | sort -nr +1
Joe 56789
Sam 45678
Bob 34567
Wendy 23456
Adam 12345
1.3.7 tr
tr (abbreviated from translate or transliterate) is a command in Unix-like operating systems.
When executed, the program reads from the standard input and writes to the standard output. It takes as parameters
two sets of characters, and replaces occurrences of the characters in the first set with the corresponding elements
from the other set. The following inputs, for instance, shift the input letters of the alphabet back by one character.
$ echo "ibm 9000" >computer.txt
$ tr a-z za-y <computer.txt
hal 9000
Note: when ever we are using the “tr” operator we have to use inpur rediction operator
In some older versions of tr (not POSIX-compliant), the character ranges must be enclosed in brackets, which must
then be quoted against interpretation by the shell:
$ tr "[a-z]" "z[a-y]" <computer.txt
Basic Unix – Pipes and Filters Page 29 of 64
WK: 2 - Day: 1
If it's not known in advance which variant of tr is being invoked, then in this example one would have to write the
ranges in unabbreviated form (tr abcdefghijklmnopqrstuvwxyz zabcdefghijklmnopqrstuvwxy). For some applications,
a single invocation is portable despite containing ranges: ROT13 can be portably implemented as tr "[A-M][N-Z][a-
m][n-z]" "[N-Z][A-M][n-z][a-m]". (This works because the brackets, which are not part of the range syntax in POSIX tr,
align properly in the two strings, and hence will be safely interpreted as a request to map the bracket character to
itself.)
Perl also has a tr operator, which operates analogously
1.3.8 cut
cut is a Unix command which is typically used to extract a certain range of characters from a line, usually from a file.
Syntax
cut [-c] [-f list] [-d delim] [-s] [file]
Flags which may be used include
-c Characters; a list following -c specifies a range of characters which will be returned, e.g. cut -c1-66 would
return the first 66 characters of a line
-f Specifies a field list, separated by a delimiter
list A comma separated or blank separated list of integer denoted fields, incrementally ordered. The -
indicator may be supplied as shorthand to allow inclusion of ranges of fields e.g. 4-6 for ranges 4 - 6 or 5 - as
shorthand for field 5 to the end, etc.
-d Delimiter; the character immediately following the -d option is the field delimiter for use in conjunction with
the -f option; the default delimiter is tab. Space and other characters with special meanings within the context
of the shell in use must be enquoted or escaped as necessary.
-s Bypasses lines which contain no field delimiters when -f is specified, unless otherwise indicated.
file The file (and accompanying path if necessary) to process as input. If no file is specified then standard
input will be used
Example
Extract columns of data
$ cut -f -3,5,7-9 -d ' ' infile1 > outfile1
-f 2,4-6 field
-c 35-44 character
-d ':' delimiter (default is a tab)
1.3.9 paste
Paste is a Unix utility tool which is used to join files horizontally (parallel merging), e.g. to join two similar length files
which are comma delimited. It is effectively the horizontal equivalent to the utility cat command which operates on the
vertical plane of two (or more) files, i.e. by adding one file to another in order.
Example
To paste several columns of data together, enter:
$ paste who where when > www
This creates a file named www that contains the data from the names file in one column, the places file in another,
and the dates file in a third. If the names, places, and dates file look like:
who where when
Sam Detroit January 3
Dave Edgewood February 4
Sue Tampa March 19
then the www file will contain:
Sam Detroit January 3
Dave Edgewood February 4
Sue Tampa March 19
Basic Unix – Pipes and Filters Page 30 of 64
WK: 2 - Day: 1
1.3.10 uniq
uniq is a Unix utility which, when fed a text file, outputs the file with adjacent identical lines collapsed to one. It is a
kind of filter program. Typically it is used after sort. It can also output only the duplicate lines (with the -d option), or
add the number of occurrences of each line (with the -c option).
An example: To see the list of lines in a file, sorted by the number of times each occurs:
$ sort file|uniq -c|sort -n
Using uniq like this is common when building pipelines in shell scripts.
Switches
-u Print only lines which are not repeated in the original file
-d Print one copy only of each repeated line in the input file.
-c Generate an output report in default style except that each line is preceded by a count of the number of
times it occurred. If this option is specified, the -u and -d options are ignored if either or both are also present.
-i Ignore case differences when comparing lines
-s Skips a number of characters in a line
-w Specifies the number of characters to compare in lines, after any characters and fields have been skipped
--help Displays a help message
--version Displays version number on stdout and exits.
Basic Unix – vi Editor Page 31 of 64
WK: 2 - Day: 2
vi Editor
1. General Introduction
The vi editor (short for visual editor) is a screen editor which is available on almost all Unix systems. Once we have
learned vi, we will find that it is a fast and powerful editor. vi has no menus but instead uses combinations of
keystrokes in order to accomplish commands. If we are just beginning to learn Unix, we might find the Pico editor
easier to use (most command options are displayed at the bottom of the screen). If we use the Pine email application
and have composed or replied to a message we have probably already used Pico as it is used for text entry. For
more information please refer to the Pine/Pico page.
1.1 Starting vi
To start using vi, at the Unix prompt type vi followed by a file name. If we wish to edit an existing file, type in its name;
if we are creating a new file, type in the name we wish to give to the new file.
$ vi filename
Then hit Return. We will see a screen similar to one below which shows blank lines with tildes and the name and
status of the file.
~
"myfile" [New file]
If we move the cursor as far as we can in any direction, we may see a screen flash or hear a beep.
To build on this further, we can precede these commands with a number for greater movement. For example, 5w
would move us forward five words; 12b would move us backwards twelve words. [We can also use numbers with the
commands mentioned earlier. For example, 5j would move us down 5 characters.]
2.4 Shortcuts
Two short cuts for moving quickly on a line include the $ and the 0 (zero) keys. The $ key will move us to the end of a
line, while the 0 will move us quickly to the beginning of a line.
Two other useful commands for moving quickly from one end to the other of a document are G to move to the end of
the file and 1G to move to the beginning of the file. If we precede G with a number, we can move to a specific line in
the document (e.g. 15G would move us to line 15).
Basic Unix – vi Editor Page 33 of 64
WK: 2 - Day: 2
Preceding the command with a number will delete multiple characters. For example, 10x will delete the character
selected and the next 9 characters; 10X will delete the 10 characters to the left of the currently selected character.
The command 5dw will delete 5 words, while 4dd deletes four lines.
Once the desired text is yanked, place the cursor in the spot in which we wish to insert the text and then use the put
command (p for line below or P for line above) to insert the contents of the buffer.
6.5 Undoing
Be sure to remember this command. When we make a mistake we can undo it. DO NOT move the cursor from the
line where we made the change. Then try using one of the following two commands:
u undoes the last change we made anywhere in the file. Using u again will "undo the undo".
U undoes all recent changes to current line. We can not have moved from the line to recover the original line.
combination. If we wished to delete 2 words we might use 2dw or d2w. Either of these combinations would work. So,
as we can see, the general format for a command can be
(number) (command) (text object) or (command) (number) (text object)
We might wish to try out some of the following combinations of commands and objects:
Command Description
d (delete)
w (word to the left)
y (yank/copy)
b (word to the right or backward)
c (change)
e (end of word)
H (top of the screen)
L (bottom of the screen)
M (middle of the screen)
0 (zero - first character on a line)
$ (end of a line)
( (previous sentence)
) (next sentence)
[ (previous section)
] (next section)
8. Repeating a Command
If we are doing repetitive editing, we may wish to use the same command over and over. vi will allow we to use the
dot (.) to repeat the last basic command we issued. If for example, we wished to deleted several lines, we could use
dd and then . (dot) in quick succession to delete a few lines.
A Quick Word about Customizing Your vi Environment
There are several options that we can set from within vi that can affect how we use vi. For example, one option allows
us to set a right margin that will then force vi to automatically wrap your lines as we type. To do this, we would use a
variation of the :set command. The :set command can be used to change various options in vi. In the example just
described, we could, while still in vi, type :set wrapmargin=10 to specify that we wish to have a right margin of 10.
Another useful option is :set number. This command causes vi to display line numbers in the file we are working on.
9. Useful vi Commands
9.1 Cut/Paste Commands
x delete one character (destructive backspace)
dw delete the current word (Note: ndw deletes n numbered words)
dd delete the current line (Note: ndd deletes n numbered lines)
D delete all content to the right of the cursor
Basic Unix – vi Editor Page 36 of 64
WK: 2 - Day: 2
d$ same as above
:u undo last command
p,P paste line starting one line below/above current cursor location
J combine the contents of two lines
"[a-z]nyy yank next n lines into named buffer [a-z]
"[a-z]p/P place the contents of selected buffer below/above the current line
a Begin appending text, one character to the right of current cursor location
A Begin appending text at the end of the current line
o/O Begin entering text one line below\above current line
ESC Exit insertion mode and return to command mode
-ctime time interval Locates files that that were created during the specified time interval
-mtime time interval Finds files that have been modified during the specified time interval
-atime time interval Locates files that have been accessed during the specified time interval
-perm permissions Locates files with certain permission settings
-user Locates files that have specified ownership
-group Locates files that are owned by specified group
-size Locates files with specified size
-type Locates a certain type of file
Time interval in options -ctime, -mtime and -atime is an integer with optional sign.
n: If the integer n does not have sign this means exactly n days ago, 0 means today.
+n: if it has plus sing, then it means "more then n days ago", or older then n,
-n: if it has the minus sign, then it means less than n days ago (-n), or younger then n.
It's evident that -1 and 0 are the same and means "today".
It is possible to locate files and directories that match or do not match multiple conditions, for example:
a to have multiple conditions ANDed
o to have multiple conditions ORed
! to negate a condition
expression to satisfy any complex condition
It is possible to specify the action to be taken on the files or directories that are found:
print prints the names of the files on standard output (usually enabled by default)
exec command executes the specified command.
The most common reason for using the find command is to utilize its capability to recursively process the
subdirectories. For example, if we want to obtain a list of all files accessed in the last 24 hours, execute the following
command:
$ find . -atime 0 -print
If the system administrator wants a list of .profile used by all users, the following command should be executed:
Basic Unix – Find and other Useful Commands Page 40 of 64
WK: 2 - Day: 3
-type c True if the file type is type c, where c is one of the following.
The find command checks the specified options, going from left to right, once for each file or directory encountered.
The simplest invocation of find can be used to create a list of all files and directories below the current directory:
$ find . -print
Basic Unix – Find and other Useful Commands Page 41 of 64
WK: 2 - Day: 3
We can use regular expressions to select files, for example those that have a .html suffix):
$ find . -name "*.html: -print
We can search for files more recent than, older than, or exactly the same age as a specified date,
-n - more recent then n days old
+n - older then n days old
n exactly of age n
Here are some useful examples. To find html files that have been modified in the last seven days, we can use -mtime
with the argument -7 (include the hyphen):
$ find . -mtime -7 -name "*.html" -print
If we just use the number 7 (without a hyphen), we will match only html files that were modified exactly seven days
ago:
$ find . -mtime 7 -name "*.html" -print
19. We can specify more than one directory as a starting point for the search. To look across the /bin and /usr
directory trees for filenames that contain the pattern \.htm, we can use the following command:
$ find /usr /bin -name "*\.htm*" -print
20. To find a list of the directories use the -type specifier. Here's one example:
find . -type d -print
The most typical options for -type are as following:
d -Directory
f - File
l - Link
2. Mail
The mail command is a quick and easy way to send an email to someone. Just type mail and the address of the
person we want to mail. We will then be prompted for a subject and any cc's. Then just type our message and
control-d on a line by itself to send the mail
2.2 Mailboxes
UNIX uses two mailboxes to hold mail messages
system mailbox (/vsr/spool/mail/)
user mail box (..../.../mbox)
Mail arrives in the system mailbox, and is saved in our user mail box after we have read it. The user mail box is
normally located in their $HOME directory.
To list all mails which he got say command only mail with out any arguments Type
$mail
The mail program displays a title message and lists all available mail headers,
SCO System V Mail (version 3.2) Type ? for help.
"/usr/spool/mail/brianb": 3 messages 3 new
N 3 brianb Mon May 31 15:02 10/299 My message3
N 2 brianb Mon May 31 15:01 9/278
>N 1 brianb Mon May 31 15:00 12/415 My first message
&
This initial screen displays the subject fields of messages which have arrived. The format of the display is,
Type Message_number From_User Date/Time Subject
3. write
The write command is used to send on-line messages to another user on the same machine. The format of the write
command is as follows:
$ write username
text of message
Basic Unix – Find and other Useful Commands Page 43 of 64
WK: 2 - Day: 3
^D
After typing the command, we enter our message, starting on the next line, terminating with the end-of-file character.
The recipient will then hear a bleep, then receive our message on screen, with a short header attached. The following
is a typical exchange. User UserRavi types:
$ write UserRavi
Hi there - want to go to lunch?
^D
$
User lnp8zz will hear a beep and the following will appear on his/her screen:
Message from UserRavi on sun050 at 12:42
Hi there - want to go to lunch?
EOF
If UserAjay wasn't logged on, the sender would see the following:
$ write UserAjay
UserAjay not logged in.
4. talk
The command talk creates a two-way, screen-oriented communication program. It allows users to type
simultaneously, with their output displayed in separate regions of the screen. To send a talk request, type talk and
then <user name>@<their machine>. So, for example, to send a talk request to dmb, who is logged on to cslab0a,
type talk raju@wilshire. The recipient of the request is then prompted for a response.
Example:
$ talk <username>
$ talk username@<their machinename> [to send a message to remote user]
Note: The mesg command may, of course, be used to disable write,talk access to a terminal.
We can stop messages being flashed up on our screen if we wish. To turn off direct communications type:
% mesg n
It will remain off for the remainder of our session, unless we type:
% mesg y
To turn the facility back on. Typing just mesg lets we know whether it is on or off.
5. walls
wall is a Unix command line utility. That would only invoked by Administrator, it displays the contents of a file or
standard input to all logged-in users.
Invocation
wall is invoked as follows:
$ raju@wilshiresoft:# wall <filename>
To take standard input as the file, omit the filename.
Output:
When invoked, wall displays the following output:
Broadcast Message from raju@wilshiresoft
(/dev/pts/0) at 01:23 ...
Exercises:
1. Send a message to another user on our Unix system, and get them to reply.
2. Create a small text file and send it to another user.
3. When we receive a message, save it to a file other than our mailbox. (Remember we can always send
yourself a message if we don't have one.)
4. Send a message to a user on a different computer system.
5. Send a note to our course tutor telling him that we can use mail now.
Basic Unix – Advanced Filters Page 45 of 64
WK: 2 - Day: 4
Advanced Filters
1. grep
The grep program searches a file or files for lines that have a certain pattern. The syntax is:
$ grep "pattern" file(s)
The name "grep" derives from the ed (a Unix line editor) command g/re/p, which means "globally search for a regular
expression and print all lines containing it." A regular expression is either some plain text (a word, for example) and/or
special characters used for pattern matching. When we learn more about regular expressions, we can use them to
specify complex patterns of text.
The simplest use of grep is to look for a pattern consisting of a single word. It can be used in a pipe so that only those
lines of the input files containing a given string are sent to the standard output. But let's start with an example reading
from files: searching all files in the working directory for a word--say, Unix. We'll use the wildcard * to quickly give
grep all filenames in the directory.
$ grep "Unix" *
ch01:Unix is a flexible and powerful operating system
ch01:When the Unix designers started work, little did
ch05:What can we do with Unix?
$
When grep searches multiple files, it shows the filename where it finds each matching line of text. Alternatively, if we
don't give grep a filename to read, it reads its standard input; that's the way all filter programs work:
$ ls -l | grep "Aug"
-rw-rw-rw- 1 john doc 11008 Aug 6 14:10 ch02
-rw-rw-rw- 1 john doc 8515 Aug 6 15:30 ch07
-rw-rw-r-- 1 john doc 2488 Aug 15 10:51 intro
-rw-rw-r-- 1 carol doc 1605 Aug 23 07:35 macros
$
First, the example runs ls -l to list our directory. The standard output of ls -l is piped to grep, which only outputs lines
that contain the string Aug (that is, files that were last modified in August). Because the standard output of grep isn't
redirected, those lines go to the terminal screen.
grep options let we modify the search. Given table lists some of the options.
Option Description
-l Print only the names of files with matching lines (lowercase letter "L").
Next, let's use a regular expression that tells grep to find lines with carol, followed by zero or more other characters
(abbreviated in a regular expression as ".*"), then followed by Aug:
[15] Note that the regular expression for "zero or more characters," ".*", is different than the corresponding filename
wildcard "*". We can't cover regular expressions in enough depth here to explain the difference--though more-detailed
books do. As a rule of thumb, remember that the first argument to grep is a regular expression; other arguments, if
any, are filenames that can use wildcards.
$ ls -l | grep "carol.*Aug"
-rw-rw-r-- 1 carol doc 1605 Aug 23 07:35 macros
$
2. sed
Basic Unix – Advanced Filters Page 46 of 64
WK: 2 - Day: 4
sed (which stands for Stream EDitor) is a simple but powerful computer program used to apply various pre-specified
textual transformations to a sequential stream of text data. It reads input files line by line, edits each line according to
rules specified in its simple language (the sed script), and then outputs the line. While originally created as a Unix
utility by Lee E. McMahon of Bell Labs from 1973 to 1974, sed is now available for virtually every operating system
that supports a command line.
2.1 Functions
sed is often thought of as a non-interactive text editor. It differs from conventional text editors in that the processing of
the two inputs is inverted. Instead of iterating once through a list of edit commands applying each one to the whole
text file in memory, sed iterates once through the text file applying the whole list of edit commands to each line.
Because only one line at a time is in memory, sed can process text files with an arbitrarily-large number of lines.
Some implementations of sed can only process lines of limited lengths.
sed's command set is modeled after the ed editor, and most commands work similarly in this inverted paradigm. For
example, the command 25d means if this is line 25, then delete (don't output) it, rather than go to line 25 and delete it
as it does in ed. The notable exceptions are the copy and move commands, which span a range of lines and thus
don't have straightforward equivalents in sed. Instead, sed introduces an extra buffer called the hold space, and
additional commands to manipulate it. The ed command to copy line 25 to line 76 (25t76) for example would be
coded as two separate commands in sed (25h; 76g), to store the line in the hold space until the point at which it
should be retrieved.
Usage:
The following example shows a typical usage of sed, where the -e option indicates that the sed expression follows:
$ sed -e 's/oldstuff/newstuff/g' inputFileName > outputFileName
The s stands for substitute; the g stands for global, which means that all matching occurrences in the line would be
replaced. After the first slash is the regular expression to search for and after the second slash is the expression to
replace it with. The substitute command (s///) is by far the most powerful and most commonly used sed command.
Under Unix, sed is often used as a filter in a pipeline:
$ generate_data | sed -e 's/x/y/'
That is, generate the data, but make the small change of replacing x with y.
Several substitutions or other commands can be put together in a file called, for example, subst.sed and then be
applied using the -f option to read the commands from the file:
$ sed -f subst.sed inputFileName > outputFileName
Besides substitution, other forms of simple processing are possible. For example, the following deletes empty lines or
lines that only contain spaces:
$ sed -e '/^ *$/d' inputFileName
This example used some of the following regular expression metacharacters:
The caret (^) matches the beginning of the line.
The dollar sign ($) matches the end of the line.
A period (.) matches any single character.
The asterisk (*) matches zero or more occurrences of the previous character.
A bracketed expression delimited by [ and ] matches any of the characters inside the brackets.
Complex sed constructs are possible, to the extent that it can be conceived of as a highly specialised, albeit simple,
programming language. Flow of control, for example, can be managed by use of a label (a colon followed by a string
which is to be the label name) and the branch instruction b; an instruction b followed by a valid label name will move
processing to the block following the label; if the label does not exist then the branch will end the script.
A third one should be added to remove all blanks and tabs immediately before the end of line:
$ sed -e 's/#.*//' -e 's/[ ^I]*$//' -e '/^$/ d'
The character "^I" is a CRTL-I or tab character. We would have to explicitly type in the tab. Note the order of
operations above, which is in that order for a very good reason. Comments might start in the middle of a line, with
white space characters before them. Therefore comments are first removed from a line, potentially leaving white
space characters that were before the comment. The second command removes all trailing blanks, so that lines that
Basic Unix – Advanced Filters Page 47 of 64
WK: 2 - Day: 4
are now blank are converted to empty lines. The last command deletes empty lines. Together, the three commands
remove all lines containing only comments, tabs or spaces.
This demonstrates the pattern space sed uses to operate on a line. The actual operation sed uses is:
Copy the input line into the pattern space.
Apply the first sed command on the pattern space, if the address restriction is true.
Repeat with the next sed expression, again
operating on the pattern space.
When the last operation is performed, write out the pattern space
and read in the next line from the input file.
n replace nth instance of pattern with replacement
g replace all instances of pattern with replacement
p write pattern space to STDOUT if a successful substitution takes place
w file Write the pattern space to file if a successful substitution takes place
We've covered quoting and regular expression. That's 90% of the effort needed to learn the substitute command. To
put it another way, we already know how to handle 90% of the most frequent uses of sed. There are a few fine points
that must be covered.
3. awk
Basic Unix – Advanced Filters Page 48 of 64
WK: 2 - Day: 4
The name awk comes from the initials of its designers: Alfred V. Aho, Peter J. Weinberger, and Brian W. Kernighan.
The original version of awk was written in 1977 at AT&T Bell Laboratories. In 1985 a new version made the
programming language more powerful, introducing user-defined functions, multiple input streams, and computed
regular expressions. This new version became generally available with Unix System V Release 3.1.
.awk is a programming language designed to search for, match patterns, and perform actions on files. awk programs
are generally quite small, and are interpreted. This makes it a good language for prototyping.
input lines to awk:
When awk scans an input line, it breaks it down into a number of fields. Fields are separated by a space or tab
character. Fields are numbered beginning at one, and the dollar symbol ($) is used to represent a field.
For instance, the following line in a file
I like money.
Has three fields. They are
$1 I
$2 like
$3 money.
Field zero ($0) refers to the entire line. awk scans lines from a file(s) or standard input.
Note: The most frustrating thing about trying to learn awk is getting our program past the shell's parser. The proper
way is to use single quotes around the program, like so:
$ awk '{print $0}' filename
The single quotes protect almost everything from the shell. In csh or tcsh, we still have to watch out for exclamation
marks, but other than that, we're safe.
$ awk '{print $2,$1}' filename
Will print the second field, then the first. All other fields are ignored
Variables
o Need not be declared
o May contain any type of data, their data type may change over the life of the program
o As in C, case matters; since all the built-in variables are all uppercase, avoid this form.
o Some of the commonly used built-in variables are:
NR -- The current line's sequential number
NF -- The number of fields in the current line
FS -- The input field separator; defaults to whitespace and is reset
by the -F command line parameter
OFS -- Output Field Separator default ' '
o RS -- Input Record Separator default \n
o FILENAME -- The name of the file currently being processed
Fields
o Each record is separated into fields named $1, $2, etc
o $0 is the entire record
o NF contains the number of fields in the current line
Print
o print prints each of the values of $1 through $NF separated by OFS then prints a \n onto stdout; the
default value of OFS is a blank
Example:
See a file(sample) contains these contents.
ramu:8:17:d
RAMU:8:17:D
king:89:37:j
smith:8:17:c
scott:19:4:d
allen:73:99:f
Basic Unix – Advanced Filters Page 49 of 64
WK: 2 - Day: 4
We may need to refer frequently to a file called table1 in the tables subdirectory when we are in our home directory
(jsmythe). Rather than typing the whole pathname
project1/chapter3/tables/table1
we could link the file to our home directory. If we are in the home directory, the command to do this is:
$ ln project1/chapter3/tables/table1 mytable1
To create the link when we are in the tables directory, the command would have been:
$ ln table1 ~/mytable1
After issuing either of these commands, an ls command in our home directory would show an entry for mytable1. The
long format of the same command would show 2 links for the file mytable1:
$ ls -l
-rw------- 2 jsmythe 6 Jul 4 14:23 mytable1
A long format listing of the file table1 would also show 2 links. What if a file called mytable1 had already existed in
our home directory? Unix would let we know that a file by that name exists, and would not make the link.
The effect of linking is that the file now has two names. We may call up the file by either name. Creating a link does
not change the ownership, group, or permissions for a file. The inode number of two files are same.
$ ln -s project1/chapter3 linkdir
The -s option means that a symbolic link is created (instead of the default hard link). This command creates a
symbolic link called linkdir which points to the directory called chapter3.
When we list the contents of linkdir
$ ls linkdir
$ data tables
we see a listing of the contents (data and tables)
We can learn more about the new directory by using the long format and directory options with the ls command:
$ ls -ld linkdir
l--------- 1 staff 7 Jun 11 13:27 linkdir -> project1/chapter3
Symbolic links are shown with an arrow (->) in the name column at the right.
2. Unix Processes
A process is an instance of running a program. If, for example, three people are running the same program
simultaneously, there are three processes there, not just one. In fact, we might have more than one process running
even with only person executing the program, because (we will see later) the program can ``split into two,'' making
two processes out of one.
Keep in mind that all Unix commands, e.g. cc and mail, are programs, and thus contribute processes to the system
when they are running. If 10 users are running mail right now, that will be 10 processes. At any given time, a typical
Unix system will have many active processes, some of which were set up when the machine was first powered up.
Every time we issue a command, Unix starts a new process, and suspends the current process (the C-shell) until the
new process completes (except in the case of background processes, to be discussed later).
Unix identifies every process by a Process Identification Number (pid) which is assigned when the process is
initiated. When we want to perform an operation on a process, we usually refer to it by its pid.
Unix is a timesharing system, which means that the processes take turns running. Each turn is a called a timeslice;
on most systems this is set at much less than one second. The reason this turns-taking approach is used is fairness:
We don't want a 2-second job to have to wait for a 5-hour job to finish, which is what would happen if a job had the
CPU to itself until it completed.1
The TT information gives terminal names, which we can see by typing the who command. E.g. we see p2 in the TT
column above, which is the terminal listed as ttyp2 in the who command.
The state of the job is given by a sequence of four letters, for example, `RWNA'. The first of these four is typically one
of the following:
first letter runnability of the process R runnable process T stopped process S process sleeping for less than about 20
seconds we processes that are idle (sleeping longer than about 20 seconds)
A state-R process is runnable, i.e. it is be able to make use of a turn given to it, and is waiting for one. We can put a
process in state T, i.e. stop the process, by typing control-z. Suppose, for example, that we are using ftp to get some
files from some archive site, and we notice a file there called README. we can use the ftp `get' command to get the
README file, and then type C-z. This will stop (suspend) the ftp process, and get me back to the C-shell. At that
point we can read the README file, say using more, and then reactivate the ftp process, by typing `fg' to the shell.
A typical example of an S/I process is one that is waiting for user input. If we are using the emacs editor, for example,
the process will go to state S when it is waiting for me to type something; if we take more than 20 seconds to decide
what to type, the process will be in state I.
When a process is running, backgrounded or suspended, it will be entered onto a list along with a job number. To
examine this list, type
$ jobs
An example of a job list could be
[1] Suspended sleep 100
[2] Running netscape
[3] Running nedit
To restart (foreground) a suspended processes, type
$ fg %jobnumber
For example, to restart sleep 100, type
$ fg %1
Typing fg with no job number foregrounds the last suspended process.
3. Killing a process
kill (terminate or signal a process), It is sometimes necessary to kill a process (for example, when an executing
program is in an infinite loop) To kill a job running in the foreground, type ^C (control c). For example, run
$ sleep 100
^C
To kill a suspended or background process, type
$ kill %jobnumber
For example, run
$ sleep 100 &
$ jobs
If it is job number 4, type
$ kill %4
To check whether this has worked, examine the job list again to see if the process has been removed.
Shell Programming
1. Why shell programming?
Even though there are various graphical interfaces available for Linux the shell still is a very neat tool. The shell is not
just a collection of commands but a really good programming language. We can automate a lot of tasks with it, the
shell is very good for system administration tasks, we can very quickly try out if our ideas work which makes it very
useful for simple prototyping and it is very useful for small utilities that perform some relatively simple tasks where
efficiency is less important than ease of configuration, maintenance and portability.
So let's see now how it works
1.2 Comments
Comments in shell programming start with # and go until the end of the line. We really recommend to use comments.
If we have comments and we don't use a certain script for some time we will still know immediately what it is doing
and how it works.
1.3 Variables
As in other programming languages we can't live without variables. In shell programming all variables have the
datatype string and we do not need to declare them. To assign a value to a variable we write:
varname=value
To get the value back we just put a dollar sign in front of the variable:
$!/bin/sh
Assign a value:
$ a="hello world"
Now print the content of "a":
echo "A is:"
echo $a
Type this lines into our text editor and save it e.g. as first. Then make the script executable by typing chmod +x
first in the shell and then start it by typing ./first
The script will just print:
A is:
hello world
Sometimes it is possible to confuse variable names with the rest of the text:
$num=2
Basic Unix – Shell Programming Page 57 of 64
WK: 2 - Day: 6
Example:
Write these lines in a file
Echo “First argument is $1”
Basic Unix – Shell Programming Page 58 of 64
WK: 2 - Day: 6
1.6 Quoting
Before passing any arguments to a program the shell tries to expand wildcards and variables. To expand means that
the wildcard (e.g. *) is replaced by the appropriate file names or that a variable is replaced by its value. To change
this behaviour we can use quotes: Let's say we have a number of files in the current directory. Two of them are jpg-
files, mail.jpg and tux.jpg.
$ !/bin/sh
echo *.jpg
This will print "mail.jpg tux.jpg".
Quotes (single and double) will prevent this wildcard expansion:
$ !/bin/sh
echo "*.jpg"
echo '*.jpg'
This will print "*.jpg" twice.
Single quotes are most strict. They prevent even variable expansion. Double quotes prevent wildcard expansion but
allow variable expansion:
$ !/bin/sh
echo $SHELL
echo "$SHELL"
echo '$SHELL'
This will print:
/bin/bash
/bin/bash
$SHELL
Finally there is the possibility to take the special meaning of any single character away by preceeding it with a
backslash:
echo \*.jpg
echo \$SHELL
This will print:
*.jpg
$SHELL
4. Conditional Statements
4.1. if Statements
The core conditional concept of wrapping a block of statements that is only to be processed if some condition is
met.
Shells also support else and the combined elif else-if condition, 3 basic layouts for if statements are shown
below. Note the use of the then and keyword to separate the condition commands from internal command block,
and the fi keyword to mark the end of the block.
First Model:
if condition-command
then
command1
command2
...
fi
Second Model:
if condition-command
then
commandA1
commandA2
...
else
commandB1
commandB2
...
fi
Third model:
Basic Unix – Shell Programming Page 60 of 64
WK: 2 - Day: 6
if condition-command-A
then
commandA1
commandA2
...
elif condition-command-B
then
commandB1
commandB2
...
else
commandC1
commandC2
...
fi
The commands inside the blocks are used the same as any other command within the system, and it is possible to
nest other conditions statements inside those blocks. For conciseness, many people will use the semicolon (;)
command separating character to allow the then keyword to be placed on the same line as the if and the
condition-command. (Note that no semicolon is needed after the then keyword)
The condition-command is a any command, and the if statement is evaluated based on the success (exit status) of
that command.
Example:
if ls -al ; then
echo "Directory was successfully listed"
else
echo "Failed to list directory"
fi
An example would be where a program needs to output something to a file, but first checks that the file exists:
#!/bin/sh
echo “Enter the filename”
Basic Unix – Shell Programming Page 61 of 64
WK: 2 - Day: 6
read file1
if [ ! -s file1 ] or if test ! –s file1
then
echo "file1 is empty or does not exist."
ls -l > file1
exit
else
echo "File file1 already exists."
fi
exit 0
An example would checks that the given strings are same or not:
#!/bin/sh
echo “Enter the strings(string1,string2)”
read str1
read str2
if test str1 = str2
then
echo "Both Strings are equal"
exit
else
echo "Given strings are not equal"
fi
exit 0
Example:
$!/bin/sh
if test $# -le 5
then
echo Less than or equal to five parameters.
else
echo More than 5 parameters.
fi
exit 0
Basic Unix – Shell Programming Page 62 of 64
WK: 2 - Day: 6
$!/bin/sh
# an example with the case statement, which reads a command from the user and processes it
echo "Enter your command (who, list, or cal)"
read command
case "$command" in
who)
echo "Running who..."
who
;;
list)
Basic Unix – Shell Programming Page 63 of 64
WK: 2 - Day: 6
Example:
#!/bin/sh
for i in Claude Paul Wayne Roger Tom
do
echo The name is $i.
done
exit 0
Within the shell, parameters are read as $1 for the first parameter, $2 for the second parameter, $3 for the third
parameter, and so on. $* is the entire list of parameters.
If the ``in list" is omitted, the list taken is the list of parameters passed to the shell on the command line.
Note: To excite above said example programs we have to compile them in a file and excite that file. Like
$ sh <filename>
5. Debugging
The simplest debugging help is of course the command echo. We can use it to print specific variables around the
place where we suspect the mistake. This is probably what most shell programmers use 80% of the time to track
down a mistake. The advantage of a shell script is that it does not require any re-compilation and inserting an "echo"
statement is done very quickly.
The shell has a real debug mode as well. If there is a mistake in our script "strangescript" then we can debug it like
this:
$ sh -x strangescript
This will execute the script and show all the statements that get executed with the variables and wildcards already
expanded.
The shell also has a mode to check for syntax errors without actually executing the program. To use this run:
$ sh -n our_script
If this returns nothing then our program is free of syntax errors.