# HG changeset patch # User amit@thunder # Date 1265917281 -19800 # Node ID 52d12eb31c30b12ad81a602539a81a56b523169c # Parent 672eaaab9204c762f585100f0b992dd4ac989d77 Virtual enviroment for SEES-hacks added ... diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/scripts/autoid.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/scripts/autoid.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,62 @@ +#!/usr/bin/env python +# +# Add unique ID attributes to para tags. This script should only be +# run by one person, since otherwise it introduces the possibility of +# chaotic conflicts among tags. + +import glob, os, re, sys + +tagged = re.compile(']* id="x_([0-9a-f]+)"[^>]*>', re.M) +untagged = re.compile('') + +names = glob.glob('ch*.docbook') +# First pass: find the highest-numbered paragraph ID. + + +chapter=None +seen = set() +errs = 0 +beginning="p_list= " + +id_file=open('p_list.py','w') +dictionary={} +id_list=[] +for name in names: + for m in tagged.finditer(open(name).read()): + i = int(m.group(1),16) + if i in seen: + print >> sys.stderr, '%s: duplication of ID %s' % (name, i) + errs += 1 + seen.add(i) + if i > biggest_id: + biggest_id = i + + +def retag(s): + global biggest_id + global chapter + biggest_id += 1 + + id_name="%s_%x" % (chapter,biggest_id) + id_list.append(id_name) + + return '' %id_name + +# Second pass: add IDs to paragraphs that currently lack them. + +for name in names: + biggest_id=0 + chapter=name.split('.')[0] + id_list=[] + f = open(name).read() + f1 = untagged.sub(retag, f ) + dictionary[chapter]=id_list + if f1 != f: + tmpname = name + '.tmp' + fp = open(tmpname, 'w') + fp.write(f1) + fp.close() + os.rename(tmpname, name) +p_lists_string=beginning+str(dictionary) +id_file.write(p_lists_string) +sys.exit(errs) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/scripts/docbook2html.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/scripts/docbook2html.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,9 @@ +import glob, os, re, sys +import subprocess +import time +names = glob.glob('ch*.docbook') +for name in names: + command="xsltproc html-single.xsl %s > %s.html" %(name ,name.split('.')[0]) + print command + subprocess.Popen(command,shell=True) + time.sleep(10) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/scripts/html-single.xsl --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/scripts/html-single.xsl Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,32 @@ + + + + + + + + + + + .png + figs/ + + .png + images/callouts/ + + + 1 + + + 0 + 0 + yes + hgbook.css + + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/scripts/index.config --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/scripts/index.config Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,14 @@ +Introduction/Introduction.rst +basic_python/intro.rst +basic_python/list_tuples.rst +basic_python/strings_dicts.rst +basic_python/func.rst +basic_python/oop.rst +basic_python/interim_assessment.rst +ult/ult_module_plan.rst +ult/Using_Linux_Tools.rst +ult/session4.rst +ult/Section_5.rst +versionControl/index.rst +versionControl/versionControl.rst +versionControl/handOut.rst diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/scripts/myrst2xml.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/scripts/myrst2xml.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,88 @@ +#!/usr/bin/python +""" +Just a hack to convert rst to xml and then docbook . +May not containt all the required elements of a docbook . +Just done to make it run for the specific rst for our +sees documentation. +""" +import xml.etree.ElementTree as ET +import os +import re +import subprocess +import os +import pkg_resources +import time +from docutils.core import publish_file +chapterno=0 + + + + + + + + + +def convert2xml(file): +# print folder,subfolder,file + file=file.split("\n")[0] + file= file.split("\t")[0] + global chapterno + name=file.split('/')[-1] + name='ch'+str(chapterno)+name.split('.')[0] +# full_file=os.path.join(folder,file) +# if file.endswith('.rst'): + print file + xml_file=name+'.xml' + a=open(xml_file,'w') +# a.close() +# command="rst2xml.py %s > %s" %(file , xml_file) + +# a=subprocess.Popen(command , shell=True) +# time.sleep(10) +# a.kill() + publish_file(source_path=file, destination_path=xml_file,parser_name='restructuredtext', writer_name='xml') +def walk(repo): + global chapterno + mainfolder='/home/amit/sttp_latest/' + for readline in open('index.config','r').readlines(): + chapterno+=1 + filename=mainfolder+readline + convert2xml(filename) + + + + + +if __name__=='__main__': + repo='/home/amit/sphinxvirt/review/' + walk(repo) +# convert(1,2,3) + + + + + + + + + + + + + + + + + + + + + + + + + + + + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/scripts/p_list.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/scripts/p_list.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1 @@ +p_list= {'ch12index': [], 'ch6oop': ['ch6oop_1', 'ch6oop_2', 'ch6oop_3', 'ch6oop_4', 'ch6oop_5', 'ch6oop_6'], 'ch9Using_Linux_Tools': ['ch9Using_Linux_Tools_1', 'ch9Using_Linux_Tools_2', 'ch9Using_Linux_Tools_3', 'ch9Using_Linux_Tools_4', 'ch9Using_Linux_Tools_5', 'ch9Using_Linux_Tools_6', 'ch9Using_Linux_Tools_7', 'ch9Using_Linux_Tools_8', 'ch9Using_Linux_Tools_9', 'ch9Using_Linux_Tools_a', 'ch9Using_Linux_Tools_b', 'ch9Using_Linux_Tools_c', 'ch9Using_Linux_Tools_d', 'ch9Using_Linux_Tools_e', 'ch9Using_Linux_Tools_f', 'ch9Using_Linux_Tools_10', 'ch9Using_Linux_Tools_11', 'ch9Using_Linux_Tools_12', 'ch9Using_Linux_Tools_13', 'ch9Using_Linux_Tools_14', 'ch9Using_Linux_Tools_15', 'ch9Using_Linux_Tools_16', 'ch9Using_Linux_Tools_17', 'ch9Using_Linux_Tools_18', 'ch9Using_Linux_Tools_19', 'ch9Using_Linux_Tools_1a', 'ch9Using_Linux_Tools_1b', 'ch9Using_Linux_Tools_1c', 'ch9Using_Linux_Tools_1d', 'ch9Using_Linux_Tools_1e', 'ch9Using_Linux_Tools_1f', 'ch9Using_Linux_Tools_20', 'ch9Using_Linux_Tools_21', 'ch9Using_Linux_Tools_22', 'ch9Using_Linux_Tools_23', 'ch9Using_Linux_Tools_24', 'ch9Using_Linux_Tools_25', 'ch9Using_Linux_Tools_26', 'ch9Using_Linux_Tools_27', 'ch9Using_Linux_Tools_28', 'ch9Using_Linux_Tools_29', 'ch9Using_Linux_Tools_2a', 'ch9Using_Linux_Tools_2b', 'ch9Using_Linux_Tools_2c', 'ch9Using_Linux_Tools_2d', 'ch9Using_Linux_Tools_2e', 'ch9Using_Linux_Tools_2f', 'ch9Using_Linux_Tools_30', 'ch9Using_Linux_Tools_31', 'ch9Using_Linux_Tools_32', 'ch9Using_Linux_Tools_33', 'ch9Using_Linux_Tools_34', 'ch9Using_Linux_Tools_35', 'ch9Using_Linux_Tools_36', 'ch9Using_Linux_Tools_37', 'ch9Using_Linux_Tools_38', 'ch9Using_Linux_Tools_39', 'ch9Using_Linux_Tools_3a', 'ch9Using_Linux_Tools_3b', 'ch9Using_Linux_Tools_3c', 'ch9Using_Linux_Tools_3d', 'ch9Using_Linux_Tools_3e', 'ch9Using_Linux_Tools_3f', 'ch9Using_Linux_Tools_40', 'ch9Using_Linux_Tools_41', 'ch9Using_Linux_Tools_42', 'ch9Using_Linux_Tools_43', 'ch9Using_Linux_Tools_44', 'ch9Using_Linux_Tools_45', 'ch9Using_Linux_Tools_46', 'ch9Using_Linux_Tools_47', 'ch9Using_Linux_Tools_48', 'ch9Using_Linux_Tools_49', 'ch9Using_Linux_Tools_4a', 'ch9Using_Linux_Tools_4b', 'ch9Using_Linux_Tools_4c', 'ch9Using_Linux_Tools_4d', 'ch9Using_Linux_Tools_4e', 'ch9Using_Linux_Tools_4f', 'ch9Using_Linux_Tools_50', 'ch9Using_Linux_Tools_51', 'ch9Using_Linux_Tools_52', 'ch9Using_Linux_Tools_53', 'ch9Using_Linux_Tools_54', 'ch9Using_Linux_Tools_55', 'ch9Using_Linux_Tools_56', 'ch9Using_Linux_Tools_57', 'ch9Using_Linux_Tools_58', 'ch9Using_Linux_Tools_59', 'ch9Using_Linux_Tools_5a', 'ch9Using_Linux_Tools_5b', 'ch9Using_Linux_Tools_5c', 'ch9Using_Linux_Tools_5d', 'ch9Using_Linux_Tools_5e', 'ch9Using_Linux_Tools_5f', 'ch9Using_Linux_Tools_60', 'ch9Using_Linux_Tools_61', 'ch9Using_Linux_Tools_62', 'ch9Using_Linux_Tools_63', 'ch9Using_Linux_Tools_64', 'ch9Using_Linux_Tools_65', 'ch9Using_Linux_Tools_66', 'ch9Using_Linux_Tools_67', 'ch9Using_Linux_Tools_68', 'ch9Using_Linux_Tools_69', 'ch9Using_Linux_Tools_6a', 'ch9Using_Linux_Tools_6b', 'ch9Using_Linux_Tools_6c', 'ch9Using_Linux_Tools_6d', 'ch9Using_Linux_Tools_6e', 'ch9Using_Linux_Tools_6f', 'ch9Using_Linux_Tools_70', 'ch9Using_Linux_Tools_71', 'ch9Using_Linux_Tools_72', 'ch9Using_Linux_Tools_73', 'ch9Using_Linux_Tools_74', 'ch9Using_Linux_Tools_75', 'ch9Using_Linux_Tools_76', 'ch9Using_Linux_Tools_77', 'ch9Using_Linux_Tools_78', 'ch9Using_Linux_Tools_79', 'ch9Using_Linux_Tools_7a', 'ch9Using_Linux_Tools_7b', 'ch9Using_Linux_Tools_7c', 'ch9Using_Linux_Tools_7d', 'ch9Using_Linux_Tools_7e', 'ch9Using_Linux_Tools_7f', 'ch9Using_Linux_Tools_80', 'ch9Using_Linux_Tools_81', 'ch9Using_Linux_Tools_82', 'ch9Using_Linux_Tools_83', 'ch9Using_Linux_Tools_84', 'ch9Using_Linux_Tools_85', 'ch9Using_Linux_Tools_86', 'ch9Using_Linux_Tools_87', 'ch9Using_Linux_Tools_88', 'ch9Using_Linux_Tools_89', 'ch9Using_Linux_Tools_8a', 'ch9Using_Linux_Tools_8b', 'ch9Using_Linux_Tools_8c', 'ch9Using_Linux_Tools_8d', 'ch9Using_Linux_Tools_8e', 'ch9Using_Linux_Tools_8f', 'ch9Using_Linux_Tools_90', 'ch9Using_Linux_Tools_91', 'ch9Using_Linux_Tools_92', 'ch9Using_Linux_Tools_93', 'ch9Using_Linux_Tools_94', 'ch9Using_Linux_Tools_95', 'ch9Using_Linux_Tools_96', 'ch9Using_Linux_Tools_97', 'ch9Using_Linux_Tools_98', 'ch9Using_Linux_Tools_99', 'ch9Using_Linux_Tools_9a', 'ch9Using_Linux_Tools_9b', 'ch9Using_Linux_Tools_9c', 'ch9Using_Linux_Tools_9d', 'ch9Using_Linux_Tools_9e', 'ch9Using_Linux_Tools_9f', 'ch9Using_Linux_Tools_a0', 'ch9Using_Linux_Tools_a1', 'ch9Using_Linux_Tools_a2', 'ch9Using_Linux_Tools_a3', 'ch9Using_Linux_Tools_a4', 'ch9Using_Linux_Tools_a5', 'ch9Using_Linux_Tools_a6', 'ch9Using_Linux_Tools_a7', 'ch9Using_Linux_Tools_a8', 'ch9Using_Linux_Tools_a9', 'ch9Using_Linux_Tools_aa', 'ch9Using_Linux_Tools_ab', 'ch9Using_Linux_Tools_ac', 'ch9Using_Linux_Tools_ad', 'ch9Using_Linux_Tools_ae', 'ch9Using_Linux_Tools_af', 'ch9Using_Linux_Tools_b0', 'ch9Using_Linux_Tools_b1', 'ch9Using_Linux_Tools_b2', 'ch9Using_Linux_Tools_b3', 'ch9Using_Linux_Tools_b4', 'ch9Using_Linux_Tools_b5', 'ch9Using_Linux_Tools_b6', 'ch9Using_Linux_Tools_b7', 'ch9Using_Linux_Tools_b8', 'ch9Using_Linux_Tools_b9', 'ch9Using_Linux_Tools_ba', 'ch9Using_Linux_Tools_bb', 'ch9Using_Linux_Tools_bc', 'ch9Using_Linux_Tools_bd', 'ch9Using_Linux_Tools_be', 'ch9Using_Linux_Tools_bf', 'ch9Using_Linux_Tools_c0', 'ch9Using_Linux_Tools_c1', 'ch9Using_Linux_Tools_c2', 'ch9Using_Linux_Tools_c3', 'ch9Using_Linux_Tools_c4', 'ch9Using_Linux_Tools_c5', 'ch9Using_Linux_Tools_c6', 'ch9Using_Linux_Tools_c7', 'ch9Using_Linux_Tools_c8', 'ch9Using_Linux_Tools_c9', 'ch9Using_Linux_Tools_ca', 'ch9Using_Linux_Tools_cb', 'ch9Using_Linux_Tools_cc', 'ch9Using_Linux_Tools_cd', 'ch9Using_Linux_Tools_ce', 'ch9Using_Linux_Tools_cf', 'ch9Using_Linux_Tools_d0', 'ch9Using_Linux_Tools_d1', 'ch9Using_Linux_Tools_d2', 'ch9Using_Linux_Tools_d3', 'ch9Using_Linux_Tools_d4', 'ch9Using_Linux_Tools_d5', 'ch9Using_Linux_Tools_d6', 'ch9Using_Linux_Tools_d7', 'ch9Using_Linux_Tools_d8', 'ch9Using_Linux_Tools_d9', 'ch9Using_Linux_Tools_da', 'ch9Using_Linux_Tools_db', 'ch9Using_Linux_Tools_dc', 'ch9Using_Linux_Tools_dd', 'ch9Using_Linux_Tools_de', 'ch9Using_Linux_Tools_df', 'ch9Using_Linux_Tools_e0', 'ch9Using_Linux_Tools_e1', 'ch9Using_Linux_Tools_e2', 'ch9Using_Linux_Tools_e3', 'ch9Using_Linux_Tools_e4', 'ch9Using_Linux_Tools_e5', 'ch9Using_Linux_Tools_e6', 'ch9Using_Linux_Tools_e7', 'ch9Using_Linux_Tools_e8', 'ch9Using_Linux_Tools_e9', 'ch9Using_Linux_Tools_ea', 'ch9Using_Linux_Tools_eb', 'ch9Using_Linux_Tools_ec', 'ch9Using_Linux_Tools_ed', 'ch9Using_Linux_Tools_ee', 'ch9Using_Linux_Tools_ef', 'ch9Using_Linux_Tools_f0', 'ch9Using_Linux_Tools_f1', 'ch9Using_Linux_Tools_f2', 'ch9Using_Linux_Tools_f3', 'ch9Using_Linux_Tools_f4', 'ch9Using_Linux_Tools_f5', 'ch9Using_Linux_Tools_f6', 'ch9Using_Linux_Tools_f7', 'ch9Using_Linux_Tools_f8', 'ch9Using_Linux_Tools_f9', 'ch9Using_Linux_Tools_fa', 'ch9Using_Linux_Tools_fb', 'ch9Using_Linux_Tools_fc', 'ch9Using_Linux_Tools_fd', 'ch9Using_Linux_Tools_fe', 'ch9Using_Linux_Tools_ff', 'ch9Using_Linux_Tools_100', 'ch9Using_Linux_Tools_101', 'ch9Using_Linux_Tools_102', 'ch9Using_Linux_Tools_103', 'ch9Using_Linux_Tools_104', 'ch9Using_Linux_Tools_105', 'ch9Using_Linux_Tools_106', 'ch9Using_Linux_Tools_107', 'ch9Using_Linux_Tools_108', 'ch9Using_Linux_Tools_109', 'ch9Using_Linux_Tools_10a', 'ch9Using_Linux_Tools_10b', 'ch9Using_Linux_Tools_10c', 'ch9Using_Linux_Tools_10d', 'ch9Using_Linux_Tools_10e', 'ch9Using_Linux_Tools_10f', 'ch9Using_Linux_Tools_110', 'ch9Using_Linux_Tools_111', 'ch9Using_Linux_Tools_112', 'ch9Using_Linux_Tools_113', 'ch9Using_Linux_Tools_114', 'ch9Using_Linux_Tools_115', 'ch9Using_Linux_Tools_116', 'ch9Using_Linux_Tools_117', 'ch9Using_Linux_Tools_118', 'ch9Using_Linux_Tools_119', 'ch9Using_Linux_Tools_11a', 'ch9Using_Linux_Tools_11b', 'ch9Using_Linux_Tools_11c', 'ch9Using_Linux_Tools_11d', 'ch9Using_Linux_Tools_11e', 'ch9Using_Linux_Tools_11f', 'ch9Using_Linux_Tools_120', 'ch9Using_Linux_Tools_121', 'ch9Using_Linux_Tools_122', 'ch9Using_Linux_Tools_123', 'ch9Using_Linux_Tools_124', 'ch9Using_Linux_Tools_125', 'ch9Using_Linux_Tools_126', 'ch9Using_Linux_Tools_127', 'ch9Using_Linux_Tools_128', 'ch9Using_Linux_Tools_129', 'ch9Using_Linux_Tools_12a', 'ch9Using_Linux_Tools_12b', 'ch9Using_Linux_Tools_12c', 'ch9Using_Linux_Tools_12d', 'ch9Using_Linux_Tools_12e', 'ch9Using_Linux_Tools_12f', 'ch9Using_Linux_Tools_130', 'ch9Using_Linux_Tools_131', 'ch9Using_Linux_Tools_132', 'ch9Using_Linux_Tools_133', 'ch9Using_Linux_Tools_134', 'ch9Using_Linux_Tools_135', 'ch9Using_Linux_Tools_136', 'ch9Using_Linux_Tools_137', 'ch9Using_Linux_Tools_138', 'ch9Using_Linux_Tools_139', 'ch9Using_Linux_Tools_13a', 'ch9Using_Linux_Tools_13b', 'ch9Using_Linux_Tools_13c', 'ch9Using_Linux_Tools_13d', 'ch9Using_Linux_Tools_13e', 'ch9Using_Linux_Tools_13f', 'ch9Using_Linux_Tools_140', 'ch9Using_Linux_Tools_141', 'ch9Using_Linux_Tools_142', 'ch9Using_Linux_Tools_143', 'ch9Using_Linux_Tools_144', 'ch9Using_Linux_Tools_145', 'ch9Using_Linux_Tools_146', 'ch9Using_Linux_Tools_147', 'ch9Using_Linux_Tools_148', 'ch9Using_Linux_Tools_149', 'ch9Using_Linux_Tools_14a', 'ch9Using_Linux_Tools_14b', 'ch9Using_Linux_Tools_14c', 'ch9Using_Linux_Tools_14d', 'ch9Using_Linux_Tools_14e', 'ch9Using_Linux_Tools_14f', 'ch9Using_Linux_Tools_150', 'ch9Using_Linux_Tools_151', 'ch9Using_Linux_Tools_152', 'ch9Using_Linux_Tools_153', 'ch9Using_Linux_Tools_154', 'ch9Using_Linux_Tools_155', 'ch9Using_Linux_Tools_156', 'ch9Using_Linux_Tools_157', 'ch9Using_Linux_Tools_158', 'ch9Using_Linux_Tools_159', 'ch9Using_Linux_Tools_15a', 'ch9Using_Linux_Tools_15b', 'ch9Using_Linux_Tools_15c', 'ch9Using_Linux_Tools_15d', 'ch9Using_Linux_Tools_15e', 'ch9Using_Linux_Tools_15f', 'ch9Using_Linux_Tools_160', 'ch9Using_Linux_Tools_161', 'ch9Using_Linux_Tools_162', 'ch9Using_Linux_Tools_163', 'ch9Using_Linux_Tools_164', 'ch9Using_Linux_Tools_165', 'ch9Using_Linux_Tools_166', 'ch9Using_Linux_Tools_167', 'ch9Using_Linux_Tools_168', 'ch9Using_Linux_Tools_169', 'ch9Using_Linux_Tools_16a', 'ch9Using_Linux_Tools_16b', 'ch9Using_Linux_Tools_16c', 'ch9Using_Linux_Tools_16d', 'ch9Using_Linux_Tools_16e', 'ch9Using_Linux_Tools_16f', 'ch9Using_Linux_Tools_170', 'ch9Using_Linux_Tools_171', 'ch9Using_Linux_Tools_172', 'ch9Using_Linux_Tools_173', 'ch9Using_Linux_Tools_174', 'ch9Using_Linux_Tools_175', 'ch9Using_Linux_Tools_176', 'ch9Using_Linux_Tools_177', 'ch9Using_Linux_Tools_178', 'ch9Using_Linux_Tools_179', 'ch9Using_Linux_Tools_17a', 'ch9Using_Linux_Tools_17b', 'ch9Using_Linux_Tools_17c', 'ch9Using_Linux_Tools_17d', 'ch9Using_Linux_Tools_17e', 'ch9Using_Linux_Tools_17f', 'ch9Using_Linux_Tools_180', 'ch9Using_Linux_Tools_181', 'ch9Using_Linux_Tools_182', 'ch9Using_Linux_Tools_183', 'ch9Using_Linux_Tools_184', 'ch9Using_Linux_Tools_185', 'ch9Using_Linux_Tools_186', 'ch9Using_Linux_Tools_187', 'ch9Using_Linux_Tools_188', 'ch9Using_Linux_Tools_189', 'ch9Using_Linux_Tools_18a', 'ch9Using_Linux_Tools_18b', 'ch9Using_Linux_Tools_18c', 'ch9Using_Linux_Tools_18d', 'ch9Using_Linux_Tools_18e', 'ch9Using_Linux_Tools_18f', 'ch9Using_Linux_Tools_190', 'ch9Using_Linux_Tools_191', 'ch9Using_Linux_Tools_192', 'ch9Using_Linux_Tools_193', 'ch9Using_Linux_Tools_194', 'ch9Using_Linux_Tools_195', 'ch9Using_Linux_Tools_196', 'ch9Using_Linux_Tools_197', 'ch9Using_Linux_Tools_198', 'ch9Using_Linux_Tools_199', 'ch9Using_Linux_Tools_19a', 'ch9Using_Linux_Tools_19b', 'ch9Using_Linux_Tools_19c', 'ch9Using_Linux_Tools_19d', 'ch9Using_Linux_Tools_19e', 'ch9Using_Linux_Tools_19f', 'ch9Using_Linux_Tools_1a0', 'ch9Using_Linux_Tools_1a1', 'ch9Using_Linux_Tools_1a2', 'ch9Using_Linux_Tools_1a3', 'ch9Using_Linux_Tools_1a4', 'ch9Using_Linux_Tools_1a5', 'ch9Using_Linux_Tools_1a6', 'ch9Using_Linux_Tools_1a7', 'ch9Using_Linux_Tools_1a8', 'ch9Using_Linux_Tools_1a9', 'ch9Using_Linux_Tools_1aa', 'ch9Using_Linux_Tools_1ab', 'ch9Using_Linux_Tools_1ac', 'ch9Using_Linux_Tools_1ad', 'ch9Using_Linux_Tools_1ae', 'ch9Using_Linux_Tools_1af', 'ch9Using_Linux_Tools_1b0', 'ch9Using_Linux_Tools_1b1', 'ch9Using_Linux_Tools_1b2', 'ch9Using_Linux_Tools_1b3', 'ch9Using_Linux_Tools_1b4', 'ch9Using_Linux_Tools_1b5', 'ch9Using_Linux_Tools_1b6', 'ch9Using_Linux_Tools_1b7', 'ch9Using_Linux_Tools_1b8', 'ch9Using_Linux_Tools_1b9', 'ch9Using_Linux_Tools_1ba', 'ch9Using_Linux_Tools_1bb', 'ch9Using_Linux_Tools_1bc', 'ch9Using_Linux_Tools_1bd', 'ch9Using_Linux_Tools_1be', 'ch9Using_Linux_Tools_1bf', 'ch9Using_Linux_Tools_1c0', 'ch9Using_Linux_Tools_1c1', 'ch9Using_Linux_Tools_1c2', 'ch9Using_Linux_Tools_1c3', 'ch9Using_Linux_Tools_1c4', 'ch9Using_Linux_Tools_1c5', 'ch9Using_Linux_Tools_1c6', 'ch9Using_Linux_Tools_1c7', 'ch9Using_Linux_Tools_1c8', 'ch9Using_Linux_Tools_1c9', 'ch9Using_Linux_Tools_1ca', 'ch9Using_Linux_Tools_1cb', 'ch9Using_Linux_Tools_1cc', 'ch9Using_Linux_Tools_1cd', 'ch9Using_Linux_Tools_1ce', 'ch9Using_Linux_Tools_1cf', 'ch9Using_Linux_Tools_1d0', 'ch9Using_Linux_Tools_1d1', 'ch9Using_Linux_Tools_1d2', 'ch9Using_Linux_Tools_1d3', 'ch9Using_Linux_Tools_1d4', 'ch9Using_Linux_Tools_1d5', 'ch9Using_Linux_Tools_1d6', 'ch9Using_Linux_Tools_1d7', 'ch9Using_Linux_Tools_1d8', 'ch9Using_Linux_Tools_1d9', 'ch9Using_Linux_Tools_1da', 'ch9Using_Linux_Tools_1db', 'ch9Using_Linux_Tools_1dc', 'ch9Using_Linux_Tools_1dd', 'ch9Using_Linux_Tools_1de', 'ch9Using_Linux_Tools_1df', 'ch9Using_Linux_Tools_1e0', 'ch9Using_Linux_Tools_1e1'], 'ch1Introduction': ['ch1Introduction_1', 'ch1Introduction_2', 'ch1Introduction_3', 'ch1Introduction_4', 'ch1Introduction_5', 'ch1Introduction_6', 'ch1Introduction_7', 'ch1Introduction_8', 'ch1Introduction_9', 'ch1Introduction_a', 'ch1Introduction_b', 'ch1Introduction_c', 'ch1Introduction_d'], 'ch5func': ['ch5func_1', 'ch5func_2', 'ch5func_3', 'ch5func_4', 'ch5func_5', 'ch5func_6', 'ch5func_7', 'ch5func_8', 'ch5func_9', 'ch5func_a', 'ch5func_b', 'ch5func_c', 'ch5func_d', 'ch5func_e', 'ch5func_f', 'ch5func_10', 'ch5func_11', 'ch5func_12', 'ch5func_13', 'ch5func_14', 'ch5func_15', 'ch5func_16', 'ch5func_17', 'ch5func_18', 'ch5func_19', 'ch5func_1a', 'ch5func_1b', 'ch5func_1c', 'ch5func_1d', 'ch5func_1e', 'ch5func_1f'], 'ch3list_tuples': ['ch3list_tuples_1', 'ch3list_tuples_2', 'ch3list_tuples_3', 'ch3list_tuples_4', 'ch3list_tuples_5', 'ch3list_tuples_6', 'ch3list_tuples_7', 'ch3list_tuples_8', 'ch3list_tuples_9', 'ch3list_tuples_a', 'ch3list_tuples_b', 'ch3list_tuples_c', 'ch3list_tuples_d', 'ch3list_tuples_e', 'ch3list_tuples_f', 'ch3list_tuples_10', 'ch3list_tuples_11', 'ch3list_tuples_12', 'ch3list_tuples_13', 'ch3list_tuples_14', 'ch3list_tuples_15', 'ch3list_tuples_16', 'ch3list_tuples_17', 'ch3list_tuples_18', 'ch3list_tuples_19', 'ch3list_tuples_1a', 'ch3list_tuples_1b', 'ch3list_tuples_1c', 'ch3list_tuples_1d', 'ch3list_tuples_1e', 'ch3list_tuples_1f', 'ch3list_tuples_20', 'ch3list_tuples_21', 'ch3list_tuples_22', 'ch3list_tuples_23', 'ch3list_tuples_24', 'ch3list_tuples_25', 'ch3list_tuples_26', 'ch3list_tuples_27', 'ch3list_tuples_28', 'ch3list_tuples_29', 'ch3list_tuples_2a', 'ch3list_tuples_2b', 'ch3list_tuples_2c', 'ch3list_tuples_2d', 'ch3list_tuples_2e', 'ch3list_tuples_2f', 'ch3list_tuples_30', 'ch3list_tuples_31', 'ch3list_tuples_32', 'ch3list_tuples_33', 'ch3list_tuples_34', 'ch3list_tuples_35', 'ch3list_tuples_36', 'ch3list_tuples_37'], 'ch2intro': ['ch2intro_1', 'ch2intro_2', 'ch2intro_3', 'ch2intro_4', 'ch2intro_5', 'ch2intro_6', 'ch2intro_7', 'ch2intro_8', 'ch2intro_9', 'ch2intro_a', 'ch2intro_b', 'ch2intro_c', 'ch2intro_d', 'ch2intro_e', 'ch2intro_f', 'ch2intro_10', 'ch2intro_11', 'ch2intro_12', 'ch2intro_13', 'ch2intro_14', 'ch2intro_15', 'ch2intro_16', 'ch2intro_17', 'ch2intro_18', 'ch2intro_19', 'ch2intro_1a', 'ch2intro_1b', 'ch2intro_1c', 'ch2intro_1d', 'ch2intro_1e', 'ch2intro_1f', 'ch2intro_20', 'ch2intro_21', 'ch2intro_22', 'ch2intro_23', 'ch2intro_24', 'ch2intro_25', 'ch2intro_26', 'ch2intro_27', 'ch2intro_28', 'ch2intro_29', 'ch2intro_2a', 'ch2intro_2b', 'ch2intro_2c', 'ch2intro_2d', 'ch2intro_2e', 'ch2intro_2f', 'ch2intro_30', 'ch2intro_31', 'ch2intro_32', 'ch2intro_33', 'ch2intro_34', 'ch2intro_35', 'ch2intro_36', 'ch2intro_37', 'ch2intro_38', 'ch2intro_39', 'ch2intro_3a', 'ch2intro_3b', 'ch2intro_3c', 'ch2intro_3d', 'ch2intro_3e', 'ch2intro_3f', 'ch2intro_40', 'ch2intro_41', 'ch2intro_42', 'ch2intro_43', 'ch2intro_44', 'ch2intro_45', 'ch2intro_46', 'ch2intro_47', 'ch2intro_48', 'ch2intro_49', 'ch2intro_4a', 'ch2intro_4b', 'ch2intro_4c', 'ch2intro_4d', 'ch2intro_4e', 'ch2intro_4f', 'ch2intro_50', 'ch2intro_51', 'ch2intro_52', 'ch2intro_53', 'ch2intro_54', 'ch2intro_55', 'ch2intro_56', 'ch2intro_57', 'ch2intro_58', 'ch2intro_59', 'ch2intro_5a', 'ch2intro_5b', 'ch2intro_5c', 'ch2intro_5d', 'ch2intro_5e', 'ch2intro_5f', 'ch2intro_60', 'ch2intro_61', 'ch2intro_62', 'ch2intro_63', 'ch2intro_64', 'ch2intro_65', 'ch2intro_66', 'ch2intro_67'], 'ch4strings_dicts': ['ch4strings_dicts_1', 'ch4strings_dicts_2', 'ch4strings_dicts_3', 'ch4strings_dicts_4', 'ch4strings_dicts_5', 'ch4strings_dicts_6', 'ch4strings_dicts_7', 'ch4strings_dicts_8', 'ch4strings_dicts_9', 'ch4strings_dicts_a', 'ch4strings_dicts_b', 'ch4strings_dicts_c', 'ch4strings_dicts_d', 'ch4strings_dicts_e', 'ch4strings_dicts_f', 'ch4strings_dicts_10', 'ch4strings_dicts_11', 'ch4strings_dicts_12', 'ch4strings_dicts_13', 'ch4strings_dicts_14', 'ch4strings_dicts_15', 'ch4strings_dicts_16', 'ch4strings_dicts_17', 'ch4strings_dicts_18', 'ch4strings_dicts_19', 'ch4strings_dicts_1a', 'ch4strings_dicts_1b', 'ch4strings_dicts_1c', 'ch4strings_dicts_1d', 'ch4strings_dicts_1e', 'ch4strings_dicts_1f', 'ch4strings_dicts_20', 'ch4strings_dicts_21', 'ch4strings_dicts_22', 'ch4strings_dicts_23', 'ch4strings_dicts_24', 'ch4strings_dicts_25', 'ch4strings_dicts_26', 'ch4strings_dicts_27', 'ch4strings_dicts_28', 'ch4strings_dicts_29', 'ch4strings_dicts_2a', 'ch4strings_dicts_2b', 'ch4strings_dicts_2c', 'ch4strings_dicts_2d', 'ch4strings_dicts_2e', 'ch4strings_dicts_2f', 'ch4strings_dicts_30'], 'ch11Section_5': ['ch11Section_5_1', 'ch11Section_5_2', 'ch11Section_5_3', 'ch11Section_5_4', 'ch11Section_5_5', 'ch11Section_5_6', 'ch11Section_5_7', 'ch11Section_5_8', 'ch11Section_5_9', 'ch11Section_5_a', 'ch11Section_5_b', 'ch11Section_5_c', 'ch11Section_5_d', 'ch11Section_5_e', 'ch11Section_5_f', 'ch11Section_5_10', 'ch11Section_5_11', 'ch11Section_5_12', 'ch11Section_5_13', 'ch11Section_5_14', 'ch11Section_5_15', 'ch11Section_5_16', 'ch11Section_5_17', 'ch11Section_5_18', 'ch11Section_5_19', 'ch11Section_5_1a', 'ch11Section_5_1b', 'ch11Section_5_1c', 'ch11Section_5_1d', 'ch11Section_5_1e', 'ch11Section_5_1f', 'ch11Section_5_20', 'ch11Section_5_21', 'ch11Section_5_22', 'ch11Section_5_23', 'ch11Section_5_24', 'ch11Section_5_25', 'ch11Section_5_26', 'ch11Section_5_27', 'ch11Section_5_28', 'ch11Section_5_29', 'ch11Section_5_2a', 'ch11Section_5_2b', 'ch11Section_5_2c', 'ch11Section_5_2d', 'ch11Section_5_2e', 'ch11Section_5_2f', 'ch11Section_5_30', 'ch11Section_5_31', 'ch11Section_5_32', 'ch11Section_5_33', 'ch11Section_5_34', 'ch11Section_5_35', 'ch11Section_5_36', 'ch11Section_5_37', 'ch11Section_5_38', 'ch11Section_5_39', 'ch11Section_5_3a', 'ch11Section_5_3b', 'ch11Section_5_3c', 'ch11Section_5_3d', 'ch11Section_5_3e', 'ch11Section_5_3f', 'ch11Section_5_40', 'ch11Section_5_41', 'ch11Section_5_42', 'ch11Section_5_43', 'ch11Section_5_44', 'ch11Section_5_45', 'ch11Section_5_46', 'ch11Section_5_47', 'ch11Section_5_48', 'ch11Section_5_49', 'ch11Section_5_4a', 'ch11Section_5_4b', 'ch11Section_5_4c', 'ch11Section_5_4d', 'ch11Section_5_4e', 'ch11Section_5_4f', 'ch11Section_5_50', 'ch11Section_5_51', 'ch11Section_5_52', 'ch11Section_5_53', 'ch11Section_5_54', 'ch11Section_5_55', 'ch11Section_5_56', 'ch11Section_5_57', 'ch11Section_5_58', 'ch11Section_5_59', 'ch11Section_5_5a', 'ch11Section_5_5b', 'ch11Section_5_5c', 'ch11Section_5_5d', 'ch11Section_5_5e', 'ch11Section_5_5f', 'ch11Section_5_60', 'ch11Section_5_61', 'ch11Section_5_62', 'ch11Section_5_63', 'ch11Section_5_64', 'ch11Section_5_65', 'ch11Section_5_66', 'ch11Section_5_67', 'ch11Section_5_68', 'ch11Section_5_69', 'ch11Section_5_6a', 'ch11Section_5_6b', 'ch11Section_5_6c', 'ch11Section_5_6d', 'ch11Section_5_6e'], 'ch7interim_assessment': [], 'ch8ult_module_plan': ['ch8ult_module_plan_1', 'ch8ult_module_plan_2', 'ch8ult_module_plan_3', 'ch8ult_module_plan_4', 'ch8ult_module_plan_5', 'ch8ult_module_plan_6', 'ch8ult_module_plan_7', 'ch8ult_module_plan_8', 'ch8ult_module_plan_9', 'ch8ult_module_plan_a', 'ch8ult_module_plan_b', 'ch8ult_module_plan_c', 'ch8ult_module_plan_d', 'ch8ult_module_plan_e', 'ch8ult_module_plan_f', 'ch8ult_module_plan_10', 'ch8ult_module_plan_11', 'ch8ult_module_plan_12', 'ch8ult_module_plan_13', 'ch8ult_module_plan_14', 'ch8ult_module_plan_15', 'ch8ult_module_plan_16', 'ch8ult_module_plan_17', 'ch8ult_module_plan_18', 'ch8ult_module_plan_19', 'ch8ult_module_plan_1a', 'ch8ult_module_plan_1b', 'ch8ult_module_plan_1c', 'ch8ult_module_plan_1d', 'ch8ult_module_plan_1e', 'ch8ult_module_plan_1f', 'ch8ult_module_plan_20', 'ch8ult_module_plan_21', 'ch8ult_module_plan_22', 'ch8ult_module_plan_23', 'ch8ult_module_plan_24', 'ch8ult_module_plan_25', 'ch8ult_module_plan_26', 'ch8ult_module_plan_27', 'ch8ult_module_plan_28', 'ch8ult_module_plan_29', 'ch8ult_module_plan_2a', 'ch8ult_module_plan_2b', 'ch8ult_module_plan_2c', 'ch8ult_module_plan_2d', 'ch8ult_module_plan_2e', 'ch8ult_module_plan_2f', 'ch8ult_module_plan_30', 'ch8ult_module_plan_31', 'ch8ult_module_plan_32', 'ch8ult_module_plan_33', 'ch8ult_module_plan_34', 'ch8ult_module_plan_35', 'ch8ult_module_plan_36', 'ch8ult_module_plan_37', 'ch8ult_module_plan_38', 'ch8ult_module_plan_39', 'ch8ult_module_plan_3a', 'ch8ult_module_plan_3b', 'ch8ult_module_plan_3c', 'ch8ult_module_plan_3d', 'ch8ult_module_plan_3e', 'ch8ult_module_plan_3f', 'ch8ult_module_plan_40', 'ch8ult_module_plan_41', 'ch8ult_module_plan_42', 'ch8ult_module_plan_43', 'ch8ult_module_plan_44'], 'ch10session4': ['ch10session4_1', 'ch10session4_2', 'ch10session4_3', 'ch10session4_4', 'ch10session4_5', 'ch10session4_6', 'ch10session4_7', 'ch10session4_8', 'ch10session4_9', 'ch10session4_a', 'ch10session4_b', 'ch10session4_c', 'ch10session4_d', 'ch10session4_e', 'ch10session4_f', 'ch10session4_10', 'ch10session4_11', 'ch10session4_12', 'ch10session4_13', 'ch10session4_14', 'ch10session4_15', 'ch10session4_16', 'ch10session4_17', 'ch10session4_18', 'ch10session4_19', 'ch10session4_1a', 'ch10session4_1b', 'ch10session4_1c', 'ch10session4_1d', 'ch10session4_1e', 'ch10session4_1f', 'ch10session4_20', 'ch10session4_21', 'ch10session4_22', 'ch10session4_23', 'ch10session4_24', 'ch10session4_25', 'ch10session4_26', 'ch10session4_27', 'ch10session4_28', 'ch10session4_29', 'ch10session4_2a', 'ch10session4_2b', 'ch10session4_2c', 'ch10session4_2d', 'ch10session4_2e', 'ch10session4_2f', 'ch10session4_30', 'ch10session4_31', 'ch10session4_32', 'ch10session4_33', 'ch10session4_34', 'ch10session4_35', 'ch10session4_36', 'ch10session4_37', 'ch10session4_38', 'ch10session4_39', 'ch10session4_3a', 'ch10session4_3b', 'ch10session4_3c', 'ch10session4_3d', 'ch10session4_3e', 'ch10session4_3f', 'ch10session4_40', 'ch10session4_41', 'ch10session4_42', 'ch10session4_43', 'ch10session4_44', 'ch10session4_45', 'ch10session4_46', 'ch10session4_47', 'ch10session4_48', 'ch10session4_49', 'ch10session4_4a', 'ch10session4_4b', 'ch10session4_4c', 'ch10session4_4d', 'ch10session4_4e', 'ch10session4_4f', 'ch10session4_50', 'ch10session4_51', 'ch10session4_52', 'ch10session4_53', 'ch10session4_54', 'ch10session4_55', 'ch10session4_56', 'ch10session4_57', 'ch10session4_58', 'ch10session4_59', 'ch10session4_5a', 'ch10session4_5b', 'ch10session4_5c', 'ch10session4_5d', 'ch10session4_5e', 'ch10session4_5f', 'ch10session4_60', 'ch10session4_61', 'ch10session4_62', 'ch10session4_63', 'ch10session4_64', 'ch10session4_65', 'ch10session4_66', 'ch10session4_67', 'ch10session4_68', 'ch10session4_69', 'ch10session4_6a', 'ch10session4_6b', 'ch10session4_6c', 'ch10session4_6d', 'ch10session4_6e', 'ch10session4_6f', 'ch10session4_70', 'ch10session4_71', 'ch10session4_72', 'ch10session4_73', 'ch10session4_74', 'ch10session4_75', 'ch10session4_76', 'ch10session4_77', 'ch10session4_78', 'ch10session4_79', 'ch10session4_7a', 'ch10session4_7b', 'ch10session4_7c', 'ch10session4_7d', 'ch10session4_7e', 'ch10session4_7f', 'ch10session4_80', 'ch10session4_81', 'ch10session4_82', 'ch10session4_83', 'ch10session4_84', 'ch10session4_85', 'ch10session4_86', 'ch10session4_87', 'ch10session4_88', 'ch10session4_89', 'ch10session4_8a', 'ch10session4_8b', 'ch10session4_8c', 'ch10session4_8d', 'ch10session4_8e', 'ch10session4_8f', 'ch10session4_90', 'ch10session4_91', 'ch10session4_92', 'ch10session4_93', 'ch10session4_94', 'ch10session4_95', 'ch10session4_96', 'ch10session4_97', 'ch10session4_98', 'ch10session4_99', 'ch10session4_9a', 'ch10session4_9b', 'ch10session4_9c', 'ch10session4_9d', 'ch10session4_9e', 'ch10session4_9f', 'ch10session4_a0', 'ch10session4_a1', 'ch10session4_a2', 'ch10session4_a3', 'ch10session4_a4', 'ch10session4_a5', 'ch10session4_a6', 'ch10session4_a7', 'ch10session4_a8', 'ch10session4_a9', 'ch10session4_aa', 'ch10session4_ab', 'ch10session4_ac', 'ch10session4_ad', 'ch10session4_ae', 'ch10session4_af', 'ch10session4_b0', 'ch10session4_b1', 'ch10session4_b2', 'ch10session4_b3', 'ch10session4_b4', 'ch10session4_b5', 'ch10session4_b6']} \ No newline at end of file diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/scripts/rst2docbook.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/scripts/rst2docbook.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,126 @@ +#!/usr/bin/python +""" +Just a hack to convert rst to xml and then docbook . +May not containt all the required elements of a docbook . +Just done to make it run for the specific rst for our +sees documentation. +""" +import xml.etree.ElementTree as ET +import os +import re +import subprocess +import os +import pkg_resources +import glob, os, re, sys +names = glob.glob('ch*.xml') +""" +chapterno=0 + +def convert2xml(file): +# print folder,subfolder,file + global chapterno + name=file.split('/')[-1] + name=str(chapterno)+name.split('.')[0] +# full_file=os.path.join(folder,file) +# if file.endswith('.rst'): + print file + xml_file=name+'.xml' + command="rst2xml.py %s > %s" %(file , xml_file) + print command + a=subprocess.Popen(command , shell=True) + + + +def walk(repo): + global chapterno + mainfolder='/home/amit/sttp_latest/' + for readline in open('index.config','r').readlines(): + chapterno+=1 + filename=mainfolder+readline + convert2xml(filename) + + +""" +def convert2docbook(xml_string): +# xml_file_obj=open(file,'r') +# xml_string=xml_file_obj.read() +# xml_file_obj.close() + try: + xml_string=re.sub('','', xml_string) + xml_string=re.sub('',xml_string) + xml_string=re.sub('','',xml_string) + xml_string=re.sub('xml:space="preserve">','',xml_string) + xml_string=re.sub('' ,'',xml_string) + xml_string=re.sub('' ,'',xml_string) + xml_string=re.sub('' ,'',xml_string) + chapter= ET.Element("chapter") + article=ET.SubElement(chapter,"article") + articleinfo=ET.SubElement(article,"articleinfo") + tree = ET.fromstring(xml_string) + title= ET.SubElement(articleinfo,"title") + try: + title.text=tree.items()[1][1] + except: + pass + article.insert(1,tree) + xml_string=ET.tostring(chapter) + xml_string=re.sub('' ,'',xml_string) + xml_string=re.sub('' ,'',xml_string) + return xml_string +# docbook_file=file.split('.')[0]+'.docbook' +# f=open(docbook_file,'w') +# f.write(xml_string) +# xml_string=None + except : + pass + + + + +if __name__=='__main__': +# repo='/home/amit/sphinxvirt/review/' +# walk(repo) +# convert(1,2,3) + for name in names: + print name + xml_string=open(name,'r').read() + + xml_string=convert2docbook(xml_string) + docbook_file=name.split('.')[0]+'.docbook' + f=open(docbook_file,'w') + try: + f.write(xml_string) + except: + pass + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/README --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/README Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,5 @@ +This directory contains web-related files. Surprise! + +javascript - files used by the comment system, based on jQuery +hgbook - Django app that acts as the comment back end +styles.css - style file diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/genindex.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/genindex.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,61 @@ +#!/usr/bin/env python + +import glob, os, re + +chapter_re = re.compile(r'<(chapter|appendix|preface)\s+id="([^"]+)">') +filename_re = re.compile(r'<\?dbhtml filename="([^"]+)"\?>') +title_re = re.compile(r'(.*)') + +chapters = (sorted(glob.glob('../en/ch*.xml')) + + sorted(glob.glob('../en/app*.xml'))) + +fp = open('index-read.html.in', 'w') + +print >> fp, ''' +{% extends "boilerplate.html" %} +{% block bodycontent %} + +
    ''' + +ch = 0 +app = 0 +ab = 0 +for c in chapters: + filename = None + title = None + chapid = None + chaptype = None + for line in open(c): + m = chapter_re.search(line) + if m: + chaptype, chapid = m.groups() + m = filename_re.search(line) + if m: + filename = m.group(1) + m = title_re.search(line) + if m: + title = m.group(1) + if filename and title and chapid: + if chaptype == 'appendix': + num = chr(ord('A') + app) + app += 1 + else: + num = ch + ch += 1 + ab += 1 + date = os.popen('hg log -l1 --template "{date|isodate}" ' + c).read().split(None, 1)[0] + args = { + 'ab': "ab"[ab % 2], + 'date': date, + 'chapid': chapid, + 'num': num, + 'filename': filename, + 'title': title, + } + print >> fp, '
  • %(date)s%(num)s. %(title)s
  • ' % args + break + +print >> fp, '''
+{% endblock %}''' + +fp.close() diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook.conf --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook.conf Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,111 @@ +# -*- apache -*- + + +# ServerName hgbook.red-bean.com +# ServerAdmin bos@serpentine.com + ServerAdmin webmaster@localhost + ErrorLog /var/log/apache2/error.log +# ErrorLog logs/hgbook-error_log + # Debian: + # CustomLog logs/hgbook-access_log full + # Fedora: +# CustomLog logs/hgbook-access_log combined + CustomLog /var/log/apache2/access.log combined +# Options +MultiViews +# DirectoryIndex index.html.var index.html + DocumentRoot "/home/amit/hgbook-alqua/web/html" + + +# Redirect permanent /hgbook.html /index.html +# Redirect permanent /hgbookch1.html /read/preface.html +# Redirect permanent /hgbookch2.html /read/a-tour-of-mercurial-the-basics.html +# Redirect permanent /hgbookch3.html /read/a-tour-of-mercurial-merging-work.html +# Redirect permanent /hgbookch4.html /read/behind-the-scenes.html +# Redirect permanent /hgbookch5.html /read/mercurial-in-daily-use.html +# Redirect permanent /hgbookch6.html /read/file-names-and-pattern-matching.html +# Redirect permanent /hgbookch6.html /read/managing-releases-and-branchy-development.html +# Redirect permanent /hgbookch7.html /read/finding-and-fixing-mistakes.html +# Redirect permanent /hgbookch8.html /read/handling-repository-events-with-hooks.html +# Redirect permanent /hgbookch9.html /read/customizing-the-output-of-mercurial.html +# Redirect permanent /hgbookch10.html /read/managing-change-with-mercurial-queues.html +# Redirect permanent /hgbookch11.html /read/advanced-uses-of-mercurial-queues.html +# Redirect permanent /hgbookch12.html /read/adding-functionality-with-extensions.html +# Redirect permanent /hgbookap1.html /read/command-reference.html +# Redirect permanent /hgbookap2.html /read/mercurial-queues-reference.html +# Redirect permanent /hgbookap3.html /read/installing-mercurial-from-source.html +# Redirect permanent /hgbookap4.html /read/open-publication-license.html +# Redirect permanent /hgbookli1.html /read/index.html +# Redirect permanent /hgbookli2.html /read/index.html +# Redirect permanent /hgbookli3.html /read/index.html +# Redirect permanent /hgbookli4.html /read/index.html + + WSGIScriptAlias /review /home/amit/hgbook-alqua/web/hgbook/run.wsgi + + + + # Actively redirect requests via a ServerAlias to the canonical hostname. + RewriteEngine On +# RewriteCond %{HTTP_HOST} !=hgbook.red-bean.com +# RewriteRule ^(.*) http://hgbook.red-bean.com$1 [R] + +# +# SetHandler python-program + # hg clone http://bitbucket.org/mirror/django-trunk/ +# PythonPath "['/home/amit/hgbook-alqua/web/hgbook'] + sys.path" +# PythonHandler django.core.handlers.modpython +# PythonAutoReload Off +# SetEnv DJANGO_SETTINGS_MODULE hgbook.settings +# PythonDebug Off +# + + + + + SetHandler None + DirectoryIndex index.html + + + + SetHandler None + + + + SetHandler None + + + + SetHandler None + + + + SetHandler None + + + + SetHandler None + + +# Alias /media /home/bos/hg/django-trunk/django/contrib/admin/media + + + Options Indexes FollowSymlinks + AllowOverride None + Order allow,deny + Allow from all + + + + AllowOverride AuthConfig + + + + Options None + + + +# +# Options None +# AllowOverride None +# Order allow,deny +# Allow from all +# diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook.conf~ --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook.conf~ Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,111 @@ +# -*- apache -*- + + +# ServerName hgbook.red-bean.com +# ServerAdmin bos@serpentine.com + ServerAdmin webmaster@localhost + ErrorLog /var/log/apache2/error.log +# ErrorLog logs/hgbook-error_log + # Debian: + # CustomLog logs/hgbook-access_log full + # Fedora: +# CustomLog logs/hgbook-access_log combined + CustomLog /var/log/apache2/access.log combined +# Options +MultiViews +# DirectoryIndex index.html.var index.html +# DocumentRoot "/home/amit/hgbook-alqua/web/html" + + +# Redirect permanent /hgbook.html /index.html +# Redirect permanent /hgbookch1.html /read/preface.html +# Redirect permanent /hgbookch2.html /read/a-tour-of-mercurial-the-basics.html +# Redirect permanent /hgbookch3.html /read/a-tour-of-mercurial-merging-work.html +# Redirect permanent /hgbookch4.html /read/behind-the-scenes.html +# Redirect permanent /hgbookch5.html /read/mercurial-in-daily-use.html +# Redirect permanent /hgbookch6.html /read/file-names-and-pattern-matching.html +# Redirect permanent /hgbookch6.html /read/managing-releases-and-branchy-development.html +# Redirect permanent /hgbookch7.html /read/finding-and-fixing-mistakes.html +# Redirect permanent /hgbookch8.html /read/handling-repository-events-with-hooks.html +# Redirect permanent /hgbookch9.html /read/customizing-the-output-of-mercurial.html +# Redirect permanent /hgbookch10.html /read/managing-change-with-mercurial-queues.html +# Redirect permanent /hgbookch11.html /read/advanced-uses-of-mercurial-queues.html +# Redirect permanent /hgbookch12.html /read/adding-functionality-with-extensions.html +# Redirect permanent /hgbookap1.html /read/command-reference.html +# Redirect permanent /hgbookap2.html /read/mercurial-queues-reference.html +# Redirect permanent /hgbookap3.html /read/installing-mercurial-from-source.html +# Redirect permanent /hgbookap4.html /read/open-publication-license.html +# Redirect permanent /hgbookli1.html /read/index.html +# Redirect permanent /hgbookli2.html /read/index.html +# Redirect permanent /hgbookli3.html /read/index.html +# Redirect permanent /hgbookli4.html /read/index.html + + WSGIScriptAlias /review /home/amit/hgbook-alqua/web/hgbook/run.wsgi + + + + # Actively redirect requests via a ServerAlias to the canonical hostname. + RewriteEngine On +# RewriteCond %{HTTP_HOST} !=hgbook.red-bean.com +# RewriteRule ^(.*) http://hgbook.red-bean.com$1 [R] + +# +# SetHandler python-program + # hg clone http://bitbucket.org/mirror/django-trunk/ +# PythonPath "['/home/amit/hgbook-alqua/web/hgbook'] + sys.path" +# PythonHandler django.core.handlers.modpython +# PythonAutoReload Off +# SetEnv DJANGO_SETTINGS_MODULE hgbook.settings +# PythonDebug Off +# + + + + + SetHandler None + DirectoryIndex index.html + + + + SetHandler None + + + + SetHandler None + + + + SetHandler None + + + + SetHandler None + + + + SetHandler None + + +# Alias /media /home/bos/hg/django-trunk/django/contrib/admin/media + + + Options Indexes FollowSymlinks + AllowOverride None + Order allow,deny + Allow from all + + + + AllowOverride AuthConfig + + + + Options None + + + +# +# Options None +# AllowOverride None +# Order allow,deny +# Allow from all +# diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/.database.sqlite3 Binary file SEESenv/web/hgbook/.database.sqlite3 has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/__init__.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/__init__.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1 @@ +import admin diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/__init__.pyc Binary file SEESenv/web/hgbook/__init__.pyc has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/admin.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/admin.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,24 @@ +from django.contrib import admin +from hgbook.comments.models import Comment, Element + +class CommentAdmin(admin.ModelAdmin): + list_display = ['element', 'submitter_name', 'comment', 'reviewed', + 'hidden', 'date'] + search_fields = ['comment'] + date_hierarchy = 'date' + list_filter = ['date', 'submitter_name'] + search_fields = ['title', 'submitter_name', 'submitter_url'] + fieldsets = ( + (None, {'fields': ('submitter_name', 'element', 'comment')}), + ('Review and presentation state', {'fields': ('reviewed', 'hidden')}), + ('Other info', {'fields': ('submitter_url', 'ip')}), + ) + # XXX: adding 'date' to the 'Other info' fieldset results in a + # ImproperlyConfigured error. :S + +class ElementAdmin(admin.ModelAdmin): + search_fields = ['id', 'chapter'] + list_filter = ['chapter', 'title'] +# search_fields = ['id' +admin.site.register(Comment, CommentAdmin) +admin.site.register(Element, ElementAdmin) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/admin.pyc Binary file SEESenv/web/hgbook/admin.pyc has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/admin.py~ --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/admin.py~ Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,24 @@ +from django.contrib import admin +from hgbook.comments.models import Comment, Element + +class CommentAdmin(admin.ModelAdmin): + list_display = ['element', 'submitter_name', 'comment', 'reviewed', + 'hidden', 'date'] + search_fields = ['comment'] + date_hierarchy = 'date' + list_filter = ['date', 'submitter_name'] + search_fields = ['title', 'submitter_name', 'submitter_url'] + fieldsets = ( + (None, {'fields': ('submitter_name', 'element', 'comment')}), + ('Review and presentation state', {'fields': ('reviewed', 'hidden')}), + ('Other info', {'fields': ('submitter_url', 'ip')}), + ) + # XXX: adding 'date' to the 'Other info' fieldset results in a + # ImproperlyConfigured error. :S + +class ElementAdmin(admin.ModelAdmin): + search_fields = ['id'] +# list_filter = ['chapter', 'title'] + +admin.site.register(Comment, CommentAdmin) +admin.site.register(Element, ElementAdmin) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/__init__.py diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/__init__.pyc Binary file SEESenv/web/hgbook/comments/__init__.pyc has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/feeds.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/comments/feeds.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,35 @@ +from django.core.exceptions import ObjectDoesNotExist +from django.utils.feedgenerator import Atom1Feed +from django.contrib.syndication.feeds import Feed +from hgbook.comments.models import Comment, Element + +class MyAtomFeed(Atom1Feed): + title_type = u'html' + +class Comments(Feed): + feed_type = MyAtomFeed + title = 'Mercurial - The Definitive Guide: recent comments' + subtitle = ('Recent comments on the text of “Mercurial: The ' + 'Definitive Guide”, from our readers') + link = '/feeds/comments/' + author_name = 'Our readers' + + def feedfilter(self, queryset): + return queryset.order_by('-date')[:20] + + def items(self): + return self.feedfilter(Comment.objects) + + def item_author_name(self, obj): + return obj.submitter_name + + def item_pubdate(self, obj): + return obj.date + + def get_object(self, bits): + if len(bits) == 0: + return self.items() + elif len(bits) > 1: + raise ObjectDoesNotExist + return self.feedfilter(Comment.objects.filter(element__chapter=bits[0], + hidden=False)) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/feeds.pyc Binary file SEESenv/web/hgbook/comments/feeds.pyc has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/models.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/comments/models.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,86 @@ +from django.db import models +import hashlib +from django.conf import settings +from django.contrib.flatpages.models import FlatPage +import os +from BeautifulSoup import BeautifulSoup +import django +from django.utils.html import strip_tags +import djapian +from djapian.indexer import CompositeIndexer ,Indexer +mutable = True +#standard_url='localhost/review/html/' + + + +class Element(models.Model): + id = models.CharField('ID attribute', max_length=64, editable=False, + primary_key=True) + + chapter = models.CharField('Chapter ID', max_length=100, editable=False, + db_index=True) + title = models.CharField('Section title', max_length=256, editable=False) + + def __unicode__(self): + return self.id + +class Comment(models.Model): + element = models.ForeignKey(Element, + help_text='ID of paragraph that was commented on') + comment = models.TextField(editable=mutable, + help_text='Text of submitted comment (please do not modify)') + submitter_name = models.CharField('Submitter', max_length=64, + help_text='Self-reported name of submitter (may be bogus)') + submitter_url = models.URLField('URL', blank=True, editable=mutable, + help_text='Self-reported URL of submitter (may be empty or bogus)') + ip = models.IPAddressField('IP address', editable=mutable, + help_text='IP address from which comment was submitted') + date = models.DateTimeField('date submitted', auto_now=True, + auto_now_add=True) + reviewed = models.BooleanField(default=False, db_index=True, + help_text='Has this comment been reviewed by an author?') + hidden = models.BooleanField(default=False, db_index=True, + help_text='Has this comment been hidden from public display?') + + def __unicode__(self): + return self.comment[:32] + +# def get_absolute_url(self): +# s = hashlib.new() +# s.update(repr(self.comment)) +# s.update(repr(self.submitter_name)) +# s.update(str(self.date)) +# print '/read/%s.html#%s?comment=%s&uuid=%s' % ( +# self.element.chapter, self.element.id, self.id, s.hexdigest()[:20] +# ) +# return '/read/%s.html#%s?comment=%s&uuid=%s' % ( +# self.element.chapter, self.element.id, self.id, s.hexdigest()[:20] +# ) + + +""" +for directory,lists,files in os.walk('../html'): + if directory=='../html': + for file in files: + if file.endswith('.html'): + filename=os.path.join('../html/'+file) + f=open(filename,'r') + soup = BeautifulSoup(''.join(f.read())) + titletag=soup.html.head.title + body=strip_tags(soup.html.body) + c=FlatPage(url=standard_url+file,title=titletag.string,content=body) + c.save() + +class FlatPageIndexer( djapian.Indexer ): + fields = [ 'title', 'content' ] + + +djapian.space.add_index(FlatPage, FlatPageIndexer, attach_as="indexer") +FlatPage.indexer.update() +result = FlatPage.indexer.search('lists') +for row in result : + print row.instance.url +print result.count() + +""" + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/models.pyc Binary file SEESenv/web/hgbook/comments/models.pyc has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/models.py~ --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/comments/models.py~ Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,86 @@ +from django.db import models +import hashlib +from django.conf import settings +from django.contrib.flatpages.models import FlatPage +import os +from BeautifulSoup import BeautifulSoup +import django +from django.utils.html import strip_tags +import djapian +from djapian.indexer import CompositeIndexer ,Indexer +mutable = True +standard_url='localhost/review/html/' + + + +class Element(models.Model): + id = models.CharField('ID attribute', max_length=64, editable=False, + primary_key=True) + + chapter = models.CharField('Chapter ID', max_length=100, editable=False, + db_index=True) + title = models.CharField('Section title', max_length=256, editable=False) + + def __unicode__(self): + return self.id + +class Comment(models.Model): + element = models.ForeignKey(Element, + help_text='ID of paragraph that was commented on') + comment = models.TextField(editable=mutable, + help_text='Text of submitted comment (please do not modify)') + submitter_name = models.CharField('Submitter', max_length=64, + help_text='Self-reported name of submitter (may be bogus)') + submitter_url = models.URLField('URL', blank=True, editable=mutable, + help_text='Self-reported URL of submitter (may be empty or bogus)') + ip = models.IPAddressField('IP address', editable=mutable, + help_text='IP address from which comment was submitted') + date = models.DateTimeField('date submitted', auto_now=True, + auto_now_add=True) + reviewed = models.BooleanField(default=False, db_index=True, + help_text='Has this comment been reviewed by an author?') + hidden = models.BooleanField(default=False, db_index=True, + help_text='Has this comment been hidden from public display?') + + def __unicode__(self): + return self.comment[:32] + +# def get_absolute_url(self): +# s = hashlib.new() +# s.update(repr(self.comment)) +# s.update(repr(self.submitter_name)) +# s.update(str(self.date)) +# print '/read/%s.html#%s?comment=%s&uuid=%s' % ( +# self.element.chapter, self.element.id, self.id, s.hexdigest()[:20] +# ) +# return '/read/%s.html#%s?comment=%s&uuid=%s' % ( +# self.element.chapter, self.element.id, self.id, s.hexdigest()[:20] +# ) + + +""" +for directory,lists,files in os.walk('../html'): + if directory=='../html': + for file in files: + if file.endswith('.html'): + filename=os.path.join('../html/'+file) + f=open(filename,'r') + soup = BeautifulSoup(''.join(f.read())) + titletag=soup.html.head.title + body=strip_tags(soup.html.body) + c=FlatPage(url=standard_url+file,title=titletag.string,content=body) + c.save() + +class FlatPageIndexer( djapian.Indexer ): + fields = [ 'title', 'content' ] + + +djapian.space.add_index(FlatPage, FlatPageIndexer, attach_as="indexer") +FlatPage.indexer.update() +result = FlatPage.indexer.search('lists') +for row in result : + print row.instance.url +print result.count() + +""" + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/p_list.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/comments/p_list.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1 @@ +p_list= {'ch6oop': ['ch6oop_1', 'ch6oop_2', 'ch6oop_3', 'ch6oop_4', 'ch6oop_5', 'ch6oop_6', 'ch6oop_7', 'ch6oop_8', 'ch6oop_9', 'ch6oop_a', 'ch6oop_b', 'ch6oop_c', 'ch6oop_d', 'ch6oop_e', 'ch6oop_f', 'ch6oop_10', 'ch6oop_11', 'ch6oop_12', 'ch6oop_13', 'ch6oop_14', 'ch6oop_15', 'ch6oop_16', 'ch6oop_17', 'ch6oop_18', 'ch6oop_19', 'ch6oop_1a', 'ch6oop_1b', 'ch6oop_1c', 'ch6oop_1d', 'ch6oop_1e', 'ch6oop_1f', 'ch6oop_20', 'ch6oop_21', 'ch6oop_22', 'ch6oop_23', 'ch6oop_24', 'ch6oop_25', 'ch6oop_26', 'ch6oop_27', 'ch6oop_28', 'ch6oop_29', 'ch6oop_2a', 'ch6oop_2b', 'ch6oop_2c', 'ch6oop_2d', 'ch6oop_2e', 'ch6oop_2f', 'ch6oop_30', 'ch6oop_31', 'ch6oop_32', 'ch6oop_33', 'ch6oop_34', 'ch6oop_35', 'ch6oop_36', 'ch6oop_37', 'ch6oop_38', 'ch6oop_39', 'ch6oop_3a', 'ch6oop_3b', 'ch6oop_3c', 'ch6oop_3d', 'ch6oop_3e', 'ch6oop_3f', 'ch6oop_40', 'ch6oop_41', 'ch6oop_42', 'ch6oop_43', 'ch6oop_44', 'ch6oop_45', 'ch6oop_46', 'ch6oop_47', 'ch6oop_48', 'ch6oop_49', 'ch6oop_4a', 'ch6oop_4b', 'ch6oop_4c', 'ch6oop_4d', 'ch6oop_4e', 'ch6oop_4f', 'ch6oop_50', 'ch6oop_51', 'ch6oop_52', 'ch6oop_53', 'ch6oop_54', 'ch6oop_55', 'ch6oop_56', 'ch6oop_57', 'ch6oop_58', 'ch6oop_59', 'ch6oop_5a', 'ch6oop_5b', 'ch6oop_5c', 'ch6oop_5d', 'ch6oop_5e', 'ch6oop_5f', 'ch6oop_60', 'ch6oop_61', 'ch6oop_62', 'ch6oop_63', 'ch6oop_64', 'ch6oop_65', 'ch6oop_66', 'ch6oop_67', 'ch6oop_68', 'ch6oop_69', 'ch6oop_6a', 'ch6oop_6b', 'ch6oop_6c', 'ch6oop_6d', 'ch6oop_6e', 'ch6oop_6f', 'ch6oop_70', 'ch6oop_71', 'ch6oop_72', 'ch6oop_73', 'ch6oop_74', 'ch6oop_75', 'ch6oop_76', 'ch6oop_77', 'ch6oop_78', 'ch6oop_79', 'ch6oop_7a', 'ch6oop_7b', 'ch6oop_7c', 'ch6oop_7d', 'ch6oop_7e', 'ch6oop_7f', 'ch6oop_80', 'ch6oop_81', 'ch6oop_82', 'ch6oop_83', 'ch6oop_84', 'ch6oop_85', 'ch6oop_86', 'ch6oop_87', 'ch6oop_88', 'ch6oop_89', 'ch6oop_8a', 'ch6oop_8b', 'ch6oop_8c', 'ch6oop_8d', 'ch6oop_8e', 'ch6oop_8f', 'ch6oop_90', 'ch6oop_91', 'ch6oop_92', 'ch6oop_93', 'ch6oop_94', 'ch6oop_95', 'ch6oop_96', 'ch6oop_97', 'ch6oop_98', 'ch6oop_99', 'ch6oop_9a', 'ch6oop_9b', 'ch6oop_9c', 'ch6oop_9d', 'ch6oop_9e', 'ch6oop_9f', 'ch6oop_a0', 'ch6oop_a1', 'ch6oop_a2', 'ch6oop_a3', 'ch6oop_a4', 'ch6oop_a5', 'ch6oop_a6', 'ch6oop_a7', 'ch6oop_a8', 'ch6oop_a9', 'ch6oop_aa', 'ch6oop_ab', 'ch6oop_ac', 'ch6oop_ad', 'ch6oop_ae', 'ch6oop_af', 'ch6oop_b0', 'ch6oop_b1', 'ch6oop_b2', 'ch6oop_b3', 'ch6oop_b4', 'ch6oop_b5', 'ch6oop_b6', 'ch6oop_b7', 'ch6oop_b8', 'ch6oop_b9', 'ch6oop_ba', 'ch6oop_bb', 'ch6oop_bc', 'ch6oop_bd', 'ch6oop_be', 'ch6oop_bf', 'ch6oop_c0', 'ch6oop_c1', 'ch6oop_c2', 'ch6oop_c3', 'ch6oop_c4', 'ch6oop_c5', 'ch6oop_c6', 'ch6oop_c7', 'ch6oop_c8', 'ch6oop_c9', 'ch6oop_ca', 'ch6oop_cb', 'ch6oop_cc', 'ch6oop_cd', 'ch6oop_ce', 'ch6oop_cf', 'ch6oop_d0', 'ch6oop_d1', 'ch6oop_d2', 'ch6oop_d3', 'ch6oop_d4', 'ch6oop_d5', 'ch6oop_d6', 'ch6oop_d7', 'ch6oop_d8', 'ch6oop_d9', 'ch6oop_da', 'ch6oop_db', 'ch6oop_dc', 'ch6oop_dd', 'ch6oop_de', 'ch6oop_df', 'ch6oop_e0', 'ch6oop_e1', 'ch6oop_e2', 'ch6oop_e3', 'ch6oop_e4', 'ch6oop_e5', 'ch6oop_e6', 'ch6oop_e7', 'ch6oop_e8', 'ch6oop_e9', 'ch6oop_ea', 'ch6oop_eb', 'ch6oop_ec', 'ch6oop_ed', 'ch6oop_ee', 'ch6oop_ef', 'ch6oop_f0', 'ch6oop_f1', 'ch6oop_f2', 'ch6oop_f3', 'ch6oop_f4', 'ch6oop_f5', 'ch6oop_f6', 'ch6oop_f7', 'ch6oop_f8', 'ch6oop_f9', 'ch6oop_fa', 'ch6oop_fb', 'ch6oop_fc', 'ch6oop_fd', 'ch6oop_fe', 'ch6oop_ff', 'ch6oop_100', 'ch6oop_101', 'ch6oop_102', 'ch6oop_103', 'ch6oop_104', 'ch6oop_105', 'ch6oop_106', 'ch6oop_107', 'ch6oop_108', 'ch6oop_109', 'ch6oop_10a', 'ch6oop_10b', 'ch6oop_10c', 'ch6oop_10d', 'ch6oop_10e', 'ch6oop_10f', 'ch6oop_110', 'ch6oop_111', 'ch6oop_112', 'ch6oop_113', 'ch6oop_114', 'ch6oop_115', 'ch6oop_116', 'ch6oop_117', 'ch6oop_118', 'ch6oop_119', 'ch6oop_11a', 'ch6oop_11b', 'ch6oop_11c', 'ch6oop_11d', 'ch6oop_11e', 'ch6oop_11f', 'ch6oop_120', 'ch6oop_121', 'ch6oop_122', 'ch6oop_123', 'ch6oop_124', 'ch6oop_125', 'ch6oop_126', 'ch6oop_127', 'ch6oop_128', 'ch6oop_129', 'ch6oop_12a', 'ch6oop_12b', 'ch6oop_12c', 'ch6oop_12d', 'ch6oop_12e', 'ch6oop_12f', 'ch6oop_130', 'ch6oop_131', 'ch6oop_132', 'ch6oop_133', 'ch6oop_134', 'ch6oop_135', 'ch6oop_136', 'ch6oop_137', 'ch6oop_138', 'ch6oop_139', 'ch6oop_13a', 'ch6oop_13b', 'ch6oop_13c', 'ch6oop_13d', 'ch6oop_13e', 'ch6oop_13f', 'ch6oop_140', 'ch6oop_141', 'ch6oop_142', 'ch6oop_143', 'ch6oop_144', 'ch6oop_145', 'ch6oop_146', 'ch6oop_147', 'ch6oop_148', 'ch6oop_149', 'ch6oop_14a', 'ch6oop_14b', 'ch6oop_14c', 'ch6oop_14d', 'ch6oop_14e', 'ch6oop_14f', 'ch6oop_150', 'ch6oop_151', 'ch6oop_152', 'ch6oop_153', 'ch6oop_154', 'ch6oop_155', 'ch6oop_156', 'ch6oop_157', 'ch6oop_158', 'ch6oop_159', 'ch6oop_15a', 'ch6oop_15b', 'ch6oop_15c', 'ch6oop_15d', 'ch6oop_15e', 'ch6oop_15f', 'ch6oop_160', 'ch6oop_161', 'ch6oop_162', 'ch6oop_163', 'ch6oop_164', 'ch6oop_165', 'ch6oop_166', 'ch6oop_167', 'ch6oop_168', 'ch6oop_169', 'ch6oop_16a', 'ch6oop_16b', 'ch6oop_16c', 'ch6oop_16d', 'ch6oop_16e', 'ch6oop_16f', 'ch6oop_170', 'ch6oop_171', 'ch6oop_172', 'ch6oop_173', 'ch6oop_174', 'ch6oop_175', 'ch6oop_176', 'ch6oop_177', 'ch6oop_178', 'ch6oop_179', 'ch6oop_17a', 'ch6oop_17b', 'ch6oop_17c', 'ch6oop_17d', 'ch6oop_17e', 'ch6oop_17f', 'ch6oop_180', 'ch6oop_181', 'ch6oop_182', 'ch6oop_183', 'ch6oop_184', 'ch6oop_185', 'ch6oop_186', 'ch6oop_187', 'ch6oop_188', 'ch6oop_189', 'ch6oop_18a', 'ch6oop_18b', 'ch6oop_18c', 'ch6oop_18d', 'ch6oop_18e', 'ch6oop_18f', 'ch6oop_190', 'ch6oop_191', 'ch6oop_192', 'ch6oop_193', 'ch6oop_194', 'ch6oop_195', 'ch6oop_196', 'ch6oop_197', 'ch6oop_198', 'ch6oop_199', 'ch6oop_19a', 'ch6oop_19b', 'ch6oop_19c', 'ch6oop_19d', 'ch6oop_19e', 'ch6oop_19f', 'ch6oop_1a0', 'ch6oop_1a1', 'ch6oop_1a2', 'ch6oop_1a3', 'ch6oop_1a4', 'ch6oop_1a5', 'ch6oop_1a6', 'ch6oop_1a7', 'ch6oop_1a8', 'ch6oop_1a9', 'ch6oop_1aa', 'ch6oop_1ab', 'ch6oop_1ac', 'ch6oop_1ad', 'ch6oop_1ae', 'ch6oop_1af', 'ch6oop_1b0', 'ch6oop_1b1', 'ch6oop_1b2', 'ch6oop_1b3', 'ch6oop_1b4', 'ch6oop_1b5', 'ch6oop_1b6', 'ch6oop_1b7', 'ch6oop_1b8', 'ch6oop_1b9', 'ch6oop_1ba', 'ch6oop_1bb', 'ch6oop_1bc', 'ch6oop_1bd', 'ch6oop_1be', 'ch6oop_1bf', 'ch6oop_1c0', 'ch6oop_1c1', 'ch6oop_1c2', 'ch6oop_1c3', 'ch6oop_1c4', 'ch6oop_1c5', 'ch6oop_1c6', 'ch6oop_1c7', 'ch6oop_1c8', 'ch6oop_1c9', 'ch6oop_1ca', 'ch6oop_1cb', 'ch6oop_1cc', 'ch6oop_1cd', 'ch6oop_1ce', 'ch6oop_1cf', 'ch6oop_1d0', 'ch6oop_1d1', 'ch6oop_1d2', 'ch6oop_1d3', 'ch6oop_1d4', 'ch6oop_1d5', 'ch6oop_1d6', 'ch6oop_1d7', 'ch6oop_1d8', 'ch6oop_1d9', 'ch6oop_1da', 'ch6oop_1db', 'ch6oop_1dc', 'ch6oop_1dd', 'ch6oop_1de', 'ch6oop_1df', 'ch6oop_1e0', 'ch6oop_1e1'], 'ch9Using_Linux_Tools': ['ch9Using_Linux_Tools_1', 'ch9Using_Linux_Tools_2', 'ch9Using_Linux_Tools_3', 'ch9Using_Linux_Tools_4', 'ch9Using_Linux_Tools_5', 'ch9Using_Linux_Tools_6', 'ch9Using_Linux_Tools_7', 'ch9Using_Linux_Tools_8', 'ch9Using_Linux_Tools_9', 'ch9Using_Linux_Tools_a', 'ch9Using_Linux_Tools_b', 'ch9Using_Linux_Tools_c', 'ch9Using_Linux_Tools_d', 'ch9Using_Linux_Tools_e', 'ch9Using_Linux_Tools_f', 'ch9Using_Linux_Tools_10', 'ch9Using_Linux_Tools_11', 'ch9Using_Linux_Tools_12', 'ch9Using_Linux_Tools_13', 'ch9Using_Linux_Tools_14', 'ch9Using_Linux_Tools_15', 'ch9Using_Linux_Tools_16', 'ch9Using_Linux_Tools_17', 'ch9Using_Linux_Tools_18', 'ch9Using_Linux_Tools_19', 'ch9Using_Linux_Tools_1a', 'ch9Using_Linux_Tools_1b', 'ch9Using_Linux_Tools_1c', 'ch9Using_Linux_Tools_1d', 'ch9Using_Linux_Tools_1e', 'ch9Using_Linux_Tools_1f', 'ch9Using_Linux_Tools_20', 'ch9Using_Linux_Tools_21', 'ch9Using_Linux_Tools_22', 'ch9Using_Linux_Tools_23', 'ch9Using_Linux_Tools_24', 'ch9Using_Linux_Tools_25', 'ch9Using_Linux_Tools_26', 'ch9Using_Linux_Tools_27', 'ch9Using_Linux_Tools_28', 'ch9Using_Linux_Tools_29', 'ch9Using_Linux_Tools_2a', 'ch9Using_Linux_Tools_2b', 'ch9Using_Linux_Tools_2c', 'ch9Using_Linux_Tools_2d', 'ch9Using_Linux_Tools_2e', 'ch9Using_Linux_Tools_2f', 'ch9Using_Linux_Tools_30', 'ch9Using_Linux_Tools_31', 'ch9Using_Linux_Tools_32', 'ch9Using_Linux_Tools_33', 'ch9Using_Linux_Tools_34', 'ch9Using_Linux_Tools_35', 'ch9Using_Linux_Tools_36', 'ch9Using_Linux_Tools_37', 'ch9Using_Linux_Tools_38', 'ch9Using_Linux_Tools_39', 'ch9Using_Linux_Tools_3a', 'ch9Using_Linux_Tools_3b', 'ch9Using_Linux_Tools_3c', 'ch9Using_Linux_Tools_3d', 'ch9Using_Linux_Tools_3e', 'ch9Using_Linux_Tools_3f', 'ch9Using_Linux_Tools_40', 'ch9Using_Linux_Tools_41', 'ch9Using_Linux_Tools_42', 'ch9Using_Linux_Tools_43', 'ch9Using_Linux_Tools_44', 'ch9Using_Linux_Tools_45', 'ch9Using_Linux_Tools_46', 'ch9Using_Linux_Tools_47', 'ch9Using_Linux_Tools_48', 'ch9Using_Linux_Tools_49', 'ch9Using_Linux_Tools_4a', 'ch9Using_Linux_Tools_4b', 'ch9Using_Linux_Tools_4c', 'ch9Using_Linux_Tools_4d', 'ch9Using_Linux_Tools_4e', 'ch9Using_Linux_Tools_4f', 'ch9Using_Linux_Tools_50', 'ch9Using_Linux_Tools_51', 'ch9Using_Linux_Tools_52', 'ch9Using_Linux_Tools_53', 'ch9Using_Linux_Tools_54', 'ch9Using_Linux_Tools_55', 'ch9Using_Linux_Tools_56', 'ch9Using_Linux_Tools_57', 'ch9Using_Linux_Tools_58', 'ch9Using_Linux_Tools_59', 'ch9Using_Linux_Tools_5a', 'ch9Using_Linux_Tools_5b', 'ch9Using_Linux_Tools_5c', 'ch9Using_Linux_Tools_5d', 'ch9Using_Linux_Tools_5e', 'ch9Using_Linux_Tools_5f', 'ch9Using_Linux_Tools_60', 'ch9Using_Linux_Tools_61', 'ch9Using_Linux_Tools_62', 'ch9Using_Linux_Tools_63', 'ch9Using_Linux_Tools_64', 'ch9Using_Linux_Tools_65', 'ch9Using_Linux_Tools_66', 'ch9Using_Linux_Tools_67', 'ch9Using_Linux_Tools_68', 'ch9Using_Linux_Tools_69', 'ch9Using_Linux_Tools_6a', 'ch9Using_Linux_Tools_6b', 'ch9Using_Linux_Tools_6c', 'ch9Using_Linux_Tools_6d', 'ch9Using_Linux_Tools_6e', 'ch9Using_Linux_Tools_6f', 'ch9Using_Linux_Tools_70', 'ch9Using_Linux_Tools_71', 'ch9Using_Linux_Tools_72', 'ch9Using_Linux_Tools_73', 'ch9Using_Linux_Tools_74', 'ch9Using_Linux_Tools_75', 'ch9Using_Linux_Tools_76', 'ch9Using_Linux_Tools_77', 'ch9Using_Linux_Tools_78', 'ch9Using_Linux_Tools_79', 'ch9Using_Linux_Tools_7a', 'ch9Using_Linux_Tools_7b', 'ch9Using_Linux_Tools_7c', 'ch9Using_Linux_Tools_7d', 'ch9Using_Linux_Tools_7e', 'ch9Using_Linux_Tools_7f', 'ch9Using_Linux_Tools_80', 'ch9Using_Linux_Tools_81', 'ch9Using_Linux_Tools_82', 'ch9Using_Linux_Tools_83', 'ch9Using_Linux_Tools_84', 'ch9Using_Linux_Tools_85', 'ch9Using_Linux_Tools_86', 'ch9Using_Linux_Tools_87', 'ch9Using_Linux_Tools_88', 'ch9Using_Linux_Tools_89', 'ch9Using_Linux_Tools_8a', 'ch9Using_Linux_Tools_8b', 'ch9Using_Linux_Tools_8c', 'ch9Using_Linux_Tools_8d', 'ch9Using_Linux_Tools_8e', 'ch9Using_Linux_Tools_8f', 'ch9Using_Linux_Tools_90', 'ch9Using_Linux_Tools_91', 'ch9Using_Linux_Tools_92', 'ch9Using_Linux_Tools_93', 'ch9Using_Linux_Tools_94', 'ch9Using_Linux_Tools_95', 'ch9Using_Linux_Tools_96', 'ch9Using_Linux_Tools_97', 'ch9Using_Linux_Tools_98', 'ch9Using_Linux_Tools_99', 'ch9Using_Linux_Tools_9a', 'ch9Using_Linux_Tools_9b', 'ch9Using_Linux_Tools_9c', 'ch9Using_Linux_Tools_9d', 'ch9Using_Linux_Tools_9e', 'ch9Using_Linux_Tools_9f', 'ch9Using_Linux_Tools_a0', 'ch9Using_Linux_Tools_a1', 'ch9Using_Linux_Tools_a2', 'ch9Using_Linux_Tools_a3', 'ch9Using_Linux_Tools_a4', 'ch9Using_Linux_Tools_a5', 'ch9Using_Linux_Tools_a6', 'ch9Using_Linux_Tools_a7', 'ch9Using_Linux_Tools_a8', 'ch9Using_Linux_Tools_a9', 'ch9Using_Linux_Tools_aa', 'ch9Using_Linux_Tools_ab', 'ch9Using_Linux_Tools_ac', 'ch9Using_Linux_Tools_ad', 'ch9Using_Linux_Tools_ae', 'ch9Using_Linux_Tools_af', 'ch9Using_Linux_Tools_b0', 'ch9Using_Linux_Tools_b1', 'ch9Using_Linux_Tools_b2', 'ch9Using_Linux_Tools_b3', 'ch9Using_Linux_Tools_b4', 'ch9Using_Linux_Tools_b5', 'ch9Using_Linux_Tools_b6', 'ch9Using_Linux_Tools_b7', 'ch9Using_Linux_Tools_b8', 'ch9Using_Linux_Tools_b9', 'ch9Using_Linux_Tools_ba', 'ch9Using_Linux_Tools_bb', 'ch9Using_Linux_Tools_bc', 'ch9Using_Linux_Tools_bd', 'ch9Using_Linux_Tools_be', 'ch9Using_Linux_Tools_bf', 'ch9Using_Linux_Tools_c0', 'ch9Using_Linux_Tools_c1', 'ch9Using_Linux_Tools_c2', 'ch9Using_Linux_Tools_c3', 'ch9Using_Linux_Tools_c4', 'ch9Using_Linux_Tools_c5', 'ch9Using_Linux_Tools_c6', 'ch9Using_Linux_Tools_c7', 'ch9Using_Linux_Tools_c8', 'ch9Using_Linux_Tools_c9', 'ch9Using_Linux_Tools_ca', 'ch9Using_Linux_Tools_cb', 'ch9Using_Linux_Tools_cc', 'ch9Using_Linux_Tools_cd', 'ch9Using_Linux_Tools_ce', 'ch9Using_Linux_Tools_cf', 'ch9Using_Linux_Tools_d0', 'ch9Using_Linux_Tools_d1', 'ch9Using_Linux_Tools_d2', 'ch9Using_Linux_Tools_d3', 'ch9Using_Linux_Tools_d4', 'ch9Using_Linux_Tools_d5', 'ch9Using_Linux_Tools_d6', 'ch9Using_Linux_Tools_d7', 'ch9Using_Linux_Tools_d8', 'ch9Using_Linux_Tools_d9', 'ch9Using_Linux_Tools_da', 'ch9Using_Linux_Tools_db', 'ch9Using_Linux_Tools_dc', 'ch9Using_Linux_Tools_dd', 'ch9Using_Linux_Tools_de', 'ch9Using_Linux_Tools_df', 'ch9Using_Linux_Tools_e0', 'ch9Using_Linux_Tools_e1', 'ch9Using_Linux_Tools_e2', 'ch9Using_Linux_Tools_e3', 'ch9Using_Linux_Tools_e4', 'ch9Using_Linux_Tools_e5', 'ch9Using_Linux_Tools_e6', 'ch9Using_Linux_Tools_e7', 'ch9Using_Linux_Tools_e8', 'ch9Using_Linux_Tools_e9', 'ch9Using_Linux_Tools_ea', 'ch9Using_Linux_Tools_eb', 'ch9Using_Linux_Tools_ec', 'ch9Using_Linux_Tools_ed', 'ch9Using_Linux_Tools_ee', 'ch9Using_Linux_Tools_ef', 'ch9Using_Linux_Tools_f0', 'ch9Using_Linux_Tools_f1', 'ch9Using_Linux_Tools_f2', 'ch9Using_Linux_Tools_f3', 'ch9Using_Linux_Tools_f4', 'ch9Using_Linux_Tools_f5', 'ch9Using_Linux_Tools_f6', 'ch9Using_Linux_Tools_f7', 'ch9Using_Linux_Tools_f8', 'ch9Using_Linux_Tools_f9', 'ch9Using_Linux_Tools_fa', 'ch9Using_Linux_Tools_fb', 'ch9Using_Linux_Tools_fc', 'ch9Using_Linux_Tools_fd', 'ch9Using_Linux_Tools_fe', 'ch9Using_Linux_Tools_ff', 'ch9Using_Linux_Tools_100', 'ch9Using_Linux_Tools_101', 'ch9Using_Linux_Tools_102', 'ch9Using_Linux_Tools_103', 'ch9Using_Linux_Tools_104', 'ch9Using_Linux_Tools_105', 'ch9Using_Linux_Tools_106', 'ch9Using_Linux_Tools_107', 'ch9Using_Linux_Tools_108', 'ch9Using_Linux_Tools_109', 'ch9Using_Linux_Tools_10a', 'ch9Using_Linux_Tools_10b', 'ch9Using_Linux_Tools_10c', 'ch9Using_Linux_Tools_10d', 'ch9Using_Linux_Tools_10e', 'ch9Using_Linux_Tools_10f', 'ch9Using_Linux_Tools_110', 'ch9Using_Linux_Tools_111', 'ch9Using_Linux_Tools_112', 'ch9Using_Linux_Tools_113', 'ch9Using_Linux_Tools_114', 'ch9Using_Linux_Tools_115', 'ch9Using_Linux_Tools_116', 'ch9Using_Linux_Tools_117', 'ch9Using_Linux_Tools_118', 'ch9Using_Linux_Tools_119', 'ch9Using_Linux_Tools_11a', 'ch9Using_Linux_Tools_11b', 'ch9Using_Linux_Tools_11c', 'ch9Using_Linux_Tools_11d', 'ch9Using_Linux_Tools_11e', 'ch9Using_Linux_Tools_11f', 'ch9Using_Linux_Tools_120', 'ch9Using_Linux_Tools_121', 'ch9Using_Linux_Tools_122', 'ch9Using_Linux_Tools_123', 'ch9Using_Linux_Tools_124', 'ch9Using_Linux_Tools_125', 'ch9Using_Linux_Tools_126', 'ch9Using_Linux_Tools_127', 'ch9Using_Linux_Tools_128', 'ch9Using_Linux_Tools_129', 'ch9Using_Linux_Tools_12a', 'ch9Using_Linux_Tools_12b', 'ch9Using_Linux_Tools_12c', 'ch9Using_Linux_Tools_12d', 'ch9Using_Linux_Tools_12e', 'ch9Using_Linux_Tools_12f', 'ch9Using_Linux_Tools_130', 'ch9Using_Linux_Tools_131', 'ch9Using_Linux_Tools_132', 'ch9Using_Linux_Tools_133', 'ch9Using_Linux_Tools_134', 'ch9Using_Linux_Tools_135', 'ch9Using_Linux_Tools_136', 'ch9Using_Linux_Tools_137', 'ch9Using_Linux_Tools_138', 'ch9Using_Linux_Tools_139', 'ch9Using_Linux_Tools_13a', 'ch9Using_Linux_Tools_13b', 'ch9Using_Linux_Tools_13c', 'ch9Using_Linux_Tools_13d', 'ch9Using_Linux_Tools_13e', 'ch9Using_Linux_Tools_13f', 'ch9Using_Linux_Tools_140', 'ch9Using_Linux_Tools_141', 'ch9Using_Linux_Tools_142', 'ch9Using_Linux_Tools_143', 'ch9Using_Linux_Tools_144', 'ch9Using_Linux_Tools_145', 'ch9Using_Linux_Tools_146', 'ch9Using_Linux_Tools_147', 'ch9Using_Linux_Tools_148', 'ch9Using_Linux_Tools_149', 'ch9Using_Linux_Tools_14a', 'ch9Using_Linux_Tools_14b', 'ch9Using_Linux_Tools_14c', 'ch9Using_Linux_Tools_14d', 'ch9Using_Linux_Tools_14e', 'ch9Using_Linux_Tools_14f', 'ch9Using_Linux_Tools_150', 'ch9Using_Linux_Tools_151', 'ch9Using_Linux_Tools_152', 'ch9Using_Linux_Tools_153', 'ch9Using_Linux_Tools_154', 'ch9Using_Linux_Tools_155', 'ch9Using_Linux_Tools_156', 'ch9Using_Linux_Tools_157', 'ch9Using_Linux_Tools_158', 'ch9Using_Linux_Tools_159', 'ch9Using_Linux_Tools_15a', 'ch9Using_Linux_Tools_15b', 'ch9Using_Linux_Tools_15c', 'ch9Using_Linux_Tools_15d', 'ch9Using_Linux_Tools_15e', 'ch9Using_Linux_Tools_15f', 'ch9Using_Linux_Tools_160', 'ch9Using_Linux_Tools_161', 'ch9Using_Linux_Tools_162', 'ch9Using_Linux_Tools_163', 'ch9Using_Linux_Tools_164', 'ch9Using_Linux_Tools_165', 'ch9Using_Linux_Tools_166', 'ch9Using_Linux_Tools_167', 'ch9Using_Linux_Tools_168', 'ch9Using_Linux_Tools_169', 'ch9Using_Linux_Tools_16a', 'ch9Using_Linux_Tools_16b', 'ch9Using_Linux_Tools_16c', 'ch9Using_Linux_Tools_16d', 'ch9Using_Linux_Tools_16e', 'ch9Using_Linux_Tools_16f', 'ch9Using_Linux_Tools_170', 'ch9Using_Linux_Tools_171', 'ch9Using_Linux_Tools_172', 'ch9Using_Linux_Tools_173', 'ch9Using_Linux_Tools_174', 'ch9Using_Linux_Tools_175', 'ch9Using_Linux_Tools_176', 'ch9Using_Linux_Tools_177', 'ch9Using_Linux_Tools_178', 'ch9Using_Linux_Tools_179', 'ch9Using_Linux_Tools_17a', 'ch9Using_Linux_Tools_17b', 'ch9Using_Linux_Tools_17c', 'ch9Using_Linux_Tools_17d', 'ch9Using_Linux_Tools_17e', 'ch9Using_Linux_Tools_17f', 'ch9Using_Linux_Tools_180', 'ch9Using_Linux_Tools_181', 'ch9Using_Linux_Tools_182', 'ch9Using_Linux_Tools_183', 'ch9Using_Linux_Tools_184', 'ch9Using_Linux_Tools_185', 'ch9Using_Linux_Tools_186', 'ch9Using_Linux_Tools_187', 'ch9Using_Linux_Tools_188', 'ch9Using_Linux_Tools_189', 'ch9Using_Linux_Tools_18a', 'ch9Using_Linux_Tools_18b', 'ch9Using_Linux_Tools_18c', 'ch9Using_Linux_Tools_18d', 'ch9Using_Linux_Tools_18e', 'ch9Using_Linux_Tools_18f', 'ch9Using_Linux_Tools_190', 'ch9Using_Linux_Tools_191', 'ch9Using_Linux_Tools_192', 'ch9Using_Linux_Tools_193', 'ch9Using_Linux_Tools_194', 'ch9Using_Linux_Tools_195', 'ch9Using_Linux_Tools_196', 'ch9Using_Linux_Tools_197', 'ch9Using_Linux_Tools_198', 'ch9Using_Linux_Tools_199', 'ch9Using_Linux_Tools_19a', 'ch9Using_Linux_Tools_19b', 'ch9Using_Linux_Tools_19c', 'ch9Using_Linux_Tools_19d', 'ch9Using_Linux_Tools_19e', 'ch9Using_Linux_Tools_19f', 'ch9Using_Linux_Tools_1a0', 'ch9Using_Linux_Tools_1a1', 'ch9Using_Linux_Tools_1a2', 'ch9Using_Linux_Tools_1a3', 'ch9Using_Linux_Tools_1a4', 'ch9Using_Linux_Tools_1a5', 'ch9Using_Linux_Tools_1a6', 'ch9Using_Linux_Tools_1a7', 'ch9Using_Linux_Tools_1a8', 'ch9Using_Linux_Tools_1a9', 'ch9Using_Linux_Tools_1aa', 'ch9Using_Linux_Tools_1ab', 'ch9Using_Linux_Tools_1ac', 'ch9Using_Linux_Tools_1ad', 'ch9Using_Linux_Tools_1ae', 'ch9Using_Linux_Tools_1af', 'ch9Using_Linux_Tools_1b0', 'ch9Using_Linux_Tools_1b1', 'ch9Using_Linux_Tools_1b2', 'ch9Using_Linux_Tools_1b3', 'ch9Using_Linux_Tools_1b4', 'ch9Using_Linux_Tools_1b5', 'ch9Using_Linux_Tools_1b6', 'ch9Using_Linux_Tools_1b7', 'ch9Using_Linux_Tools_1b8', 'ch9Using_Linux_Tools_1b9', 'ch9Using_Linux_Tools_1ba', 'ch9Using_Linux_Tools_1bb', 'ch9Using_Linux_Tools_1bc', 'ch9Using_Linux_Tools_1bd', 'ch9Using_Linux_Tools_1be', 'ch9Using_Linux_Tools_1bf', 'ch9Using_Linux_Tools_1c0', 'ch9Using_Linux_Tools_1c1', 'ch9Using_Linux_Tools_1c2', 'ch9Using_Linux_Tools_1c3', 'ch9Using_Linux_Tools_1c4', 'ch9Using_Linux_Tools_1c5', 'ch9Using_Linux_Tools_1c6', 'ch9Using_Linux_Tools_1c7', 'ch9Using_Linux_Tools_1c8', 'ch9Using_Linux_Tools_1c9', 'ch9Using_Linux_Tools_1ca', 'ch9Using_Linux_Tools_1cb', 'ch9Using_Linux_Tools_1cc', 'ch9Using_Linux_Tools_1cd', 'ch9Using_Linux_Tools_1ce', 'ch9Using_Linux_Tools_1cf', 'ch9Using_Linux_Tools_1d0', 'ch9Using_Linux_Tools_1d1', 'ch9Using_Linux_Tools_1d2', 'ch9Using_Linux_Tools_1d3', 'ch9Using_Linux_Tools_1d4', 'ch9Using_Linux_Tools_1d5', 'ch9Using_Linux_Tools_1d6', 'ch9Using_Linux_Tools_1d7', 'ch9Using_Linux_Tools_1d8', 'ch9Using_Linux_Tools_1d9', 'ch9Using_Linux_Tools_1da', 'ch9Using_Linux_Tools_1db', 'ch9Using_Linux_Tools_1dc', 'ch9Using_Linux_Tools_1dd', 'ch9Using_Linux_Tools_1de', 'ch9Using_Linux_Tools_1df', 'ch9Using_Linux_Tools_1e0', 'ch9Using_Linux_Tools_1e1'], 'ch1Introduction': ['ch1Introduction_1', 'ch1Introduction_2', 'ch1Introduction_3', 'ch1Introduction_4', 'ch1Introduction_5', 'ch1Introduction_6', 'ch1Introduction_7', 'ch1Introduction_8', 'ch1Introduction_9', 'ch1Introduction_a', 'ch1Introduction_b', 'ch1Introduction_c', 'ch1Introduction_d', 'ch1Introduction_e', 'ch1Introduction_f', 'ch1Introduction_10', 'ch1Introduction_11', 'ch1Introduction_12', 'ch1Introduction_13', 'ch1Introduction_14', 'ch1Introduction_15', 'ch1Introduction_16', 'ch1Introduction_17', 'ch1Introduction_18', 'ch1Introduction_19', 'ch1Introduction_1a', 'ch1Introduction_1b', 'ch1Introduction_1c', 'ch1Introduction_1d', 'ch1Introduction_1e', 'ch1Introduction_1f', 'ch1Introduction_20', 'ch1Introduction_21', 'ch1Introduction_22', 'ch1Introduction_23', 'ch1Introduction_24', 'ch1Introduction_25', 'ch1Introduction_26', 'ch1Introduction_27', 'ch1Introduction_28', 'ch1Introduction_29', 'ch1Introduction_2a', 'ch1Introduction_2b', 'ch1Introduction_2c', 'ch1Introduction_2d', 'ch1Introduction_2e', 'ch1Introduction_2f', 'ch1Introduction_30', 'ch1Introduction_31', 'ch1Introduction_32', 'ch1Introduction_33', 'ch1Introduction_34', 'ch1Introduction_35', 'ch1Introduction_36', 'ch1Introduction_37', 'ch1Introduction_38', 'ch1Introduction_39', 'ch1Introduction_3a', 'ch1Introduction_3b', 'ch1Introduction_3c', 'ch1Introduction_3d', 'ch1Introduction_3e', 'ch1Introduction_3f', 'ch1Introduction_40', 'ch1Introduction_41', 'ch1Introduction_42', 'ch1Introduction_43', 'ch1Introduction_44', 'ch1Introduction_45', 'ch1Introduction_46', 'ch1Introduction_47', 'ch1Introduction_48', 'ch1Introduction_49', 'ch1Introduction_4a', 'ch1Introduction_4b', 'ch1Introduction_4c', 'ch1Introduction_4d', 'ch1Introduction_4e', 'ch1Introduction_4f', 'ch1Introduction_50', 'ch1Introduction_51', 'ch1Introduction_52', 'ch1Introduction_53', 'ch1Introduction_54', 'ch1Introduction_55', 'ch1Introduction_56', 'ch1Introduction_57', 'ch1Introduction_58', 'ch1Introduction_59', 'ch1Introduction_5a', 'ch1Introduction_5b', 'ch1Introduction_5c', 'ch1Introduction_5d', 'ch1Introduction_5e', 'ch1Introduction_5f', 'ch1Introduction_60', 'ch1Introduction_61', 'ch1Introduction_62', 'ch1Introduction_63', 'ch1Introduction_64', 'ch1Introduction_65', 'ch1Introduction_66', 'ch1Introduction_67', 'ch1Introduction_68', 'ch1Introduction_69', 'ch1Introduction_6a', 'ch1Introduction_6b', 'ch1Introduction_6c', 'ch1Introduction_6d', 'ch1Introduction_6e', 'ch1Introduction_6f', 'ch1Introduction_70', 'ch1Introduction_71', 'ch1Introduction_72', 'ch1Introduction_73', 'ch1Introduction_74', 'ch1Introduction_75', 'ch1Introduction_76', 'ch1Introduction_77', 'ch1Introduction_78', 'ch1Introduction_79', 'ch1Introduction_7a', 'ch1Introduction_7b', 'ch1Introduction_7c', 'ch1Introduction_7d', 'ch1Introduction_7e', 'ch1Introduction_7f', 'ch1Introduction_80', 'ch1Introduction_81', 'ch1Introduction_82', 'ch1Introduction_83', 'ch1Introduction_84', 'ch1Introduction_85', 'ch1Introduction_86', 'ch1Introduction_87', 'ch1Introduction_88', 'ch1Introduction_89', 'ch1Introduction_8a', 'ch1Introduction_8b', 'ch1Introduction_8c', 'ch1Introduction_8d', 'ch1Introduction_8e', 'ch1Introduction_8f', 'ch1Introduction_90', 'ch1Introduction_91', 'ch1Introduction_92', 'ch1Introduction_93', 'ch1Introduction_94', 'ch1Introduction_95', 'ch1Introduction_96', 'ch1Introduction_97', 'ch1Introduction_98', 'ch1Introduction_99', 'ch1Introduction_9a', 'ch1Introduction_9b', 'ch1Introduction_9c', 'ch1Introduction_9d', 'ch1Introduction_9e', 'ch1Introduction_9f', 'ch1Introduction_a0', 'ch1Introduction_a1', 'ch1Introduction_a2', 'ch1Introduction_a3', 'ch1Introduction_a4', 'ch1Introduction_a5', 'ch1Introduction_a6', 'ch1Introduction_a7', 'ch1Introduction_a8', 'ch1Introduction_a9', 'ch1Introduction_aa', 'ch1Introduction_ab', 'ch1Introduction_ac', 'ch1Introduction_ad', 'ch1Introduction_ae', 'ch1Introduction_af', 'ch1Introduction_b0', 'ch1Introduction_b1', 'ch1Introduction_b2', 'ch1Introduction_b3', 'ch1Introduction_b4', 'ch1Introduction_b5', 'ch1Introduction_b6', 'ch1Introduction_b7', 'ch1Introduction_b8', 'ch1Introduction_b9', 'ch1Introduction_ba', 'ch1Introduction_bb', 'ch1Introduction_bc', 'ch1Introduction_bd', 'ch1Introduction_be', 'ch1Introduction_bf', 'ch1Introduction_c0', 'ch1Introduction_c1', 'ch1Introduction_c2', 'ch1Introduction_c3', 'ch1Introduction_c4', 'ch1Introduction_c5', 'ch1Introduction_c6', 'ch1Introduction_c7', 'ch1Introduction_c8', 'ch1Introduction_c9', 'ch1Introduction_ca', 'ch1Introduction_cb', 'ch1Introduction_cc', 'ch1Introduction_cd', 'ch1Introduction_ce', 'ch1Introduction_cf', 'ch1Introduction_d0', 'ch1Introduction_d1', 'ch1Introduction_d2', 'ch1Introduction_d3', 'ch1Introduction_d4', 'ch1Introduction_d5', 'ch1Introduction_d6', 'ch1Introduction_d7', 'ch1Introduction_d8', 'ch1Introduction_d9', 'ch1Introduction_da', 'ch1Introduction_db', 'ch1Introduction_dc', 'ch1Introduction_dd', 'ch1Introduction_de', 'ch1Introduction_df', 'ch1Introduction_e0', 'ch1Introduction_e1', 'ch1Introduction_e2', 'ch1Introduction_e3', 'ch1Introduction_e4', 'ch1Introduction_e5', 'ch1Introduction_e6', 'ch1Introduction_e7', 'ch1Introduction_e8', 'ch1Introduction_e9', 'ch1Introduction_ea', 'ch1Introduction_eb', 'ch1Introduction_ec', 'ch1Introduction_ed', 'ch1Introduction_ee', 'ch1Introduction_ef', 'ch1Introduction_f0', 'ch1Introduction_f1', 'ch1Introduction_f2', 'ch1Introduction_f3', 'ch1Introduction_f4', 'ch1Introduction_f5', 'ch1Introduction_f6', 'ch1Introduction_f7', 'ch1Introduction_f8', 'ch1Introduction_f9', 'ch1Introduction_fa', 'ch1Introduction_fb', 'ch1Introduction_fc', 'ch1Introduction_fd', 'ch1Introduction_fe', 'ch1Introduction_ff', 'ch1Introduction_100', 'ch1Introduction_101', 'ch1Introduction_102', 'ch1Introduction_103', 'ch1Introduction_104', 'ch1Introduction_105', 'ch1Introduction_106', 'ch1Introduction_107', 'ch1Introduction_108', 'ch1Introduction_109', 'ch1Introduction_10a', 'ch1Introduction_10b', 'ch1Introduction_10c', 'ch1Introduction_10d', 'ch1Introduction_10e', 'ch1Introduction_10f', 'ch1Introduction_110', 'ch1Introduction_111', 'ch1Introduction_112', 'ch1Introduction_113', 'ch1Introduction_114', 'ch1Introduction_115', 'ch1Introduction_116', 'ch1Introduction_117', 'ch1Introduction_118', 'ch1Introduction_119', 'ch1Introduction_11a', 'ch1Introduction_11b', 'ch1Introduction_11c', 'ch1Introduction_11d', 'ch1Introduction_11e', 'ch1Introduction_11f', 'ch1Introduction_120', 'ch1Introduction_121', 'ch1Introduction_122', 'ch1Introduction_123', 'ch1Introduction_124', 'ch1Introduction_125', 'ch1Introduction_126', 'ch1Introduction_127', 'ch1Introduction_128', 'ch1Introduction_129', 'ch1Introduction_12a', 'ch1Introduction_12b', 'ch1Introduction_12c', 'ch1Introduction_12d', 'ch1Introduction_12e', 'ch1Introduction_12f', 'ch1Introduction_130', 'ch1Introduction_131', 'ch1Introduction_132', 'ch1Introduction_133', 'ch1Introduction_134', 'ch1Introduction_135', 'ch1Introduction_136', 'ch1Introduction_137', 'ch1Introduction_138', 'ch1Introduction_139', 'ch1Introduction_13a', 'ch1Introduction_13b', 'ch1Introduction_13c', 'ch1Introduction_13d', 'ch1Introduction_13e', 'ch1Introduction_13f', 'ch1Introduction_140', 'ch1Introduction_141', 'ch1Introduction_142', 'ch1Introduction_143', 'ch1Introduction_144', 'ch1Introduction_145', 'ch1Introduction_146', 'ch1Introduction_147', 'ch1Introduction_148', 'ch1Introduction_149', 'ch1Introduction_14a', 'ch1Introduction_14b', 'ch1Introduction_14c', 'ch1Introduction_14d', 'ch1Introduction_14e', 'ch1Introduction_14f', 'ch1Introduction_150', 'ch1Introduction_151', 'ch1Introduction_152', 'ch1Introduction_153', 'ch1Introduction_154', 'ch1Introduction_155', 'ch1Introduction_156', 'ch1Introduction_157', 'ch1Introduction_158', 'ch1Introduction_159', 'ch1Introduction_15a', 'ch1Introduction_15b', 'ch1Introduction_15c', 'ch1Introduction_15d', 'ch1Introduction_15e', 'ch1Introduction_15f', 'ch1Introduction_160', 'ch1Introduction_161', 'ch1Introduction_162', 'ch1Introduction_163', 'ch1Introduction_164', 'ch1Introduction_165', 'ch1Introduction_166', 'ch1Introduction_167', 'ch1Introduction_168', 'ch1Introduction_169', 'ch1Introduction_16a', 'ch1Introduction_16b', 'ch1Introduction_16c', 'ch1Introduction_16d', 'ch1Introduction_16e', 'ch1Introduction_16f', 'ch1Introduction_170', 'ch1Introduction_171', 'ch1Introduction_172', 'ch1Introduction_173', 'ch1Introduction_174', 'ch1Introduction_175', 'ch1Introduction_176', 'ch1Introduction_177', 'ch1Introduction_178', 'ch1Introduction_179', 'ch1Introduction_17a', 'ch1Introduction_17b', 'ch1Introduction_17c', 'ch1Introduction_17d', 'ch1Introduction_17e', 'ch1Introduction_17f', 'ch1Introduction_180', 'ch1Introduction_181', 'ch1Introduction_182', 'ch1Introduction_183', 'ch1Introduction_184', 'ch1Introduction_185', 'ch1Introduction_186', 'ch1Introduction_187', 'ch1Introduction_188', 'ch1Introduction_189', 'ch1Introduction_18a', 'ch1Introduction_18b', 'ch1Introduction_18c', 'ch1Introduction_18d', 'ch1Introduction_18e', 'ch1Introduction_18f', 'ch1Introduction_190', 'ch1Introduction_191', 'ch1Introduction_192', 'ch1Introduction_193', 'ch1Introduction_194', 'ch1Introduction_195', 'ch1Introduction_196', 'ch1Introduction_197', 'ch1Introduction_198', 'ch1Introduction_199', 'ch1Introduction_19a', 'ch1Introduction_19b', 'ch1Introduction_19c', 'ch1Introduction_19d', 'ch1Introduction_19e', 'ch1Introduction_19f', 'ch1Introduction_1a0', 'ch1Introduction_1a1', 'ch1Introduction_1a2', 'ch1Introduction_1a3', 'ch1Introduction_1a4', 'ch1Introduction_1a5', 'ch1Introduction_1a6', 'ch1Introduction_1a7', 'ch1Introduction_1a8', 'ch1Introduction_1a9', 'ch1Introduction_1aa', 'ch1Introduction_1ab', 'ch1Introduction_1ac', 'ch1Introduction_1ad', 'ch1Introduction_1ae', 'ch1Introduction_1af', 'ch1Introduction_1b0', 'ch1Introduction_1b1', 'ch1Introduction_1b2', 'ch1Introduction_1b3', 'ch1Introduction_1b4', 'ch1Introduction_1b5', 'ch1Introduction_1b6', 'ch1Introduction_1b7', 'ch1Introduction_1b8', 'ch1Introduction_1b9', 'ch1Introduction_1ba', 'ch1Introduction_1bb', 'ch1Introduction_1bc', 'ch1Introduction_1bd', 'ch1Introduction_1be', 'ch1Introduction_1bf', 'ch1Introduction_1c0', 'ch1Introduction_1c1', 'ch1Introduction_1c2', 'ch1Introduction_1c3', 'ch1Introduction_1c4', 'ch1Introduction_1c5', 'ch1Introduction_1c6', 'ch1Introduction_1c7', 'ch1Introduction_1c8', 'ch1Introduction_1c9', 'ch1Introduction_1ca', 'ch1Introduction_1cb', 'ch1Introduction_1cc', 'ch1Introduction_1cd', 'ch1Introduction_1ce', 'ch1Introduction_1cf', 'ch1Introduction_1d0', 'ch1Introduction_1d1', 'ch1Introduction_1d2', 'ch1Introduction_1d3', 'ch1Introduction_1d4', 'ch1Introduction_1d5', 'ch1Introduction_1d6', 'ch1Introduction_1d7', 'ch1Introduction_1d8', 'ch1Introduction_1d9', 'ch1Introduction_1da', 'ch1Introduction_1db', 'ch1Introduction_1dc', 'ch1Introduction_1dd', 'ch1Introduction_1de', 'ch1Introduction_1df', 'ch1Introduction_1e0', 'ch1Introduction_1e1'], 'ch5func': ['ch5func_1', 'ch5func_2', 'ch5func_3', 'ch5func_4', 'ch5func_5', 'ch5func_6', 'ch5func_7', 'ch5func_8', 'ch5func_9', 'ch5func_a', 'ch5func_b', 'ch5func_c', 'ch5func_d', 'ch5func_e', 'ch5func_f', 'ch5func_10', 'ch5func_11', 'ch5func_12', 'ch5func_13', 'ch5func_14', 'ch5func_15', 'ch5func_16', 'ch5func_17', 'ch5func_18', 'ch5func_19', 'ch5func_1a', 'ch5func_1b', 'ch5func_1c', 'ch5func_1d', 'ch5func_1e', 'ch5func_1f', 'ch5func_20', 'ch5func_21', 'ch5func_22', 'ch5func_23', 'ch5func_24', 'ch5func_25', 'ch5func_26', 'ch5func_27', 'ch5func_28', 'ch5func_29', 'ch5func_2a', 'ch5func_2b', 'ch5func_2c', 'ch5func_2d', 'ch5func_2e', 'ch5func_2f', 'ch5func_30', 'ch5func_31', 'ch5func_32', 'ch5func_33', 'ch5func_34', 'ch5func_35', 'ch5func_36', 'ch5func_37', 'ch5func_38', 'ch5func_39', 'ch5func_3a', 'ch5func_3b', 'ch5func_3c', 'ch5func_3d', 'ch5func_3e', 'ch5func_3f', 'ch5func_40', 'ch5func_41', 'ch5func_42', 'ch5func_43', 'ch5func_44', 'ch5func_45', 'ch5func_46', 'ch5func_47', 'ch5func_48', 'ch5func_49', 'ch5func_4a', 'ch5func_4b', 'ch5func_4c', 'ch5func_4d', 'ch5func_4e', 'ch5func_4f', 'ch5func_50', 'ch5func_51', 'ch5func_52', 'ch5func_53', 'ch5func_54', 'ch5func_55', 'ch5func_56', 'ch5func_57', 'ch5func_58', 'ch5func_59', 'ch5func_5a', 'ch5func_5b', 'ch5func_5c', 'ch5func_5d', 'ch5func_5e', 'ch5func_5f', 'ch5func_60', 'ch5func_61', 'ch5func_62', 'ch5func_63', 'ch5func_64', 'ch5func_65', 'ch5func_66', 'ch5func_67', 'ch5func_68', 'ch5func_69', 'ch5func_6a', 'ch5func_6b', 'ch5func_6c', 'ch5func_6d', 'ch5func_6e', 'ch5func_6f', 'ch5func_70', 'ch5func_71', 'ch5func_72', 'ch5func_73', 'ch5func_74', 'ch5func_75', 'ch5func_76', 'ch5func_77', 'ch5func_78', 'ch5func_79', 'ch5func_7a', 'ch5func_7b', 'ch5func_7c', 'ch5func_7d', 'ch5func_7e', 'ch5func_7f', 'ch5func_80', 'ch5func_81', 'ch5func_82', 'ch5func_83', 'ch5func_84', 'ch5func_85', 'ch5func_86', 'ch5func_87', 'ch5func_88', 'ch5func_89', 'ch5func_8a', 'ch5func_8b', 'ch5func_8c', 'ch5func_8d', 'ch5func_8e', 'ch5func_8f', 'ch5func_90', 'ch5func_91', 'ch5func_92', 'ch5func_93', 'ch5func_94', 'ch5func_95', 'ch5func_96', 'ch5func_97', 'ch5func_98', 'ch5func_99', 'ch5func_9a', 'ch5func_9b', 'ch5func_9c', 'ch5func_9d', 'ch5func_9e', 'ch5func_9f', 'ch5func_a0', 'ch5func_a1', 'ch5func_a2', 'ch5func_a3', 'ch5func_a4', 'ch5func_a5', 'ch5func_a6', 'ch5func_a7', 'ch5func_a8', 'ch5func_a9', 'ch5func_aa', 'ch5func_ab', 'ch5func_ac', 'ch5func_ad', 'ch5func_ae', 'ch5func_af', 'ch5func_b0', 'ch5func_b1', 'ch5func_b2', 'ch5func_b3', 'ch5func_b4', 'ch5func_b5', 'ch5func_b6', 'ch5func_b7', 'ch5func_b8', 'ch5func_b9', 'ch5func_ba', 'ch5func_bb', 'ch5func_bc', 'ch5func_bd', 'ch5func_be', 'ch5func_bf', 'ch5func_c0', 'ch5func_c1', 'ch5func_c2', 'ch5func_c3', 'ch5func_c4', 'ch5func_c5', 'ch5func_c6', 'ch5func_c7', 'ch5func_c8', 'ch5func_c9', 'ch5func_ca', 'ch5func_cb', 'ch5func_cc', 'ch5func_cd', 'ch5func_ce', 'ch5func_cf', 'ch5func_d0', 'ch5func_d1', 'ch5func_d2', 'ch5func_d3', 'ch5func_d4', 'ch5func_d5', 'ch5func_d6', 'ch5func_d7', 'ch5func_d8', 'ch5func_d9', 'ch5func_da', 'ch5func_db', 'ch5func_dc', 'ch5func_dd', 'ch5func_de', 'ch5func_df', 'ch5func_e0', 'ch5func_e1', 'ch5func_e2', 'ch5func_e3', 'ch5func_e4', 'ch5func_e5', 'ch5func_e6', 'ch5func_e7', 'ch5func_e8', 'ch5func_e9', 'ch5func_ea', 'ch5func_eb', 'ch5func_ec', 'ch5func_ed', 'ch5func_ee', 'ch5func_ef', 'ch5func_f0', 'ch5func_f1', 'ch5func_f2', 'ch5func_f3', 'ch5func_f4', 'ch5func_f5', 'ch5func_f6', 'ch5func_f7', 'ch5func_f8', 'ch5func_f9', 'ch5func_fa', 'ch5func_fb', 'ch5func_fc', 'ch5func_fd', 'ch5func_fe', 'ch5func_ff', 'ch5func_100', 'ch5func_101', 'ch5func_102', 'ch5func_103', 'ch5func_104', 'ch5func_105', 'ch5func_106', 'ch5func_107', 'ch5func_108', 'ch5func_109', 'ch5func_10a', 'ch5func_10b', 'ch5func_10c', 'ch5func_10d', 'ch5func_10e', 'ch5func_10f', 'ch5func_110', 'ch5func_111', 'ch5func_112', 'ch5func_113', 'ch5func_114', 'ch5func_115', 'ch5func_116', 'ch5func_117', 'ch5func_118', 'ch5func_119', 'ch5func_11a', 'ch5func_11b', 'ch5func_11c', 'ch5func_11d', 'ch5func_11e', 'ch5func_11f', 'ch5func_120', 'ch5func_121', 'ch5func_122', 'ch5func_123', 'ch5func_124', 'ch5func_125', 'ch5func_126', 'ch5func_127', 'ch5func_128', 'ch5func_129', 'ch5func_12a', 'ch5func_12b', 'ch5func_12c', 'ch5func_12d', 'ch5func_12e', 'ch5func_12f', 'ch5func_130', 'ch5func_131', 'ch5func_132', 'ch5func_133', 'ch5func_134', 'ch5func_135', 'ch5func_136', 'ch5func_137', 'ch5func_138', 'ch5func_139', 'ch5func_13a', 'ch5func_13b', 'ch5func_13c', 'ch5func_13d', 'ch5func_13e', 'ch5func_13f', 'ch5func_140', 'ch5func_141', 'ch5func_142', 'ch5func_143', 'ch5func_144', 'ch5func_145', 'ch5func_146', 'ch5func_147', 'ch5func_148', 'ch5func_149', 'ch5func_14a', 'ch5func_14b', 'ch5func_14c', 'ch5func_14d', 'ch5func_14e', 'ch5func_14f', 'ch5func_150', 'ch5func_151', 'ch5func_152', 'ch5func_153', 'ch5func_154', 'ch5func_155', 'ch5func_156', 'ch5func_157', 'ch5func_158', 'ch5func_159', 'ch5func_15a', 'ch5func_15b', 'ch5func_15c', 'ch5func_15d', 'ch5func_15e', 'ch5func_15f', 'ch5func_160', 'ch5func_161', 'ch5func_162', 'ch5func_163', 'ch5func_164', 'ch5func_165', 'ch5func_166', 'ch5func_167', 'ch5func_168', 'ch5func_169', 'ch5func_16a', 'ch5func_16b', 'ch5func_16c', 'ch5func_16d', 'ch5func_16e', 'ch5func_16f', 'ch5func_170', 'ch5func_171', 'ch5func_172', 'ch5func_173', 'ch5func_174', 'ch5func_175', 'ch5func_176', 'ch5func_177', 'ch5func_178', 'ch5func_179', 'ch5func_17a', 'ch5func_17b', 'ch5func_17c', 'ch5func_17d', 'ch5func_17e', 'ch5func_17f', 'ch5func_180', 'ch5func_181', 'ch5func_182', 'ch5func_183', 'ch5func_184', 'ch5func_185', 'ch5func_186', 'ch5func_187', 'ch5func_188', 'ch5func_189', 'ch5func_18a', 'ch5func_18b', 'ch5func_18c', 'ch5func_18d', 'ch5func_18e', 'ch5func_18f', 'ch5func_190', 'ch5func_191', 'ch5func_192', 'ch5func_193', 'ch5func_194', 'ch5func_195', 'ch5func_196', 'ch5func_197', 'ch5func_198', 'ch5func_199', 'ch5func_19a', 'ch5func_19b', 'ch5func_19c', 'ch5func_19d', 'ch5func_19e', 'ch5func_19f', 'ch5func_1a0', 'ch5func_1a1', 'ch5func_1a2', 'ch5func_1a3', 'ch5func_1a4', 'ch5func_1a5', 'ch5func_1a6', 'ch5func_1a7', 'ch5func_1a8', 'ch5func_1a9', 'ch5func_1aa', 'ch5func_1ab', 'ch5func_1ac', 'ch5func_1ad', 'ch5func_1ae', 'ch5func_1af', 'ch5func_1b0', 'ch5func_1b1', 'ch5func_1b2', 'ch5func_1b3', 'ch5func_1b4', 'ch5func_1b5', 'ch5func_1b6', 'ch5func_1b7', 'ch5func_1b8', 'ch5func_1b9', 'ch5func_1ba', 'ch5func_1bb', 'ch5func_1bc', 'ch5func_1bd', 'ch5func_1be', 'ch5func_1bf', 'ch5func_1c0', 'ch5func_1c1', 'ch5func_1c2', 'ch5func_1c3', 'ch5func_1c4', 'ch5func_1c5', 'ch5func_1c6', 'ch5func_1c7', 'ch5func_1c8', 'ch5func_1c9', 'ch5func_1ca', 'ch5func_1cb', 'ch5func_1cc', 'ch5func_1cd', 'ch5func_1ce', 'ch5func_1cf', 'ch5func_1d0', 'ch5func_1d1', 'ch5func_1d2', 'ch5func_1d3', 'ch5func_1d4', 'ch5func_1d5', 'ch5func_1d6', 'ch5func_1d7', 'ch5func_1d8', 'ch5func_1d9', 'ch5func_1da', 'ch5func_1db', 'ch5func_1dc', 'ch5func_1dd', 'ch5func_1de', 'ch5func_1df', 'ch5func_1e0', 'ch5func_1e1'], 'ch3list_tuples': ['ch3list_tuples_1', 'ch3list_tuples_2', 'ch3list_tuples_3', 'ch3list_tuples_4', 'ch3list_tuples_5', 'ch3list_tuples_6', 'ch3list_tuples_7', 'ch3list_tuples_8', 'ch3list_tuples_9', 'ch3list_tuples_a', 'ch3list_tuples_b', 'ch3list_tuples_c', 'ch3list_tuples_d', 'ch3list_tuples_e', 'ch3list_tuples_f', 'ch3list_tuples_10', 'ch3list_tuples_11', 'ch3list_tuples_12', 'ch3list_tuples_13', 'ch3list_tuples_14', 'ch3list_tuples_15', 'ch3list_tuples_16', 'ch3list_tuples_17', 'ch3list_tuples_18', 'ch3list_tuples_19', 'ch3list_tuples_1a', 'ch3list_tuples_1b', 'ch3list_tuples_1c', 'ch3list_tuples_1d', 'ch3list_tuples_1e', 'ch3list_tuples_1f', 'ch3list_tuples_20', 'ch3list_tuples_21', 'ch3list_tuples_22', 'ch3list_tuples_23', 'ch3list_tuples_24', 'ch3list_tuples_25', 'ch3list_tuples_26', 'ch3list_tuples_27', 'ch3list_tuples_28', 'ch3list_tuples_29', 'ch3list_tuples_2a', 'ch3list_tuples_2b', 'ch3list_tuples_2c', 'ch3list_tuples_2d', 'ch3list_tuples_2e', 'ch3list_tuples_2f', 'ch3list_tuples_30', 'ch3list_tuples_31', 'ch3list_tuples_32', 'ch3list_tuples_33', 'ch3list_tuples_34', 'ch3list_tuples_35', 'ch3list_tuples_36', 'ch3list_tuples_37', 'ch3list_tuples_38', 'ch3list_tuples_39', 'ch3list_tuples_3a', 'ch3list_tuples_3b', 'ch3list_tuples_3c', 'ch3list_tuples_3d', 'ch3list_tuples_3e', 'ch3list_tuples_3f', 'ch3list_tuples_40', 'ch3list_tuples_41', 'ch3list_tuples_42', 'ch3list_tuples_43', 'ch3list_tuples_44', 'ch3list_tuples_45', 'ch3list_tuples_46', 'ch3list_tuples_47', 'ch3list_tuples_48', 'ch3list_tuples_49', 'ch3list_tuples_4a', 'ch3list_tuples_4b', 'ch3list_tuples_4c', 'ch3list_tuples_4d', 'ch3list_tuples_4e', 'ch3list_tuples_4f', 'ch3list_tuples_50', 'ch3list_tuples_51', 'ch3list_tuples_52', 'ch3list_tuples_53', 'ch3list_tuples_54', 'ch3list_tuples_55', 'ch3list_tuples_56', 'ch3list_tuples_57', 'ch3list_tuples_58', 'ch3list_tuples_59', 'ch3list_tuples_5a', 'ch3list_tuples_5b', 'ch3list_tuples_5c', 'ch3list_tuples_5d', 'ch3list_tuples_5e', 'ch3list_tuples_5f', 'ch3list_tuples_60', 'ch3list_tuples_61', 'ch3list_tuples_62', 'ch3list_tuples_63', 'ch3list_tuples_64', 'ch3list_tuples_65', 'ch3list_tuples_66', 'ch3list_tuples_67', 'ch3list_tuples_68', 'ch3list_tuples_69', 'ch3list_tuples_6a', 'ch3list_tuples_6b', 'ch3list_tuples_6c', 'ch3list_tuples_6d', 'ch3list_tuples_6e', 'ch3list_tuples_6f', 'ch3list_tuples_70', 'ch3list_tuples_71', 'ch3list_tuples_72', 'ch3list_tuples_73', 'ch3list_tuples_74', 'ch3list_tuples_75', 'ch3list_tuples_76', 'ch3list_tuples_77', 'ch3list_tuples_78', 'ch3list_tuples_79', 'ch3list_tuples_7a', 'ch3list_tuples_7b', 'ch3list_tuples_7c', 'ch3list_tuples_7d', 'ch3list_tuples_7e', 'ch3list_tuples_7f', 'ch3list_tuples_80', 'ch3list_tuples_81', 'ch3list_tuples_82', 'ch3list_tuples_83', 'ch3list_tuples_84', 'ch3list_tuples_85', 'ch3list_tuples_86', 'ch3list_tuples_87', 'ch3list_tuples_88', 'ch3list_tuples_89', 'ch3list_tuples_8a', 'ch3list_tuples_8b', 'ch3list_tuples_8c', 'ch3list_tuples_8d', 'ch3list_tuples_8e', 'ch3list_tuples_8f', 'ch3list_tuples_90', 'ch3list_tuples_91', 'ch3list_tuples_92', 'ch3list_tuples_93', 'ch3list_tuples_94', 'ch3list_tuples_95', 'ch3list_tuples_96', 'ch3list_tuples_97', 'ch3list_tuples_98', 'ch3list_tuples_99', 'ch3list_tuples_9a', 'ch3list_tuples_9b', 'ch3list_tuples_9c', 'ch3list_tuples_9d', 'ch3list_tuples_9e', 'ch3list_tuples_9f', 'ch3list_tuples_a0', 'ch3list_tuples_a1', 'ch3list_tuples_a2', 'ch3list_tuples_a3', 'ch3list_tuples_a4', 'ch3list_tuples_a5', 'ch3list_tuples_a6', 'ch3list_tuples_a7', 'ch3list_tuples_a8', 'ch3list_tuples_a9', 'ch3list_tuples_aa', 'ch3list_tuples_ab', 'ch3list_tuples_ac', 'ch3list_tuples_ad', 'ch3list_tuples_ae', 'ch3list_tuples_af', 'ch3list_tuples_b0', 'ch3list_tuples_b1', 'ch3list_tuples_b2', 'ch3list_tuples_b3', 'ch3list_tuples_b4', 'ch3list_tuples_b5', 'ch3list_tuples_b6', 'ch3list_tuples_b7', 'ch3list_tuples_b8', 'ch3list_tuples_b9', 'ch3list_tuples_ba', 'ch3list_tuples_bb', 'ch3list_tuples_bc', 'ch3list_tuples_bd', 'ch3list_tuples_be', 'ch3list_tuples_bf', 'ch3list_tuples_c0', 'ch3list_tuples_c1', 'ch3list_tuples_c2', 'ch3list_tuples_c3', 'ch3list_tuples_c4', 'ch3list_tuples_c5', 'ch3list_tuples_c6', 'ch3list_tuples_c7', 'ch3list_tuples_c8', 'ch3list_tuples_c9', 'ch3list_tuples_ca', 'ch3list_tuples_cb', 'ch3list_tuples_cc', 'ch3list_tuples_cd', 'ch3list_tuples_ce', 'ch3list_tuples_cf', 'ch3list_tuples_d0', 'ch3list_tuples_d1', 'ch3list_tuples_d2', 'ch3list_tuples_d3', 'ch3list_tuples_d4', 'ch3list_tuples_d5', 'ch3list_tuples_d6', 'ch3list_tuples_d7', 'ch3list_tuples_d8', 'ch3list_tuples_d9', 'ch3list_tuples_da', 'ch3list_tuples_db', 'ch3list_tuples_dc', 'ch3list_tuples_dd', 'ch3list_tuples_de', 'ch3list_tuples_df', 'ch3list_tuples_e0', 'ch3list_tuples_e1', 'ch3list_tuples_e2', 'ch3list_tuples_e3', 'ch3list_tuples_e4', 'ch3list_tuples_e5', 'ch3list_tuples_e6', 'ch3list_tuples_e7', 'ch3list_tuples_e8', 'ch3list_tuples_e9', 'ch3list_tuples_ea', 'ch3list_tuples_eb', 'ch3list_tuples_ec', 'ch3list_tuples_ed', 'ch3list_tuples_ee', 'ch3list_tuples_ef', 'ch3list_tuples_f0', 'ch3list_tuples_f1', 'ch3list_tuples_f2', 'ch3list_tuples_f3', 'ch3list_tuples_f4', 'ch3list_tuples_f5', 'ch3list_tuples_f6', 'ch3list_tuples_f7', 'ch3list_tuples_f8', 'ch3list_tuples_f9', 'ch3list_tuples_fa', 'ch3list_tuples_fb', 'ch3list_tuples_fc', 'ch3list_tuples_fd', 'ch3list_tuples_fe', 'ch3list_tuples_ff', 'ch3list_tuples_100', 'ch3list_tuples_101', 'ch3list_tuples_102', 'ch3list_tuples_103', 'ch3list_tuples_104', 'ch3list_tuples_105', 'ch3list_tuples_106', 'ch3list_tuples_107', 'ch3list_tuples_108', 'ch3list_tuples_109', 'ch3list_tuples_10a', 'ch3list_tuples_10b', 'ch3list_tuples_10c', 'ch3list_tuples_10d', 'ch3list_tuples_10e', 'ch3list_tuples_10f', 'ch3list_tuples_110', 'ch3list_tuples_111', 'ch3list_tuples_112', 'ch3list_tuples_113', 'ch3list_tuples_114', 'ch3list_tuples_115', 'ch3list_tuples_116', 'ch3list_tuples_117', 'ch3list_tuples_118', 'ch3list_tuples_119', 'ch3list_tuples_11a', 'ch3list_tuples_11b', 'ch3list_tuples_11c', 'ch3list_tuples_11d', 'ch3list_tuples_11e', 'ch3list_tuples_11f', 'ch3list_tuples_120', 'ch3list_tuples_121', 'ch3list_tuples_122', 'ch3list_tuples_123', 'ch3list_tuples_124', 'ch3list_tuples_125', 'ch3list_tuples_126', 'ch3list_tuples_127', 'ch3list_tuples_128', 'ch3list_tuples_129', 'ch3list_tuples_12a', 'ch3list_tuples_12b', 'ch3list_tuples_12c', 'ch3list_tuples_12d', 'ch3list_tuples_12e', 'ch3list_tuples_12f', 'ch3list_tuples_130', 'ch3list_tuples_131', 'ch3list_tuples_132', 'ch3list_tuples_133', 'ch3list_tuples_134', 'ch3list_tuples_135', 'ch3list_tuples_136', 'ch3list_tuples_137', 'ch3list_tuples_138', 'ch3list_tuples_139', 'ch3list_tuples_13a', 'ch3list_tuples_13b', 'ch3list_tuples_13c', 'ch3list_tuples_13d', 'ch3list_tuples_13e', 'ch3list_tuples_13f', 'ch3list_tuples_140', 'ch3list_tuples_141', 'ch3list_tuples_142', 'ch3list_tuples_143', 'ch3list_tuples_144', 'ch3list_tuples_145', 'ch3list_tuples_146', 'ch3list_tuples_147', 'ch3list_tuples_148', 'ch3list_tuples_149', 'ch3list_tuples_14a', 'ch3list_tuples_14b', 'ch3list_tuples_14c', 'ch3list_tuples_14d', 'ch3list_tuples_14e', 'ch3list_tuples_14f', 'ch3list_tuples_150', 'ch3list_tuples_151', 'ch3list_tuples_152', 'ch3list_tuples_153', 'ch3list_tuples_154', 'ch3list_tuples_155', 'ch3list_tuples_156', 'ch3list_tuples_157', 'ch3list_tuples_158', 'ch3list_tuples_159', 'ch3list_tuples_15a', 'ch3list_tuples_15b', 'ch3list_tuples_15c', 'ch3list_tuples_15d', 'ch3list_tuples_15e', 'ch3list_tuples_15f', 'ch3list_tuples_160', 'ch3list_tuples_161', 'ch3list_tuples_162', 'ch3list_tuples_163', 'ch3list_tuples_164', 'ch3list_tuples_165', 'ch3list_tuples_166', 'ch3list_tuples_167', 'ch3list_tuples_168', 'ch3list_tuples_169', 'ch3list_tuples_16a', 'ch3list_tuples_16b', 'ch3list_tuples_16c', 'ch3list_tuples_16d', 'ch3list_tuples_16e', 'ch3list_tuples_16f', 'ch3list_tuples_170', 'ch3list_tuples_171', 'ch3list_tuples_172', 'ch3list_tuples_173', 'ch3list_tuples_174', 'ch3list_tuples_175', 'ch3list_tuples_176', 'ch3list_tuples_177', 'ch3list_tuples_178', 'ch3list_tuples_179', 'ch3list_tuples_17a', 'ch3list_tuples_17b', 'ch3list_tuples_17c', 'ch3list_tuples_17d', 'ch3list_tuples_17e', 'ch3list_tuples_17f', 'ch3list_tuples_180', 'ch3list_tuples_181', 'ch3list_tuples_182', 'ch3list_tuples_183', 'ch3list_tuples_184', 'ch3list_tuples_185', 'ch3list_tuples_186', 'ch3list_tuples_187', 'ch3list_tuples_188', 'ch3list_tuples_189', 'ch3list_tuples_18a', 'ch3list_tuples_18b', 'ch3list_tuples_18c', 'ch3list_tuples_18d', 'ch3list_tuples_18e', 'ch3list_tuples_18f', 'ch3list_tuples_190', 'ch3list_tuples_191', 'ch3list_tuples_192', 'ch3list_tuples_193', 'ch3list_tuples_194', 'ch3list_tuples_195', 'ch3list_tuples_196', 'ch3list_tuples_197', 'ch3list_tuples_198', 'ch3list_tuples_199', 'ch3list_tuples_19a', 'ch3list_tuples_19b', 'ch3list_tuples_19c', 'ch3list_tuples_19d', 'ch3list_tuples_19e', 'ch3list_tuples_19f', 'ch3list_tuples_1a0', 'ch3list_tuples_1a1', 'ch3list_tuples_1a2', 'ch3list_tuples_1a3', 'ch3list_tuples_1a4', 'ch3list_tuples_1a5', 'ch3list_tuples_1a6', 'ch3list_tuples_1a7', 'ch3list_tuples_1a8', 'ch3list_tuples_1a9', 'ch3list_tuples_1aa', 'ch3list_tuples_1ab', 'ch3list_tuples_1ac', 'ch3list_tuples_1ad', 'ch3list_tuples_1ae', 'ch3list_tuples_1af', 'ch3list_tuples_1b0', 'ch3list_tuples_1b1', 'ch3list_tuples_1b2', 'ch3list_tuples_1b3', 'ch3list_tuples_1b4', 'ch3list_tuples_1b5', 'ch3list_tuples_1b6', 'ch3list_tuples_1b7', 'ch3list_tuples_1b8', 'ch3list_tuples_1b9', 'ch3list_tuples_1ba', 'ch3list_tuples_1bb', 'ch3list_tuples_1bc', 'ch3list_tuples_1bd', 'ch3list_tuples_1be', 'ch3list_tuples_1bf', 'ch3list_tuples_1c0', 'ch3list_tuples_1c1', 'ch3list_tuples_1c2', 'ch3list_tuples_1c3', 'ch3list_tuples_1c4', 'ch3list_tuples_1c5', 'ch3list_tuples_1c6', 'ch3list_tuples_1c7', 'ch3list_tuples_1c8', 'ch3list_tuples_1c9', 'ch3list_tuples_1ca', 'ch3list_tuples_1cb', 'ch3list_tuples_1cc', 'ch3list_tuples_1cd', 'ch3list_tuples_1ce', 'ch3list_tuples_1cf', 'ch3list_tuples_1d0', 'ch3list_tuples_1d1', 'ch3list_tuples_1d2', 'ch3list_tuples_1d3', 'ch3list_tuples_1d4', 'ch3list_tuples_1d5', 'ch3list_tuples_1d6', 'ch3list_tuples_1d7', 'ch3list_tuples_1d8', 'ch3list_tuples_1d9', 'ch3list_tuples_1da', 'ch3list_tuples_1db', 'ch3list_tuples_1dc', 'ch3list_tuples_1dd', 'ch3list_tuples_1de', 'ch3list_tuples_1df', 'ch3list_tuples_1e0', 'ch3list_tuples_1e1'], 'ch2intro': ['ch2intro_1', 'ch2intro_2', 'ch2intro_3', 'ch2intro_4', 'ch2intro_5', 'ch2intro_6', 'ch2intro_7', 'ch2intro_8', 'ch2intro_9', 'ch2intro_a', 'ch2intro_b', 'ch2intro_c', 'ch2intro_d', 'ch2intro_e', 'ch2intro_f', 'ch2intro_10', 'ch2intro_11', 'ch2intro_12', 'ch2intro_13', 'ch2intro_14', 'ch2intro_15', 'ch2intro_16', 'ch2intro_17', 'ch2intro_18', 'ch2intro_19', 'ch2intro_1a', 'ch2intro_1b', 'ch2intro_1c', 'ch2intro_1d', 'ch2intro_1e', 'ch2intro_1f', 'ch2intro_20', 'ch2intro_21', 'ch2intro_22', 'ch2intro_23', 'ch2intro_24', 'ch2intro_25', 'ch2intro_26', 'ch2intro_27', 'ch2intro_28', 'ch2intro_29', 'ch2intro_2a', 'ch2intro_2b', 'ch2intro_2c', 'ch2intro_2d', 'ch2intro_2e', 'ch2intro_2f', 'ch2intro_30', 'ch2intro_31', 'ch2intro_32', 'ch2intro_33', 'ch2intro_34', 'ch2intro_35', 'ch2intro_36', 'ch2intro_37', 'ch2intro_38', 'ch2intro_39', 'ch2intro_3a', 'ch2intro_3b', 'ch2intro_3c', 'ch2intro_3d', 'ch2intro_3e', 'ch2intro_3f', 'ch2intro_40', 'ch2intro_41', 'ch2intro_42', 'ch2intro_43', 'ch2intro_44', 'ch2intro_45', 'ch2intro_46', 'ch2intro_47', 'ch2intro_48', 'ch2intro_49', 'ch2intro_4a', 'ch2intro_4b', 'ch2intro_4c', 'ch2intro_4d', 'ch2intro_4e', 'ch2intro_4f', 'ch2intro_50', 'ch2intro_51', 'ch2intro_52', 'ch2intro_53', 'ch2intro_54', 'ch2intro_55', 'ch2intro_56', 'ch2intro_57', 'ch2intro_58', 'ch2intro_59', 'ch2intro_5a', 'ch2intro_5b', 'ch2intro_5c', 'ch2intro_5d', 'ch2intro_5e', 'ch2intro_5f', 'ch2intro_60', 'ch2intro_61', 'ch2intro_62', 'ch2intro_63', 'ch2intro_64', 'ch2intro_65', 'ch2intro_66', 'ch2intro_67', 'ch2intro_68', 'ch2intro_69', 'ch2intro_6a', 'ch2intro_6b', 'ch2intro_6c', 'ch2intro_6d', 'ch2intro_6e', 'ch2intro_6f', 'ch2intro_70', 'ch2intro_71', 'ch2intro_72', 'ch2intro_73', 'ch2intro_74', 'ch2intro_75', 'ch2intro_76', 'ch2intro_77', 'ch2intro_78', 'ch2intro_79', 'ch2intro_7a', 'ch2intro_7b', 'ch2intro_7c', 'ch2intro_7d', 'ch2intro_7e', 'ch2intro_7f', 'ch2intro_80', 'ch2intro_81', 'ch2intro_82', 'ch2intro_83', 'ch2intro_84', 'ch2intro_85', 'ch2intro_86', 'ch2intro_87', 'ch2intro_88', 'ch2intro_89', 'ch2intro_8a', 'ch2intro_8b', 'ch2intro_8c', 'ch2intro_8d', 'ch2intro_8e', 'ch2intro_8f', 'ch2intro_90', 'ch2intro_91', 'ch2intro_92', 'ch2intro_93', 'ch2intro_94', 'ch2intro_95', 'ch2intro_96', 'ch2intro_97', 'ch2intro_98', 'ch2intro_99', 'ch2intro_9a', 'ch2intro_9b', 'ch2intro_9c', 'ch2intro_9d', 'ch2intro_9e', 'ch2intro_9f', 'ch2intro_a0', 'ch2intro_a1', 'ch2intro_a2', 'ch2intro_a3', 'ch2intro_a4', 'ch2intro_a5', 'ch2intro_a6', 'ch2intro_a7', 'ch2intro_a8', 'ch2intro_a9', 'ch2intro_aa', 'ch2intro_ab', 'ch2intro_ac', 'ch2intro_ad', 'ch2intro_ae', 'ch2intro_af', 'ch2intro_b0', 'ch2intro_b1', 'ch2intro_b2', 'ch2intro_b3', 'ch2intro_b4', 'ch2intro_b5', 'ch2intro_b6', 'ch2intro_b7', 'ch2intro_b8', 'ch2intro_b9', 'ch2intro_ba', 'ch2intro_bb', 'ch2intro_bc', 'ch2intro_bd', 'ch2intro_be', 'ch2intro_bf', 'ch2intro_c0', 'ch2intro_c1', 'ch2intro_c2', 'ch2intro_c3', 'ch2intro_c4', 'ch2intro_c5', 'ch2intro_c6', 'ch2intro_c7', 'ch2intro_c8', 'ch2intro_c9', 'ch2intro_ca', 'ch2intro_cb', 'ch2intro_cc', 'ch2intro_cd', 'ch2intro_ce', 'ch2intro_cf', 'ch2intro_d0', 'ch2intro_d1', 'ch2intro_d2', 'ch2intro_d3', 'ch2intro_d4', 'ch2intro_d5', 'ch2intro_d6', 'ch2intro_d7', 'ch2intro_d8', 'ch2intro_d9', 'ch2intro_da', 'ch2intro_db', 'ch2intro_dc', 'ch2intro_dd', 'ch2intro_de', 'ch2intro_df', 'ch2intro_e0', 'ch2intro_e1', 'ch2intro_e2', 'ch2intro_e3', 'ch2intro_e4', 'ch2intro_e5', 'ch2intro_e6', 'ch2intro_e7', 'ch2intro_e8', 'ch2intro_e9', 'ch2intro_ea', 'ch2intro_eb', 'ch2intro_ec', 'ch2intro_ed', 'ch2intro_ee', 'ch2intro_ef', 'ch2intro_f0', 'ch2intro_f1', 'ch2intro_f2', 'ch2intro_f3', 'ch2intro_f4', 'ch2intro_f5', 'ch2intro_f6', 'ch2intro_f7', 'ch2intro_f8', 'ch2intro_f9', 'ch2intro_fa', 'ch2intro_fb', 'ch2intro_fc', 'ch2intro_fd', 'ch2intro_fe', 'ch2intro_ff', 'ch2intro_100', 'ch2intro_101', 'ch2intro_102', 'ch2intro_103', 'ch2intro_104', 'ch2intro_105', 'ch2intro_106', 'ch2intro_107', 'ch2intro_108', 'ch2intro_109', 'ch2intro_10a', 'ch2intro_10b', 'ch2intro_10c', 'ch2intro_10d', 'ch2intro_10e', 'ch2intro_10f', 'ch2intro_110', 'ch2intro_111', 'ch2intro_112', 'ch2intro_113', 'ch2intro_114', 'ch2intro_115', 'ch2intro_116', 'ch2intro_117', 'ch2intro_118', 'ch2intro_119', 'ch2intro_11a', 'ch2intro_11b', 'ch2intro_11c', 'ch2intro_11d', 'ch2intro_11e', 'ch2intro_11f', 'ch2intro_120', 'ch2intro_121', 'ch2intro_122', 'ch2intro_123', 'ch2intro_124', 'ch2intro_125', 'ch2intro_126', 'ch2intro_127', 'ch2intro_128', 'ch2intro_129', 'ch2intro_12a', 'ch2intro_12b', 'ch2intro_12c', 'ch2intro_12d', 'ch2intro_12e', 'ch2intro_12f', 'ch2intro_130', 'ch2intro_131', 'ch2intro_132', 'ch2intro_133', 'ch2intro_134', 'ch2intro_135', 'ch2intro_136', 'ch2intro_137', 'ch2intro_138', 'ch2intro_139', 'ch2intro_13a', 'ch2intro_13b', 'ch2intro_13c', 'ch2intro_13d', 'ch2intro_13e', 'ch2intro_13f', 'ch2intro_140', 'ch2intro_141', 'ch2intro_142', 'ch2intro_143', 'ch2intro_144', 'ch2intro_145', 'ch2intro_146', 'ch2intro_147', 'ch2intro_148', 'ch2intro_149', 'ch2intro_14a', 'ch2intro_14b', 'ch2intro_14c', 'ch2intro_14d', 'ch2intro_14e', 'ch2intro_14f', 'ch2intro_150', 'ch2intro_151', 'ch2intro_152', 'ch2intro_153', 'ch2intro_154', 'ch2intro_155', 'ch2intro_156', 'ch2intro_157', 'ch2intro_158', 'ch2intro_159', 'ch2intro_15a', 'ch2intro_15b', 'ch2intro_15c', 'ch2intro_15d', 'ch2intro_15e', 'ch2intro_15f', 'ch2intro_160', 'ch2intro_161', 'ch2intro_162', 'ch2intro_163', 'ch2intro_164', 'ch2intro_165', 'ch2intro_166', 'ch2intro_167', 'ch2intro_168', 'ch2intro_169', 'ch2intro_16a', 'ch2intro_16b', 'ch2intro_16c', 'ch2intro_16d', 'ch2intro_16e', 'ch2intro_16f', 'ch2intro_170', 'ch2intro_171', 'ch2intro_172', 'ch2intro_173', 'ch2intro_174', 'ch2intro_175', 'ch2intro_176', 'ch2intro_177', 'ch2intro_178', 'ch2intro_179', 'ch2intro_17a', 'ch2intro_17b', 'ch2intro_17c', 'ch2intro_17d', 'ch2intro_17e', 'ch2intro_17f', 'ch2intro_180', 'ch2intro_181', 'ch2intro_182', 'ch2intro_183', 'ch2intro_184', 'ch2intro_185', 'ch2intro_186', 'ch2intro_187', 'ch2intro_188', 'ch2intro_189', 'ch2intro_18a', 'ch2intro_18b', 'ch2intro_18c', 'ch2intro_18d', 'ch2intro_18e', 'ch2intro_18f', 'ch2intro_190', 'ch2intro_191', 'ch2intro_192', 'ch2intro_193', 'ch2intro_194', 'ch2intro_195', 'ch2intro_196', 'ch2intro_197', 'ch2intro_198', 'ch2intro_199', 'ch2intro_19a', 'ch2intro_19b', 'ch2intro_19c', 'ch2intro_19d', 'ch2intro_19e', 'ch2intro_19f', 'ch2intro_1a0', 'ch2intro_1a1', 'ch2intro_1a2', 'ch2intro_1a3', 'ch2intro_1a4', 'ch2intro_1a5', 'ch2intro_1a6', 'ch2intro_1a7', 'ch2intro_1a8', 'ch2intro_1a9', 'ch2intro_1aa', 'ch2intro_1ab', 'ch2intro_1ac', 'ch2intro_1ad', 'ch2intro_1ae', 'ch2intro_1af', 'ch2intro_1b0', 'ch2intro_1b1', 'ch2intro_1b2', 'ch2intro_1b3', 'ch2intro_1b4', 'ch2intro_1b5', 'ch2intro_1b6', 'ch2intro_1b7', 'ch2intro_1b8', 'ch2intro_1b9', 'ch2intro_1ba', 'ch2intro_1bb', 'ch2intro_1bc', 'ch2intro_1bd', 'ch2intro_1be', 'ch2intro_1bf', 'ch2intro_1c0', 'ch2intro_1c1', 'ch2intro_1c2', 'ch2intro_1c3', 'ch2intro_1c4', 'ch2intro_1c5', 'ch2intro_1c6', 'ch2intro_1c7', 'ch2intro_1c8', 'ch2intro_1c9', 'ch2intro_1ca', 'ch2intro_1cb', 'ch2intro_1cc', 'ch2intro_1cd', 'ch2intro_1ce', 'ch2intro_1cf', 'ch2intro_1d0', 'ch2intro_1d1', 'ch2intro_1d2', 'ch2intro_1d3', 'ch2intro_1d4', 'ch2intro_1d5', 'ch2intro_1d6', 'ch2intro_1d7', 'ch2intro_1d8', 'ch2intro_1d9', 'ch2intro_1da', 'ch2intro_1db', 'ch2intro_1dc', 'ch2intro_1dd', 'ch2intro_1de', 'ch2intro_1df', 'ch2intro_1e0', 'ch2intro_1e1'], 'ch4strings_dicts': ['ch4strings_dicts_1', 'ch4strings_dicts_2', 'ch4strings_dicts_3', 'ch4strings_dicts_4', 'ch4strings_dicts_5', 'ch4strings_dicts_6', 'ch4strings_dicts_7', 'ch4strings_dicts_8', 'ch4strings_dicts_9', 'ch4strings_dicts_a', 'ch4strings_dicts_b', 'ch4strings_dicts_c', 'ch4strings_dicts_d', 'ch4strings_dicts_e', 'ch4strings_dicts_f', 'ch4strings_dicts_10', 'ch4strings_dicts_11', 'ch4strings_dicts_12', 'ch4strings_dicts_13', 'ch4strings_dicts_14', 'ch4strings_dicts_15', 'ch4strings_dicts_16', 'ch4strings_dicts_17', 'ch4strings_dicts_18', 'ch4strings_dicts_19', 'ch4strings_dicts_1a', 'ch4strings_dicts_1b', 'ch4strings_dicts_1c', 'ch4strings_dicts_1d', 'ch4strings_dicts_1e', 'ch4strings_dicts_1f', 'ch4strings_dicts_20', 'ch4strings_dicts_21', 'ch4strings_dicts_22', 'ch4strings_dicts_23', 'ch4strings_dicts_24', 'ch4strings_dicts_25', 'ch4strings_dicts_26', 'ch4strings_dicts_27', 'ch4strings_dicts_28', 'ch4strings_dicts_29', 'ch4strings_dicts_2a', 'ch4strings_dicts_2b', 'ch4strings_dicts_2c', 'ch4strings_dicts_2d', 'ch4strings_dicts_2e', 'ch4strings_dicts_2f', 'ch4strings_dicts_30', 'ch4strings_dicts_31', 'ch4strings_dicts_32', 'ch4strings_dicts_33', 'ch4strings_dicts_34', 'ch4strings_dicts_35', 'ch4strings_dicts_36', 'ch4strings_dicts_37', 'ch4strings_dicts_38', 'ch4strings_dicts_39', 'ch4strings_dicts_3a', 'ch4strings_dicts_3b', 'ch4strings_dicts_3c', 'ch4strings_dicts_3d', 'ch4strings_dicts_3e', 'ch4strings_dicts_3f', 'ch4strings_dicts_40', 'ch4strings_dicts_41', 'ch4strings_dicts_42', 'ch4strings_dicts_43', 'ch4strings_dicts_44', 'ch4strings_dicts_45', 'ch4strings_dicts_46', 'ch4strings_dicts_47', 'ch4strings_dicts_48', 'ch4strings_dicts_49', 'ch4strings_dicts_4a', 'ch4strings_dicts_4b', 'ch4strings_dicts_4c', 'ch4strings_dicts_4d', 'ch4strings_dicts_4e', 'ch4strings_dicts_4f', 'ch4strings_dicts_50', 'ch4strings_dicts_51', 'ch4strings_dicts_52', 'ch4strings_dicts_53', 'ch4strings_dicts_54', 'ch4strings_dicts_55', 'ch4strings_dicts_56', 'ch4strings_dicts_57', 'ch4strings_dicts_58', 'ch4strings_dicts_59', 'ch4strings_dicts_5a', 'ch4strings_dicts_5b', 'ch4strings_dicts_5c', 'ch4strings_dicts_5d', 'ch4strings_dicts_5e', 'ch4strings_dicts_5f', 'ch4strings_dicts_60', 'ch4strings_dicts_61', 'ch4strings_dicts_62', 'ch4strings_dicts_63', 'ch4strings_dicts_64', 'ch4strings_dicts_65', 'ch4strings_dicts_66', 'ch4strings_dicts_67', 'ch4strings_dicts_68', 'ch4strings_dicts_69', 'ch4strings_dicts_6a', 'ch4strings_dicts_6b', 'ch4strings_dicts_6c', 'ch4strings_dicts_6d', 'ch4strings_dicts_6e', 'ch4strings_dicts_6f', 'ch4strings_dicts_70', 'ch4strings_dicts_71', 'ch4strings_dicts_72', 'ch4strings_dicts_73', 'ch4strings_dicts_74', 'ch4strings_dicts_75', 'ch4strings_dicts_76', 'ch4strings_dicts_77', 'ch4strings_dicts_78', 'ch4strings_dicts_79', 'ch4strings_dicts_7a', 'ch4strings_dicts_7b', 'ch4strings_dicts_7c', 'ch4strings_dicts_7d', 'ch4strings_dicts_7e', 'ch4strings_dicts_7f', 'ch4strings_dicts_80', 'ch4strings_dicts_81', 'ch4strings_dicts_82', 'ch4strings_dicts_83', 'ch4strings_dicts_84', 'ch4strings_dicts_85', 'ch4strings_dicts_86', 'ch4strings_dicts_87', 'ch4strings_dicts_88', 'ch4strings_dicts_89', 'ch4strings_dicts_8a', 'ch4strings_dicts_8b', 'ch4strings_dicts_8c', 'ch4strings_dicts_8d', 'ch4strings_dicts_8e', 'ch4strings_dicts_8f', 'ch4strings_dicts_90', 'ch4strings_dicts_91', 'ch4strings_dicts_92', 'ch4strings_dicts_93', 'ch4strings_dicts_94', 'ch4strings_dicts_95', 'ch4strings_dicts_96', 'ch4strings_dicts_97', 'ch4strings_dicts_98', 'ch4strings_dicts_99', 'ch4strings_dicts_9a', 'ch4strings_dicts_9b', 'ch4strings_dicts_9c', 'ch4strings_dicts_9d', 'ch4strings_dicts_9e', 'ch4strings_dicts_9f', 'ch4strings_dicts_a0', 'ch4strings_dicts_a1', 'ch4strings_dicts_a2', 'ch4strings_dicts_a3', 'ch4strings_dicts_a4', 'ch4strings_dicts_a5', 'ch4strings_dicts_a6', 'ch4strings_dicts_a7', 'ch4strings_dicts_a8', 'ch4strings_dicts_a9', 'ch4strings_dicts_aa', 'ch4strings_dicts_ab', 'ch4strings_dicts_ac', 'ch4strings_dicts_ad', 'ch4strings_dicts_ae', 'ch4strings_dicts_af', 'ch4strings_dicts_b0', 'ch4strings_dicts_b1', 'ch4strings_dicts_b2', 'ch4strings_dicts_b3', 'ch4strings_dicts_b4', 'ch4strings_dicts_b5', 'ch4strings_dicts_b6', 'ch4strings_dicts_b7', 'ch4strings_dicts_b8', 'ch4strings_dicts_b9', 'ch4strings_dicts_ba', 'ch4strings_dicts_bb', 'ch4strings_dicts_bc', 'ch4strings_dicts_bd', 'ch4strings_dicts_be', 'ch4strings_dicts_bf', 'ch4strings_dicts_c0', 'ch4strings_dicts_c1', 'ch4strings_dicts_c2', 'ch4strings_dicts_c3', 'ch4strings_dicts_c4', 'ch4strings_dicts_c5', 'ch4strings_dicts_c6', 'ch4strings_dicts_c7', 'ch4strings_dicts_c8', 'ch4strings_dicts_c9', 'ch4strings_dicts_ca', 'ch4strings_dicts_cb', 'ch4strings_dicts_cc', 'ch4strings_dicts_cd', 'ch4strings_dicts_ce', 'ch4strings_dicts_cf', 'ch4strings_dicts_d0', 'ch4strings_dicts_d1', 'ch4strings_dicts_d2', 'ch4strings_dicts_d3', 'ch4strings_dicts_d4', 'ch4strings_dicts_d5', 'ch4strings_dicts_d6', 'ch4strings_dicts_d7', 'ch4strings_dicts_d8', 'ch4strings_dicts_d9', 'ch4strings_dicts_da', 'ch4strings_dicts_db', 'ch4strings_dicts_dc', 'ch4strings_dicts_dd', 'ch4strings_dicts_de', 'ch4strings_dicts_df', 'ch4strings_dicts_e0', 'ch4strings_dicts_e1', 'ch4strings_dicts_e2', 'ch4strings_dicts_e3', 'ch4strings_dicts_e4', 'ch4strings_dicts_e5', 'ch4strings_dicts_e6', 'ch4strings_dicts_e7', 'ch4strings_dicts_e8', 'ch4strings_dicts_e9', 'ch4strings_dicts_ea', 'ch4strings_dicts_eb', 'ch4strings_dicts_ec', 'ch4strings_dicts_ed', 'ch4strings_dicts_ee', 'ch4strings_dicts_ef', 'ch4strings_dicts_f0', 'ch4strings_dicts_f1', 'ch4strings_dicts_f2', 'ch4strings_dicts_f3', 'ch4strings_dicts_f4', 'ch4strings_dicts_f5', 'ch4strings_dicts_f6', 'ch4strings_dicts_f7', 'ch4strings_dicts_f8', 'ch4strings_dicts_f9', 'ch4strings_dicts_fa', 'ch4strings_dicts_fb', 'ch4strings_dicts_fc', 'ch4strings_dicts_fd', 'ch4strings_dicts_fe', 'ch4strings_dicts_ff', 'ch4strings_dicts_100', 'ch4strings_dicts_101', 'ch4strings_dicts_102', 'ch4strings_dicts_103', 'ch4strings_dicts_104', 'ch4strings_dicts_105', 'ch4strings_dicts_106', 'ch4strings_dicts_107', 'ch4strings_dicts_108', 'ch4strings_dicts_109', 'ch4strings_dicts_10a', 'ch4strings_dicts_10b', 'ch4strings_dicts_10c', 'ch4strings_dicts_10d', 'ch4strings_dicts_10e', 'ch4strings_dicts_10f', 'ch4strings_dicts_110', 'ch4strings_dicts_111', 'ch4strings_dicts_112', 'ch4strings_dicts_113', 'ch4strings_dicts_114', 'ch4strings_dicts_115', 'ch4strings_dicts_116', 'ch4strings_dicts_117', 'ch4strings_dicts_118', 'ch4strings_dicts_119', 'ch4strings_dicts_11a', 'ch4strings_dicts_11b', 'ch4strings_dicts_11c', 'ch4strings_dicts_11d', 'ch4strings_dicts_11e', 'ch4strings_dicts_11f', 'ch4strings_dicts_120', 'ch4strings_dicts_121', 'ch4strings_dicts_122', 'ch4strings_dicts_123', 'ch4strings_dicts_124', 'ch4strings_dicts_125', 'ch4strings_dicts_126', 'ch4strings_dicts_127', 'ch4strings_dicts_128', 'ch4strings_dicts_129', 'ch4strings_dicts_12a', 'ch4strings_dicts_12b', 'ch4strings_dicts_12c', 'ch4strings_dicts_12d', 'ch4strings_dicts_12e', 'ch4strings_dicts_12f', 'ch4strings_dicts_130', 'ch4strings_dicts_131', 'ch4strings_dicts_132', 'ch4strings_dicts_133', 'ch4strings_dicts_134', 'ch4strings_dicts_135', 'ch4strings_dicts_136', 'ch4strings_dicts_137', 'ch4strings_dicts_138', 'ch4strings_dicts_139', 'ch4strings_dicts_13a', 'ch4strings_dicts_13b', 'ch4strings_dicts_13c', 'ch4strings_dicts_13d', 'ch4strings_dicts_13e', 'ch4strings_dicts_13f', 'ch4strings_dicts_140', 'ch4strings_dicts_141', 'ch4strings_dicts_142', 'ch4strings_dicts_143', 'ch4strings_dicts_144', 'ch4strings_dicts_145', 'ch4strings_dicts_146', 'ch4strings_dicts_147', 'ch4strings_dicts_148', 'ch4strings_dicts_149', 'ch4strings_dicts_14a', 'ch4strings_dicts_14b', 'ch4strings_dicts_14c', 'ch4strings_dicts_14d', 'ch4strings_dicts_14e', 'ch4strings_dicts_14f', 'ch4strings_dicts_150', 'ch4strings_dicts_151', 'ch4strings_dicts_152', 'ch4strings_dicts_153', 'ch4strings_dicts_154', 'ch4strings_dicts_155', 'ch4strings_dicts_156', 'ch4strings_dicts_157', 'ch4strings_dicts_158', 'ch4strings_dicts_159', 'ch4strings_dicts_15a', 'ch4strings_dicts_15b', 'ch4strings_dicts_15c', 'ch4strings_dicts_15d', 'ch4strings_dicts_15e', 'ch4strings_dicts_15f', 'ch4strings_dicts_160', 'ch4strings_dicts_161', 'ch4strings_dicts_162', 'ch4strings_dicts_163', 'ch4strings_dicts_164', 'ch4strings_dicts_165', 'ch4strings_dicts_166', 'ch4strings_dicts_167', 'ch4strings_dicts_168', 'ch4strings_dicts_169', 'ch4strings_dicts_16a', 'ch4strings_dicts_16b', 'ch4strings_dicts_16c', 'ch4strings_dicts_16d', 'ch4strings_dicts_16e', 'ch4strings_dicts_16f', 'ch4strings_dicts_170', 'ch4strings_dicts_171', 'ch4strings_dicts_172', 'ch4strings_dicts_173', 'ch4strings_dicts_174', 'ch4strings_dicts_175', 'ch4strings_dicts_176', 'ch4strings_dicts_177', 'ch4strings_dicts_178', 'ch4strings_dicts_179', 'ch4strings_dicts_17a', 'ch4strings_dicts_17b', 'ch4strings_dicts_17c', 'ch4strings_dicts_17d', 'ch4strings_dicts_17e', 'ch4strings_dicts_17f', 'ch4strings_dicts_180', 'ch4strings_dicts_181', 'ch4strings_dicts_182', 'ch4strings_dicts_183', 'ch4strings_dicts_184', 'ch4strings_dicts_185', 'ch4strings_dicts_186', 'ch4strings_dicts_187', 'ch4strings_dicts_188', 'ch4strings_dicts_189', 'ch4strings_dicts_18a', 'ch4strings_dicts_18b', 'ch4strings_dicts_18c', 'ch4strings_dicts_18d', 'ch4strings_dicts_18e', 'ch4strings_dicts_18f', 'ch4strings_dicts_190', 'ch4strings_dicts_191', 'ch4strings_dicts_192', 'ch4strings_dicts_193', 'ch4strings_dicts_194', 'ch4strings_dicts_195', 'ch4strings_dicts_196', 'ch4strings_dicts_197', 'ch4strings_dicts_198', 'ch4strings_dicts_199', 'ch4strings_dicts_19a', 'ch4strings_dicts_19b', 'ch4strings_dicts_19c', 'ch4strings_dicts_19d', 'ch4strings_dicts_19e', 'ch4strings_dicts_19f', 'ch4strings_dicts_1a0', 'ch4strings_dicts_1a1', 'ch4strings_dicts_1a2', 'ch4strings_dicts_1a3', 'ch4strings_dicts_1a4', 'ch4strings_dicts_1a5', 'ch4strings_dicts_1a6', 'ch4strings_dicts_1a7', 'ch4strings_dicts_1a8', 'ch4strings_dicts_1a9', 'ch4strings_dicts_1aa', 'ch4strings_dicts_1ab', 'ch4strings_dicts_1ac', 'ch4strings_dicts_1ad', 'ch4strings_dicts_1ae', 'ch4strings_dicts_1af', 'ch4strings_dicts_1b0', 'ch4strings_dicts_1b1', 'ch4strings_dicts_1b2', 'ch4strings_dicts_1b3', 'ch4strings_dicts_1b4', 'ch4strings_dicts_1b5', 'ch4strings_dicts_1b6', 'ch4strings_dicts_1b7', 'ch4strings_dicts_1b8', 'ch4strings_dicts_1b9', 'ch4strings_dicts_1ba', 'ch4strings_dicts_1bb', 'ch4strings_dicts_1bc', 'ch4strings_dicts_1bd', 'ch4strings_dicts_1be', 'ch4strings_dicts_1bf', 'ch4strings_dicts_1c0', 'ch4strings_dicts_1c1', 'ch4strings_dicts_1c2', 'ch4strings_dicts_1c3', 'ch4strings_dicts_1c4', 'ch4strings_dicts_1c5', 'ch4strings_dicts_1c6', 'ch4strings_dicts_1c7', 'ch4strings_dicts_1c8', 'ch4strings_dicts_1c9', 'ch4strings_dicts_1ca', 'ch4strings_dicts_1cb', 'ch4strings_dicts_1cc', 'ch4strings_dicts_1cd', 'ch4strings_dicts_1ce', 'ch4strings_dicts_1cf', 'ch4strings_dicts_1d0', 'ch4strings_dicts_1d1', 'ch4strings_dicts_1d2', 'ch4strings_dicts_1d3', 'ch4strings_dicts_1d4', 'ch4strings_dicts_1d5', 'ch4strings_dicts_1d6', 'ch4strings_dicts_1d7', 'ch4strings_dicts_1d8', 'ch4strings_dicts_1d9', 'ch4strings_dicts_1da', 'ch4strings_dicts_1db', 'ch4strings_dicts_1dc', 'ch4strings_dicts_1dd', 'ch4strings_dicts_1de', 'ch4strings_dicts_1df', 'ch4strings_dicts_1e0', 'ch4strings_dicts_1e1'], 'ch11Section_5': ['ch11Section_5_1', 'ch11Section_5_2', 'ch11Section_5_3', 'ch11Section_5_4', 'ch11Section_5_5', 'ch11Section_5_6', 'ch11Section_5_7', 'ch11Section_5_8', 'ch11Section_5_9', 'ch11Section_5_a', 'ch11Section_5_b', 'ch11Section_5_c', 'ch11Section_5_d', 'ch11Section_5_e', 'ch11Section_5_f', 'ch11Section_5_10', 'ch11Section_5_11', 'ch11Section_5_12', 'ch11Section_5_13', 'ch11Section_5_14', 'ch11Section_5_15', 'ch11Section_5_16', 'ch11Section_5_17', 'ch11Section_5_18', 'ch11Section_5_19', 'ch11Section_5_1a', 'ch11Section_5_1b', 'ch11Section_5_1c', 'ch11Section_5_1d', 'ch11Section_5_1e', 'ch11Section_5_1f', 'ch11Section_5_20', 'ch11Section_5_21', 'ch11Section_5_22', 'ch11Section_5_23', 'ch11Section_5_24', 'ch11Section_5_25', 'ch11Section_5_26', 'ch11Section_5_27', 'ch11Section_5_28', 'ch11Section_5_29', 'ch11Section_5_2a', 'ch11Section_5_2b', 'ch11Section_5_2c', 'ch11Section_5_2d', 'ch11Section_5_2e', 'ch11Section_5_2f', 'ch11Section_5_30', 'ch11Section_5_31', 'ch11Section_5_32', 'ch11Section_5_33', 'ch11Section_5_34', 'ch11Section_5_35', 'ch11Section_5_36', 'ch11Section_5_37', 'ch11Section_5_38', 'ch11Section_5_39', 'ch11Section_5_3a', 'ch11Section_5_3b', 'ch11Section_5_3c', 'ch11Section_5_3d', 'ch11Section_5_3e', 'ch11Section_5_3f', 'ch11Section_5_40', 'ch11Section_5_41', 'ch11Section_5_42', 'ch11Section_5_43', 'ch11Section_5_44', 'ch11Section_5_45', 'ch11Section_5_46', 'ch11Section_5_47', 'ch11Section_5_48', 'ch11Section_5_49', 'ch11Section_5_4a', 'ch11Section_5_4b', 'ch11Section_5_4c', 'ch11Section_5_4d', 'ch11Section_5_4e', 'ch11Section_5_4f', 'ch11Section_5_50', 'ch11Section_5_51', 'ch11Section_5_52', 'ch11Section_5_53', 'ch11Section_5_54', 'ch11Section_5_55', 'ch11Section_5_56', 'ch11Section_5_57', 'ch11Section_5_58', 'ch11Section_5_59', 'ch11Section_5_5a', 'ch11Section_5_5b', 'ch11Section_5_5c', 'ch11Section_5_5d', 'ch11Section_5_5e', 'ch11Section_5_5f', 'ch11Section_5_60', 'ch11Section_5_61', 'ch11Section_5_62', 'ch11Section_5_63', 'ch11Section_5_64', 'ch11Section_5_65', 'ch11Section_5_66', 'ch11Section_5_67', 'ch11Section_5_68', 'ch11Section_5_69', 'ch11Section_5_6a', 'ch11Section_5_6b', 'ch11Section_5_6c', 'ch11Section_5_6d', 'ch11Section_5_6e', 'ch11Section_5_6f', 'ch11Section_5_70', 'ch11Section_5_71', 'ch11Section_5_72', 'ch11Section_5_73', 'ch11Section_5_74', 'ch11Section_5_75', 'ch11Section_5_76', 'ch11Section_5_77', 'ch11Section_5_78', 'ch11Section_5_79', 'ch11Section_5_7a', 'ch11Section_5_7b', 'ch11Section_5_7c', 'ch11Section_5_7d', 'ch11Section_5_7e', 'ch11Section_5_7f', 'ch11Section_5_80', 'ch11Section_5_81', 'ch11Section_5_82', 'ch11Section_5_83', 'ch11Section_5_84', 'ch11Section_5_85', 'ch11Section_5_86', 'ch11Section_5_87', 'ch11Section_5_88', 'ch11Section_5_89', 'ch11Section_5_8a', 'ch11Section_5_8b', 'ch11Section_5_8c', 'ch11Section_5_8d', 'ch11Section_5_8e', 'ch11Section_5_8f', 'ch11Section_5_90', 'ch11Section_5_91', 'ch11Section_5_92', 'ch11Section_5_93', 'ch11Section_5_94', 'ch11Section_5_95', 'ch11Section_5_96', 'ch11Section_5_97', 'ch11Section_5_98', 'ch11Section_5_99', 'ch11Section_5_9a', 'ch11Section_5_9b', 'ch11Section_5_9c', 'ch11Section_5_9d', 'ch11Section_5_9e', 'ch11Section_5_9f', 'ch11Section_5_a0', 'ch11Section_5_a1', 'ch11Section_5_a2', 'ch11Section_5_a3', 'ch11Section_5_a4', 'ch11Section_5_a5', 'ch11Section_5_a6', 'ch11Section_5_a7', 'ch11Section_5_a8', 'ch11Section_5_a9', 'ch11Section_5_aa', 'ch11Section_5_ab', 'ch11Section_5_ac', 'ch11Section_5_ad', 'ch11Section_5_ae', 'ch11Section_5_af', 'ch11Section_5_b0', 'ch11Section_5_b1', 'ch11Section_5_b2', 'ch11Section_5_b3', 'ch11Section_5_b4', 'ch11Section_5_b5', 'ch11Section_5_b6', 'ch11Section_5_b7', 'ch11Section_5_b8', 'ch11Section_5_b9', 'ch11Section_5_ba', 'ch11Section_5_bb', 'ch11Section_5_bc', 'ch11Section_5_bd', 'ch11Section_5_be', 'ch11Section_5_bf', 'ch11Section_5_c0', 'ch11Section_5_c1', 'ch11Section_5_c2', 'ch11Section_5_c3', 'ch11Section_5_c4', 'ch11Section_5_c5', 'ch11Section_5_c6', 'ch11Section_5_c7', 'ch11Section_5_c8', 'ch11Section_5_c9', 'ch11Section_5_ca', 'ch11Section_5_cb', 'ch11Section_5_cc', 'ch11Section_5_cd', 'ch11Section_5_ce', 'ch11Section_5_cf', 'ch11Section_5_d0', 'ch11Section_5_d1', 'ch11Section_5_d2', 'ch11Section_5_d3', 'ch11Section_5_d4', 'ch11Section_5_d5', 'ch11Section_5_d6', 'ch11Section_5_d7', 'ch11Section_5_d8', 'ch11Section_5_d9', 'ch11Section_5_da', 'ch11Section_5_db', 'ch11Section_5_dc', 'ch11Section_5_dd', 'ch11Section_5_de', 'ch11Section_5_df', 'ch11Section_5_e0', 'ch11Section_5_e1', 'ch11Section_5_e2', 'ch11Section_5_e3', 'ch11Section_5_e4', 'ch11Section_5_e5', 'ch11Section_5_e6', 'ch11Section_5_e7', 'ch11Section_5_e8', 'ch11Section_5_e9', 'ch11Section_5_ea', 'ch11Section_5_eb', 'ch11Section_5_ec', 'ch11Section_5_ed', 'ch11Section_5_ee', 'ch11Section_5_ef', 'ch11Section_5_f0', 'ch11Section_5_f1', 'ch11Section_5_f2', 'ch11Section_5_f3', 'ch11Section_5_f4', 'ch11Section_5_f5', 'ch11Section_5_f6', 'ch11Section_5_f7', 'ch11Section_5_f8', 'ch11Section_5_f9', 'ch11Section_5_fa', 'ch11Section_5_fb', 'ch11Section_5_fc', 'ch11Section_5_fd', 'ch11Section_5_fe', 'ch11Section_5_ff', 'ch11Section_5_100', 'ch11Section_5_101', 'ch11Section_5_102', 'ch11Section_5_103', 'ch11Section_5_104', 'ch11Section_5_105', 'ch11Section_5_106', 'ch11Section_5_107', 'ch11Section_5_108', 'ch11Section_5_109', 'ch11Section_5_10a', 'ch11Section_5_10b', 'ch11Section_5_10c', 'ch11Section_5_10d', 'ch11Section_5_10e', 'ch11Section_5_10f', 'ch11Section_5_110', 'ch11Section_5_111', 'ch11Section_5_112', 'ch11Section_5_113', 'ch11Section_5_114', 'ch11Section_5_115', 'ch11Section_5_116', 'ch11Section_5_117', 'ch11Section_5_118', 'ch11Section_5_119', 'ch11Section_5_11a', 'ch11Section_5_11b', 'ch11Section_5_11c', 'ch11Section_5_11d', 'ch11Section_5_11e', 'ch11Section_5_11f', 'ch11Section_5_120', 'ch11Section_5_121', 'ch11Section_5_122', 'ch11Section_5_123', 'ch11Section_5_124', 'ch11Section_5_125', 'ch11Section_5_126', 'ch11Section_5_127', 'ch11Section_5_128', 'ch11Section_5_129', 'ch11Section_5_12a', 'ch11Section_5_12b', 'ch11Section_5_12c', 'ch11Section_5_12d', 'ch11Section_5_12e', 'ch11Section_5_12f', 'ch11Section_5_130', 'ch11Section_5_131', 'ch11Section_5_132', 'ch11Section_5_133', 'ch11Section_5_134', 'ch11Section_5_135', 'ch11Section_5_136', 'ch11Section_5_137', 'ch11Section_5_138', 'ch11Section_5_139', 'ch11Section_5_13a', 'ch11Section_5_13b', 'ch11Section_5_13c', 'ch11Section_5_13d', 'ch11Section_5_13e', 'ch11Section_5_13f', 'ch11Section_5_140', 'ch11Section_5_141', 'ch11Section_5_142', 'ch11Section_5_143', 'ch11Section_5_144', 'ch11Section_5_145', 'ch11Section_5_146', 'ch11Section_5_147', 'ch11Section_5_148', 'ch11Section_5_149', 'ch11Section_5_14a', 'ch11Section_5_14b', 'ch11Section_5_14c', 'ch11Section_5_14d', 'ch11Section_5_14e', 'ch11Section_5_14f', 'ch11Section_5_150', 'ch11Section_5_151', 'ch11Section_5_152', 'ch11Section_5_153', 'ch11Section_5_154', 'ch11Section_5_155', 'ch11Section_5_156', 'ch11Section_5_157', 'ch11Section_5_158', 'ch11Section_5_159', 'ch11Section_5_15a', 'ch11Section_5_15b', 'ch11Section_5_15c', 'ch11Section_5_15d', 'ch11Section_5_15e', 'ch11Section_5_15f', 'ch11Section_5_160', 'ch11Section_5_161', 'ch11Section_5_162', 'ch11Section_5_163', 'ch11Section_5_164', 'ch11Section_5_165', 'ch11Section_5_166', 'ch11Section_5_167', 'ch11Section_5_168', 'ch11Section_5_169', 'ch11Section_5_16a', 'ch11Section_5_16b', 'ch11Section_5_16c', 'ch11Section_5_16d', 'ch11Section_5_16e', 'ch11Section_5_16f', 'ch11Section_5_170', 'ch11Section_5_171', 'ch11Section_5_172', 'ch11Section_5_173', 'ch11Section_5_174', 'ch11Section_5_175', 'ch11Section_5_176', 'ch11Section_5_177', 'ch11Section_5_178', 'ch11Section_5_179', 'ch11Section_5_17a', 'ch11Section_5_17b', 'ch11Section_5_17c', 'ch11Section_5_17d', 'ch11Section_5_17e', 'ch11Section_5_17f', 'ch11Section_5_180', 'ch11Section_5_181', 'ch11Section_5_182', 'ch11Section_5_183', 'ch11Section_5_184', 'ch11Section_5_185', 'ch11Section_5_186', 'ch11Section_5_187', 'ch11Section_5_188', 'ch11Section_5_189', 'ch11Section_5_18a', 'ch11Section_5_18b', 'ch11Section_5_18c', 'ch11Section_5_18d', 'ch11Section_5_18e', 'ch11Section_5_18f', 'ch11Section_5_190', 'ch11Section_5_191', 'ch11Section_5_192', 'ch11Section_5_193', 'ch11Section_5_194', 'ch11Section_5_195', 'ch11Section_5_196', 'ch11Section_5_197', 'ch11Section_5_198', 'ch11Section_5_199', 'ch11Section_5_19a', 'ch11Section_5_19b', 'ch11Section_5_19c', 'ch11Section_5_19d', 'ch11Section_5_19e', 'ch11Section_5_19f', 'ch11Section_5_1a0', 'ch11Section_5_1a1', 'ch11Section_5_1a2', 'ch11Section_5_1a3', 'ch11Section_5_1a4', 'ch11Section_5_1a5', 'ch11Section_5_1a6', 'ch11Section_5_1a7', 'ch11Section_5_1a8', 'ch11Section_5_1a9', 'ch11Section_5_1aa', 'ch11Section_5_1ab', 'ch11Section_5_1ac', 'ch11Section_5_1ad', 'ch11Section_5_1ae', 'ch11Section_5_1af', 'ch11Section_5_1b0', 'ch11Section_5_1b1', 'ch11Section_5_1b2', 'ch11Section_5_1b3', 'ch11Section_5_1b4', 'ch11Section_5_1b5', 'ch11Section_5_1b6', 'ch11Section_5_1b7', 'ch11Section_5_1b8', 'ch11Section_5_1b9', 'ch11Section_5_1ba', 'ch11Section_5_1bb', 'ch11Section_5_1bc', 'ch11Section_5_1bd', 'ch11Section_5_1be', 'ch11Section_5_1bf', 'ch11Section_5_1c0', 'ch11Section_5_1c1', 'ch11Section_5_1c2', 'ch11Section_5_1c3', 'ch11Section_5_1c4', 'ch11Section_5_1c5', 'ch11Section_5_1c6', 'ch11Section_5_1c7', 'ch11Section_5_1c8', 'ch11Section_5_1c9', 'ch11Section_5_1ca', 'ch11Section_5_1cb', 'ch11Section_5_1cc', 'ch11Section_5_1cd', 'ch11Section_5_1ce', 'ch11Section_5_1cf', 'ch11Section_5_1d0', 'ch11Section_5_1d1', 'ch11Section_5_1d2', 'ch11Section_5_1d3', 'ch11Section_5_1d4', 'ch11Section_5_1d5', 'ch11Section_5_1d6', 'ch11Section_5_1d7', 'ch11Section_5_1d8', 'ch11Section_5_1d9', 'ch11Section_5_1da', 'ch11Section_5_1db', 'ch11Section_5_1dc', 'ch11Section_5_1dd', 'ch11Section_5_1de', 'ch11Section_5_1df', 'ch11Section_5_1e0', 'ch11Section_5_1e1'], 'ch7interim_assessment': ['ch7interim_assessment_1', 'ch7interim_assessment_2', 'ch7interim_assessment_3', 'ch7interim_assessment_4', 'ch7interim_assessment_5', 'ch7interim_assessment_6', 'ch7interim_assessment_7', 'ch7interim_assessment_8', 'ch7interim_assessment_9', 'ch7interim_assessment_a', 'ch7interim_assessment_b', 'ch7interim_assessment_c', 'ch7interim_assessment_d', 'ch7interim_assessment_e', 'ch7interim_assessment_f', 'ch7interim_assessment_10', 'ch7interim_assessment_11', 'ch7interim_assessment_12', 'ch7interim_assessment_13', 'ch7interim_assessment_14', 'ch7interim_assessment_15', 'ch7interim_assessment_16', 'ch7interim_assessment_17', 'ch7interim_assessment_18', 'ch7interim_assessment_19', 'ch7interim_assessment_1a', 'ch7interim_assessment_1b', 'ch7interim_assessment_1c', 'ch7interim_assessment_1d', 'ch7interim_assessment_1e', 'ch7interim_assessment_1f', 'ch7interim_assessment_20', 'ch7interim_assessment_21', 'ch7interim_assessment_22', 'ch7interim_assessment_23', 'ch7interim_assessment_24', 'ch7interim_assessment_25', 'ch7interim_assessment_26', 'ch7interim_assessment_27', 'ch7interim_assessment_28', 'ch7interim_assessment_29', 'ch7interim_assessment_2a', 'ch7interim_assessment_2b', 'ch7interim_assessment_2c', 'ch7interim_assessment_2d', 'ch7interim_assessment_2e', 'ch7interim_assessment_2f', 'ch7interim_assessment_30', 'ch7interim_assessment_31', 'ch7interim_assessment_32', 'ch7interim_assessment_33', 'ch7interim_assessment_34', 'ch7interim_assessment_35', 'ch7interim_assessment_36', 'ch7interim_assessment_37', 'ch7interim_assessment_38', 'ch7interim_assessment_39', 'ch7interim_assessment_3a', 'ch7interim_assessment_3b', 'ch7interim_assessment_3c', 'ch7interim_assessment_3d', 'ch7interim_assessment_3e', 'ch7interim_assessment_3f', 'ch7interim_assessment_40', 'ch7interim_assessment_41', 'ch7interim_assessment_42', 'ch7interim_assessment_43', 'ch7interim_assessment_44', 'ch7interim_assessment_45', 'ch7interim_assessment_46', 'ch7interim_assessment_47', 'ch7interim_assessment_48', 'ch7interim_assessment_49', 'ch7interim_assessment_4a', 'ch7interim_assessment_4b', 'ch7interim_assessment_4c', 'ch7interim_assessment_4d', 'ch7interim_assessment_4e', 'ch7interim_assessment_4f', 'ch7interim_assessment_50', 'ch7interim_assessment_51', 'ch7interim_assessment_52', 'ch7interim_assessment_53', 'ch7interim_assessment_54', 'ch7interim_assessment_55', 'ch7interim_assessment_56', 'ch7interim_assessment_57', 'ch7interim_assessment_58', 'ch7interim_assessment_59', 'ch7interim_assessment_5a', 'ch7interim_assessment_5b', 'ch7interim_assessment_5c', 'ch7interim_assessment_5d', 'ch7interim_assessment_5e', 'ch7interim_assessment_5f', 'ch7interim_assessment_60', 'ch7interim_assessment_61', 'ch7interim_assessment_62', 'ch7interim_assessment_63', 'ch7interim_assessment_64', 'ch7interim_assessment_65', 'ch7interim_assessment_66', 'ch7interim_assessment_67', 'ch7interim_assessment_68', 'ch7interim_assessment_69', 'ch7interim_assessment_6a', 'ch7interim_assessment_6b', 'ch7interim_assessment_6c', 'ch7interim_assessment_6d', 'ch7interim_assessment_6e', 'ch7interim_assessment_6f', 'ch7interim_assessment_70', 'ch7interim_assessment_71', 'ch7interim_assessment_72', 'ch7interim_assessment_73', 'ch7interim_assessment_74', 'ch7interim_assessment_75', 'ch7interim_assessment_76', 'ch7interim_assessment_77', 'ch7interim_assessment_78', 'ch7interim_assessment_79', 'ch7interim_assessment_7a', 'ch7interim_assessment_7b', 'ch7interim_assessment_7c', 'ch7interim_assessment_7d', 'ch7interim_assessment_7e', 'ch7interim_assessment_7f', 'ch7interim_assessment_80', 'ch7interim_assessment_81', 'ch7interim_assessment_82', 'ch7interim_assessment_83', 'ch7interim_assessment_84', 'ch7interim_assessment_85', 'ch7interim_assessment_86', 'ch7interim_assessment_87', 'ch7interim_assessment_88', 'ch7interim_assessment_89', 'ch7interim_assessment_8a', 'ch7interim_assessment_8b', 'ch7interim_assessment_8c', 'ch7interim_assessment_8d', 'ch7interim_assessment_8e', 'ch7interim_assessment_8f', 'ch7interim_assessment_90', 'ch7interim_assessment_91', 'ch7interim_assessment_92', 'ch7interim_assessment_93', 'ch7interim_assessment_94', 'ch7interim_assessment_95', 'ch7interim_assessment_96', 'ch7interim_assessment_97', 'ch7interim_assessment_98', 'ch7interim_assessment_99', 'ch7interim_assessment_9a', 'ch7interim_assessment_9b', 'ch7interim_assessment_9c', 'ch7interim_assessment_9d', 'ch7interim_assessment_9e', 'ch7interim_assessment_9f', 'ch7interim_assessment_a0', 'ch7interim_assessment_a1', 'ch7interim_assessment_a2', 'ch7interim_assessment_a3', 'ch7interim_assessment_a4', 'ch7interim_assessment_a5', 'ch7interim_assessment_a6', 'ch7interim_assessment_a7', 'ch7interim_assessment_a8', 'ch7interim_assessment_a9', 'ch7interim_assessment_aa', 'ch7interim_assessment_ab', 'ch7interim_assessment_ac', 'ch7interim_assessment_ad', 'ch7interim_assessment_ae', 'ch7interim_assessment_af', 'ch7interim_assessment_b0', 'ch7interim_assessment_b1', 'ch7interim_assessment_b2', 'ch7interim_assessment_b3', 'ch7interim_assessment_b4', 'ch7interim_assessment_b5', 'ch7interim_assessment_b6', 'ch7interim_assessment_b7', 'ch7interim_assessment_b8', 'ch7interim_assessment_b9', 'ch7interim_assessment_ba', 'ch7interim_assessment_bb', 'ch7interim_assessment_bc', 'ch7interim_assessment_bd', 'ch7interim_assessment_be', 'ch7interim_assessment_bf', 'ch7interim_assessment_c0', 'ch7interim_assessment_c1', 'ch7interim_assessment_c2', 'ch7interim_assessment_c3', 'ch7interim_assessment_c4', 'ch7interim_assessment_c5', 'ch7interim_assessment_c6', 'ch7interim_assessment_c7', 'ch7interim_assessment_c8', 'ch7interim_assessment_c9', 'ch7interim_assessment_ca', 'ch7interim_assessment_cb', 'ch7interim_assessment_cc', 'ch7interim_assessment_cd', 'ch7interim_assessment_ce', 'ch7interim_assessment_cf', 'ch7interim_assessment_d0', 'ch7interim_assessment_d1', 'ch7interim_assessment_d2', 'ch7interim_assessment_d3', 'ch7interim_assessment_d4', 'ch7interim_assessment_d5', 'ch7interim_assessment_d6', 'ch7interim_assessment_d7', 'ch7interim_assessment_d8', 'ch7interim_assessment_d9', 'ch7interim_assessment_da', 'ch7interim_assessment_db', 'ch7interim_assessment_dc', 'ch7interim_assessment_dd', 'ch7interim_assessment_de', 'ch7interim_assessment_df', 'ch7interim_assessment_e0', 'ch7interim_assessment_e1', 'ch7interim_assessment_e2', 'ch7interim_assessment_e3', 'ch7interim_assessment_e4', 'ch7interim_assessment_e5', 'ch7interim_assessment_e6', 'ch7interim_assessment_e7', 'ch7interim_assessment_e8', 'ch7interim_assessment_e9', 'ch7interim_assessment_ea', 'ch7interim_assessment_eb', 'ch7interim_assessment_ec', 'ch7interim_assessment_ed', 'ch7interim_assessment_ee', 'ch7interim_assessment_ef', 'ch7interim_assessment_f0', 'ch7interim_assessment_f1', 'ch7interim_assessment_f2', 'ch7interim_assessment_f3', 'ch7interim_assessment_f4', 'ch7interim_assessment_f5', 'ch7interim_assessment_f6', 'ch7interim_assessment_f7', 'ch7interim_assessment_f8', 'ch7interim_assessment_f9', 'ch7interim_assessment_fa', 'ch7interim_assessment_fb', 'ch7interim_assessment_fc', 'ch7interim_assessment_fd', 'ch7interim_assessment_fe', 'ch7interim_assessment_ff', 'ch7interim_assessment_100', 'ch7interim_assessment_101', 'ch7interim_assessment_102', 'ch7interim_assessment_103', 'ch7interim_assessment_104', 'ch7interim_assessment_105', 'ch7interim_assessment_106', 'ch7interim_assessment_107', 'ch7interim_assessment_108', 'ch7interim_assessment_109', 'ch7interim_assessment_10a', 'ch7interim_assessment_10b', 'ch7interim_assessment_10c', 'ch7interim_assessment_10d', 'ch7interim_assessment_10e', 'ch7interim_assessment_10f', 'ch7interim_assessment_110', 'ch7interim_assessment_111', 'ch7interim_assessment_112', 'ch7interim_assessment_113', 'ch7interim_assessment_114', 'ch7interim_assessment_115', 'ch7interim_assessment_116', 'ch7interim_assessment_117', 'ch7interim_assessment_118', 'ch7interim_assessment_119', 'ch7interim_assessment_11a', 'ch7interim_assessment_11b', 'ch7interim_assessment_11c', 'ch7interim_assessment_11d', 'ch7interim_assessment_11e', 'ch7interim_assessment_11f', 'ch7interim_assessment_120', 'ch7interim_assessment_121', 'ch7interim_assessment_122', 'ch7interim_assessment_123', 'ch7interim_assessment_124', 'ch7interim_assessment_125', 'ch7interim_assessment_126', 'ch7interim_assessment_127', 'ch7interim_assessment_128', 'ch7interim_assessment_129', 'ch7interim_assessment_12a', 'ch7interim_assessment_12b', 'ch7interim_assessment_12c', 'ch7interim_assessment_12d', 'ch7interim_assessment_12e', 'ch7interim_assessment_12f', 'ch7interim_assessment_130', 'ch7interim_assessment_131', 'ch7interim_assessment_132', 'ch7interim_assessment_133', 'ch7interim_assessment_134', 'ch7interim_assessment_135', 'ch7interim_assessment_136', 'ch7interim_assessment_137', 'ch7interim_assessment_138', 'ch7interim_assessment_139', 'ch7interim_assessment_13a', 'ch7interim_assessment_13b', 'ch7interim_assessment_13c', 'ch7interim_assessment_13d', 'ch7interim_assessment_13e', 'ch7interim_assessment_13f', 'ch7interim_assessment_140', 'ch7interim_assessment_141', 'ch7interim_assessment_142', 'ch7interim_assessment_143', 'ch7interim_assessment_144', 'ch7interim_assessment_145', 'ch7interim_assessment_146', 'ch7interim_assessment_147', 'ch7interim_assessment_148', 'ch7interim_assessment_149', 'ch7interim_assessment_14a', 'ch7interim_assessment_14b', 'ch7interim_assessment_14c', 'ch7interim_assessment_14d', 'ch7interim_assessment_14e', 'ch7interim_assessment_14f', 'ch7interim_assessment_150', 'ch7interim_assessment_151', 'ch7interim_assessment_152', 'ch7interim_assessment_153', 'ch7interim_assessment_154', 'ch7interim_assessment_155', 'ch7interim_assessment_156', 'ch7interim_assessment_157', 'ch7interim_assessment_158', 'ch7interim_assessment_159', 'ch7interim_assessment_15a', 'ch7interim_assessment_15b', 'ch7interim_assessment_15c', 'ch7interim_assessment_15d', 'ch7interim_assessment_15e', 'ch7interim_assessment_15f', 'ch7interim_assessment_160', 'ch7interim_assessment_161', 'ch7interim_assessment_162', 'ch7interim_assessment_163', 'ch7interim_assessment_164', 'ch7interim_assessment_165', 'ch7interim_assessment_166', 'ch7interim_assessment_167', 'ch7interim_assessment_168', 'ch7interim_assessment_169', 'ch7interim_assessment_16a', 'ch7interim_assessment_16b', 'ch7interim_assessment_16c', 'ch7interim_assessment_16d', 'ch7interim_assessment_16e', 'ch7interim_assessment_16f', 'ch7interim_assessment_170', 'ch7interim_assessment_171', 'ch7interim_assessment_172', 'ch7interim_assessment_173', 'ch7interim_assessment_174', 'ch7interim_assessment_175', 'ch7interim_assessment_176', 'ch7interim_assessment_177', 'ch7interim_assessment_178', 'ch7interim_assessment_179', 'ch7interim_assessment_17a', 'ch7interim_assessment_17b', 'ch7interim_assessment_17c', 'ch7interim_assessment_17d', 'ch7interim_assessment_17e', 'ch7interim_assessment_17f', 'ch7interim_assessment_180', 'ch7interim_assessment_181', 'ch7interim_assessment_182', 'ch7interim_assessment_183', 'ch7interim_assessment_184', 'ch7interim_assessment_185', 'ch7interim_assessment_186', 'ch7interim_assessment_187', 'ch7interim_assessment_188', 'ch7interim_assessment_189', 'ch7interim_assessment_18a', 'ch7interim_assessment_18b', 'ch7interim_assessment_18c', 'ch7interim_assessment_18d', 'ch7interim_assessment_18e', 'ch7interim_assessment_18f', 'ch7interim_assessment_190', 'ch7interim_assessment_191', 'ch7interim_assessment_192', 'ch7interim_assessment_193', 'ch7interim_assessment_194', 'ch7interim_assessment_195', 'ch7interim_assessment_196', 'ch7interim_assessment_197', 'ch7interim_assessment_198', 'ch7interim_assessment_199', 'ch7interim_assessment_19a', 'ch7interim_assessment_19b', 'ch7interim_assessment_19c', 'ch7interim_assessment_19d', 'ch7interim_assessment_19e', 'ch7interim_assessment_19f', 'ch7interim_assessment_1a0', 'ch7interim_assessment_1a1', 'ch7interim_assessment_1a2', 'ch7interim_assessment_1a3', 'ch7interim_assessment_1a4', 'ch7interim_assessment_1a5', 'ch7interim_assessment_1a6', 'ch7interim_assessment_1a7', 'ch7interim_assessment_1a8', 'ch7interim_assessment_1a9', 'ch7interim_assessment_1aa', 'ch7interim_assessment_1ab', 'ch7interim_assessment_1ac', 'ch7interim_assessment_1ad', 'ch7interim_assessment_1ae', 'ch7interim_assessment_1af', 'ch7interim_assessment_1b0', 'ch7interim_assessment_1b1', 'ch7interim_assessment_1b2', 'ch7interim_assessment_1b3', 'ch7interim_assessment_1b4', 'ch7interim_assessment_1b5', 'ch7interim_assessment_1b6', 'ch7interim_assessment_1b7', 'ch7interim_assessment_1b8', 'ch7interim_assessment_1b9', 'ch7interim_assessment_1ba', 'ch7interim_assessment_1bb', 'ch7interim_assessment_1bc', 'ch7interim_assessment_1bd', 'ch7interim_assessment_1be', 'ch7interim_assessment_1bf', 'ch7interim_assessment_1c0', 'ch7interim_assessment_1c1', 'ch7interim_assessment_1c2', 'ch7interim_assessment_1c3', 'ch7interim_assessment_1c4', 'ch7interim_assessment_1c5', 'ch7interim_assessment_1c6', 'ch7interim_assessment_1c7', 'ch7interim_assessment_1c8', 'ch7interim_assessment_1c9', 'ch7interim_assessment_1ca', 'ch7interim_assessment_1cb', 'ch7interim_assessment_1cc', 'ch7interim_assessment_1cd', 'ch7interim_assessment_1ce', 'ch7interim_assessment_1cf', 'ch7interim_assessment_1d0', 'ch7interim_assessment_1d1', 'ch7interim_assessment_1d2', 'ch7interim_assessment_1d3', 'ch7interim_assessment_1d4', 'ch7interim_assessment_1d5', 'ch7interim_assessment_1d6', 'ch7interim_assessment_1d7', 'ch7interim_assessment_1d8', 'ch7interim_assessment_1d9', 'ch7interim_assessment_1da', 'ch7interim_assessment_1db', 'ch7interim_assessment_1dc', 'ch7interim_assessment_1dd', 'ch7interim_assessment_1de', 'ch7interim_assessment_1df', 'ch7interim_assessment_1e0', 'ch7interim_assessment_1e1'], 'ch8ult_module_plan': ['ch8ult_module_plan_1', 'ch8ult_module_plan_2', 'ch8ult_module_plan_3', 'ch8ult_module_plan_4', 'ch8ult_module_plan_5', 'ch8ult_module_plan_6', 'ch8ult_module_plan_7', 'ch8ult_module_plan_8', 'ch8ult_module_plan_9', 'ch8ult_module_plan_a', 'ch8ult_module_plan_b', 'ch8ult_module_plan_c', 'ch8ult_module_plan_d', 'ch8ult_module_plan_e', 'ch8ult_module_plan_f', 'ch8ult_module_plan_10', 'ch8ult_module_plan_11', 'ch8ult_module_plan_12', 'ch8ult_module_plan_13', 'ch8ult_module_plan_14', 'ch8ult_module_plan_15', 'ch8ult_module_plan_16', 'ch8ult_module_plan_17', 'ch8ult_module_plan_18', 'ch8ult_module_plan_19', 'ch8ult_module_plan_1a', 'ch8ult_module_plan_1b', 'ch8ult_module_plan_1c', 'ch8ult_module_plan_1d', 'ch8ult_module_plan_1e', 'ch8ult_module_plan_1f', 'ch8ult_module_plan_20', 'ch8ult_module_plan_21', 'ch8ult_module_plan_22', 'ch8ult_module_plan_23', 'ch8ult_module_plan_24', 'ch8ult_module_plan_25', 'ch8ult_module_plan_26', 'ch8ult_module_plan_27', 'ch8ult_module_plan_28', 'ch8ult_module_plan_29', 'ch8ult_module_plan_2a', 'ch8ult_module_plan_2b', 'ch8ult_module_plan_2c', 'ch8ult_module_plan_2d', 'ch8ult_module_plan_2e', 'ch8ult_module_plan_2f', 'ch8ult_module_plan_30', 'ch8ult_module_plan_31', 'ch8ult_module_plan_32', 'ch8ult_module_plan_33', 'ch8ult_module_plan_34', 'ch8ult_module_plan_35', 'ch8ult_module_plan_36', 'ch8ult_module_plan_37', 'ch8ult_module_plan_38', 'ch8ult_module_plan_39', 'ch8ult_module_plan_3a', 'ch8ult_module_plan_3b', 'ch8ult_module_plan_3c', 'ch8ult_module_plan_3d', 'ch8ult_module_plan_3e', 'ch8ult_module_plan_3f', 'ch8ult_module_plan_40', 'ch8ult_module_plan_41', 'ch8ult_module_plan_42', 'ch8ult_module_plan_43', 'ch8ult_module_plan_44', 'ch8ult_module_plan_45', 'ch8ult_module_plan_46', 'ch8ult_module_plan_47', 'ch8ult_module_plan_48', 'ch8ult_module_plan_49', 'ch8ult_module_plan_4a', 'ch8ult_module_plan_4b', 'ch8ult_module_plan_4c', 'ch8ult_module_plan_4d', 'ch8ult_module_plan_4e', 'ch8ult_module_plan_4f', 'ch8ult_module_plan_50', 'ch8ult_module_plan_51', 'ch8ult_module_plan_52', 'ch8ult_module_plan_53', 'ch8ult_module_plan_54', 'ch8ult_module_plan_55', 'ch8ult_module_plan_56', 'ch8ult_module_plan_57', 'ch8ult_module_plan_58', 'ch8ult_module_plan_59', 'ch8ult_module_plan_5a', 'ch8ult_module_plan_5b', 'ch8ult_module_plan_5c', 'ch8ult_module_plan_5d', 'ch8ult_module_plan_5e', 'ch8ult_module_plan_5f', 'ch8ult_module_plan_60', 'ch8ult_module_plan_61', 'ch8ult_module_plan_62', 'ch8ult_module_plan_63', 'ch8ult_module_plan_64', 'ch8ult_module_plan_65', 'ch8ult_module_plan_66', 'ch8ult_module_plan_67', 'ch8ult_module_plan_68', 'ch8ult_module_plan_69', 'ch8ult_module_plan_6a', 'ch8ult_module_plan_6b', 'ch8ult_module_plan_6c', 'ch8ult_module_plan_6d', 'ch8ult_module_plan_6e', 'ch8ult_module_plan_6f', 'ch8ult_module_plan_70', 'ch8ult_module_plan_71', 'ch8ult_module_plan_72', 'ch8ult_module_plan_73', 'ch8ult_module_plan_74', 'ch8ult_module_plan_75', 'ch8ult_module_plan_76', 'ch8ult_module_plan_77', 'ch8ult_module_plan_78', 'ch8ult_module_plan_79', 'ch8ult_module_plan_7a', 'ch8ult_module_plan_7b', 'ch8ult_module_plan_7c', 'ch8ult_module_plan_7d', 'ch8ult_module_plan_7e', 'ch8ult_module_plan_7f', 'ch8ult_module_plan_80', 'ch8ult_module_plan_81', 'ch8ult_module_plan_82', 'ch8ult_module_plan_83', 'ch8ult_module_plan_84', 'ch8ult_module_plan_85', 'ch8ult_module_plan_86', 'ch8ult_module_plan_87', 'ch8ult_module_plan_88', 'ch8ult_module_plan_89', 'ch8ult_module_plan_8a', 'ch8ult_module_plan_8b', 'ch8ult_module_plan_8c', 'ch8ult_module_plan_8d', 'ch8ult_module_plan_8e', 'ch8ult_module_plan_8f', 'ch8ult_module_plan_90', 'ch8ult_module_plan_91', 'ch8ult_module_plan_92', 'ch8ult_module_plan_93', 'ch8ult_module_plan_94', 'ch8ult_module_plan_95', 'ch8ult_module_plan_96', 'ch8ult_module_plan_97', 'ch8ult_module_plan_98', 'ch8ult_module_plan_99', 'ch8ult_module_plan_9a', 'ch8ult_module_plan_9b', 'ch8ult_module_plan_9c', 'ch8ult_module_plan_9d', 'ch8ult_module_plan_9e', 'ch8ult_module_plan_9f', 'ch8ult_module_plan_a0', 'ch8ult_module_plan_a1', 'ch8ult_module_plan_a2', 'ch8ult_module_plan_a3', 'ch8ult_module_plan_a4', 'ch8ult_module_plan_a5', 'ch8ult_module_plan_a6', 'ch8ult_module_plan_a7', 'ch8ult_module_plan_a8', 'ch8ult_module_plan_a9', 'ch8ult_module_plan_aa', 'ch8ult_module_plan_ab', 'ch8ult_module_plan_ac', 'ch8ult_module_plan_ad', 'ch8ult_module_plan_ae', 'ch8ult_module_plan_af', 'ch8ult_module_plan_b0', 'ch8ult_module_plan_b1', 'ch8ult_module_plan_b2', 'ch8ult_module_plan_b3', 'ch8ult_module_plan_b4', 'ch8ult_module_plan_b5', 'ch8ult_module_plan_b6', 'ch8ult_module_plan_b7', 'ch8ult_module_plan_b8', 'ch8ult_module_plan_b9', 'ch8ult_module_plan_ba', 'ch8ult_module_plan_bb', 'ch8ult_module_plan_bc', 'ch8ult_module_plan_bd', 'ch8ult_module_plan_be', 'ch8ult_module_plan_bf', 'ch8ult_module_plan_c0', 'ch8ult_module_plan_c1', 'ch8ult_module_plan_c2', 'ch8ult_module_plan_c3', 'ch8ult_module_plan_c4', 'ch8ult_module_plan_c5', 'ch8ult_module_plan_c6', 'ch8ult_module_plan_c7', 'ch8ult_module_plan_c8', 'ch8ult_module_plan_c9', 'ch8ult_module_plan_ca', 'ch8ult_module_plan_cb', 'ch8ult_module_plan_cc', 'ch8ult_module_plan_cd', 'ch8ult_module_plan_ce', 'ch8ult_module_plan_cf', 'ch8ult_module_plan_d0', 'ch8ult_module_plan_d1', 'ch8ult_module_plan_d2', 'ch8ult_module_plan_d3', 'ch8ult_module_plan_d4', 'ch8ult_module_plan_d5', 'ch8ult_module_plan_d6', 'ch8ult_module_plan_d7', 'ch8ult_module_plan_d8', 'ch8ult_module_plan_d9', 'ch8ult_module_plan_da', 'ch8ult_module_plan_db', 'ch8ult_module_plan_dc', 'ch8ult_module_plan_dd', 'ch8ult_module_plan_de', 'ch8ult_module_plan_df', 'ch8ult_module_plan_e0', 'ch8ult_module_plan_e1', 'ch8ult_module_plan_e2', 'ch8ult_module_plan_e3', 'ch8ult_module_plan_e4', 'ch8ult_module_plan_e5', 'ch8ult_module_plan_e6', 'ch8ult_module_plan_e7', 'ch8ult_module_plan_e8', 'ch8ult_module_plan_e9', 'ch8ult_module_plan_ea', 'ch8ult_module_plan_eb', 'ch8ult_module_plan_ec', 'ch8ult_module_plan_ed', 'ch8ult_module_plan_ee', 'ch8ult_module_plan_ef', 'ch8ult_module_plan_f0', 'ch8ult_module_plan_f1', 'ch8ult_module_plan_f2', 'ch8ult_module_plan_f3', 'ch8ult_module_plan_f4', 'ch8ult_module_plan_f5', 'ch8ult_module_plan_f6', 'ch8ult_module_plan_f7', 'ch8ult_module_plan_f8', 'ch8ult_module_plan_f9', 'ch8ult_module_plan_fa', 'ch8ult_module_plan_fb', 'ch8ult_module_plan_fc', 'ch8ult_module_plan_fd', 'ch8ult_module_plan_fe', 'ch8ult_module_plan_ff', 'ch8ult_module_plan_100', 'ch8ult_module_plan_101', 'ch8ult_module_plan_102', 'ch8ult_module_plan_103', 'ch8ult_module_plan_104', 'ch8ult_module_plan_105', 'ch8ult_module_plan_106', 'ch8ult_module_plan_107', 'ch8ult_module_plan_108', 'ch8ult_module_plan_109', 'ch8ult_module_plan_10a', 'ch8ult_module_plan_10b', 'ch8ult_module_plan_10c', 'ch8ult_module_plan_10d', 'ch8ult_module_plan_10e', 'ch8ult_module_plan_10f', 'ch8ult_module_plan_110', 'ch8ult_module_plan_111', 'ch8ult_module_plan_112', 'ch8ult_module_plan_113', 'ch8ult_module_plan_114', 'ch8ult_module_plan_115', 'ch8ult_module_plan_116', 'ch8ult_module_plan_117', 'ch8ult_module_plan_118', 'ch8ult_module_plan_119', 'ch8ult_module_plan_11a', 'ch8ult_module_plan_11b', 'ch8ult_module_plan_11c', 'ch8ult_module_plan_11d', 'ch8ult_module_plan_11e', 'ch8ult_module_plan_11f', 'ch8ult_module_plan_120', 'ch8ult_module_plan_121', 'ch8ult_module_plan_122', 'ch8ult_module_plan_123', 'ch8ult_module_plan_124', 'ch8ult_module_plan_125', 'ch8ult_module_plan_126', 'ch8ult_module_plan_127', 'ch8ult_module_plan_128', 'ch8ult_module_plan_129', 'ch8ult_module_plan_12a', 'ch8ult_module_plan_12b', 'ch8ult_module_plan_12c', 'ch8ult_module_plan_12d', 'ch8ult_module_plan_12e', 'ch8ult_module_plan_12f', 'ch8ult_module_plan_130', 'ch8ult_module_plan_131', 'ch8ult_module_plan_132', 'ch8ult_module_plan_133', 'ch8ult_module_plan_134', 'ch8ult_module_plan_135', 'ch8ult_module_plan_136', 'ch8ult_module_plan_137', 'ch8ult_module_plan_138', 'ch8ult_module_plan_139', 'ch8ult_module_plan_13a', 'ch8ult_module_plan_13b', 'ch8ult_module_plan_13c', 'ch8ult_module_plan_13d', 'ch8ult_module_plan_13e', 'ch8ult_module_plan_13f', 'ch8ult_module_plan_140', 'ch8ult_module_plan_141', 'ch8ult_module_plan_142', 'ch8ult_module_plan_143', 'ch8ult_module_plan_144', 'ch8ult_module_plan_145', 'ch8ult_module_plan_146', 'ch8ult_module_plan_147', 'ch8ult_module_plan_148', 'ch8ult_module_plan_149', 'ch8ult_module_plan_14a', 'ch8ult_module_plan_14b', 'ch8ult_module_plan_14c', 'ch8ult_module_plan_14d', 'ch8ult_module_plan_14e', 'ch8ult_module_plan_14f', 'ch8ult_module_plan_150', 'ch8ult_module_plan_151', 'ch8ult_module_plan_152', 'ch8ult_module_plan_153', 'ch8ult_module_plan_154', 'ch8ult_module_plan_155', 'ch8ult_module_plan_156', 'ch8ult_module_plan_157', 'ch8ult_module_plan_158', 'ch8ult_module_plan_159', 'ch8ult_module_plan_15a', 'ch8ult_module_plan_15b', 'ch8ult_module_plan_15c', 'ch8ult_module_plan_15d', 'ch8ult_module_plan_15e', 'ch8ult_module_plan_15f', 'ch8ult_module_plan_160', 'ch8ult_module_plan_161', 'ch8ult_module_plan_162', 'ch8ult_module_plan_163', 'ch8ult_module_plan_164', 'ch8ult_module_plan_165', 'ch8ult_module_plan_166', 'ch8ult_module_plan_167', 'ch8ult_module_plan_168', 'ch8ult_module_plan_169', 'ch8ult_module_plan_16a', 'ch8ult_module_plan_16b', 'ch8ult_module_plan_16c', 'ch8ult_module_plan_16d', 'ch8ult_module_plan_16e', 'ch8ult_module_plan_16f', 'ch8ult_module_plan_170', 'ch8ult_module_plan_171', 'ch8ult_module_plan_172', 'ch8ult_module_plan_173', 'ch8ult_module_plan_174', 'ch8ult_module_plan_175', 'ch8ult_module_plan_176', 'ch8ult_module_plan_177', 'ch8ult_module_plan_178', 'ch8ult_module_plan_179', 'ch8ult_module_plan_17a', 'ch8ult_module_plan_17b', 'ch8ult_module_plan_17c', 'ch8ult_module_plan_17d', 'ch8ult_module_plan_17e', 'ch8ult_module_plan_17f', 'ch8ult_module_plan_180', 'ch8ult_module_plan_181', 'ch8ult_module_plan_182', 'ch8ult_module_plan_183', 'ch8ult_module_plan_184', 'ch8ult_module_plan_185', 'ch8ult_module_plan_186', 'ch8ult_module_plan_187', 'ch8ult_module_plan_188', 'ch8ult_module_plan_189', 'ch8ult_module_plan_18a', 'ch8ult_module_plan_18b', 'ch8ult_module_plan_18c', 'ch8ult_module_plan_18d', 'ch8ult_module_plan_18e', 'ch8ult_module_plan_18f', 'ch8ult_module_plan_190', 'ch8ult_module_plan_191', 'ch8ult_module_plan_192', 'ch8ult_module_plan_193', 'ch8ult_module_plan_194', 'ch8ult_module_plan_195', 'ch8ult_module_plan_196', 'ch8ult_module_plan_197', 'ch8ult_module_plan_198', 'ch8ult_module_plan_199', 'ch8ult_module_plan_19a', 'ch8ult_module_plan_19b', 'ch8ult_module_plan_19c', 'ch8ult_module_plan_19d', 'ch8ult_module_plan_19e', 'ch8ult_module_plan_19f', 'ch8ult_module_plan_1a0', 'ch8ult_module_plan_1a1', 'ch8ult_module_plan_1a2', 'ch8ult_module_plan_1a3', 'ch8ult_module_plan_1a4', 'ch8ult_module_plan_1a5', 'ch8ult_module_plan_1a6', 'ch8ult_module_plan_1a7', 'ch8ult_module_plan_1a8', 'ch8ult_module_plan_1a9', 'ch8ult_module_plan_1aa', 'ch8ult_module_plan_1ab', 'ch8ult_module_plan_1ac', 'ch8ult_module_plan_1ad', 'ch8ult_module_plan_1ae', 'ch8ult_module_plan_1af', 'ch8ult_module_plan_1b0', 'ch8ult_module_plan_1b1', 'ch8ult_module_plan_1b2', 'ch8ult_module_plan_1b3', 'ch8ult_module_plan_1b4', 'ch8ult_module_plan_1b5', 'ch8ult_module_plan_1b6', 'ch8ult_module_plan_1b7', 'ch8ult_module_plan_1b8', 'ch8ult_module_plan_1b9', 'ch8ult_module_plan_1ba', 'ch8ult_module_plan_1bb', 'ch8ult_module_plan_1bc', 'ch8ult_module_plan_1bd', 'ch8ult_module_plan_1be', 'ch8ult_module_plan_1bf', 'ch8ult_module_plan_1c0', 'ch8ult_module_plan_1c1', 'ch8ult_module_plan_1c2', 'ch8ult_module_plan_1c3', 'ch8ult_module_plan_1c4', 'ch8ult_module_plan_1c5', 'ch8ult_module_plan_1c6', 'ch8ult_module_plan_1c7', 'ch8ult_module_plan_1c8', 'ch8ult_module_plan_1c9', 'ch8ult_module_plan_1ca', 'ch8ult_module_plan_1cb', 'ch8ult_module_plan_1cc', 'ch8ult_module_plan_1cd', 'ch8ult_module_plan_1ce', 'ch8ult_module_plan_1cf', 'ch8ult_module_plan_1d0', 'ch8ult_module_plan_1d1', 'ch8ult_module_plan_1d2', 'ch8ult_module_plan_1d3', 'ch8ult_module_plan_1d4', 'ch8ult_module_plan_1d5', 'ch8ult_module_plan_1d6', 'ch8ult_module_plan_1d7', 'ch8ult_module_plan_1d8', 'ch8ult_module_plan_1d9', 'ch8ult_module_plan_1da', 'ch8ult_module_plan_1db', 'ch8ult_module_plan_1dc', 'ch8ult_module_plan_1dd', 'ch8ult_module_plan_1de', 'ch8ult_module_plan_1df', 'ch8ult_module_plan_1e0', 'ch8ult_module_plan_1e1'], 'ch10session4': ['ch10session4_1', 'ch10session4_2', 'ch10session4_3', 'ch10session4_4', 'ch10session4_5', 'ch10session4_6', 'ch10session4_7', 'ch10session4_8', 'ch10session4_9', 'ch10session4_a', 'ch10session4_b', 'ch10session4_c', 'ch10session4_d', 'ch10session4_e', 'ch10session4_f', 'ch10session4_10', 'ch10session4_11', 'ch10session4_12', 'ch10session4_13', 'ch10session4_14', 'ch10session4_15', 'ch10session4_16', 'ch10session4_17', 'ch10session4_18', 'ch10session4_19', 'ch10session4_1a', 'ch10session4_1b', 'ch10session4_1c', 'ch10session4_1d', 'ch10session4_1e', 'ch10session4_1f', 'ch10session4_20', 'ch10session4_21', 'ch10session4_22', 'ch10session4_23', 'ch10session4_24', 'ch10session4_25', 'ch10session4_26', 'ch10session4_27', 'ch10session4_28', 'ch10session4_29', 'ch10session4_2a', 'ch10session4_2b', 'ch10session4_2c', 'ch10session4_2d', 'ch10session4_2e', 'ch10session4_2f', 'ch10session4_30', 'ch10session4_31', 'ch10session4_32', 'ch10session4_33', 'ch10session4_34', 'ch10session4_35', 'ch10session4_36', 'ch10session4_37', 'ch10session4_38', 'ch10session4_39', 'ch10session4_3a', 'ch10session4_3b', 'ch10session4_3c', 'ch10session4_3d', 'ch10session4_3e', 'ch10session4_3f', 'ch10session4_40', 'ch10session4_41', 'ch10session4_42', 'ch10session4_43', 'ch10session4_44', 'ch10session4_45', 'ch10session4_46', 'ch10session4_47', 'ch10session4_48', 'ch10session4_49', 'ch10session4_4a', 'ch10session4_4b', 'ch10session4_4c', 'ch10session4_4d', 'ch10session4_4e', 'ch10session4_4f', 'ch10session4_50', 'ch10session4_51', 'ch10session4_52', 'ch10session4_53', 'ch10session4_54', 'ch10session4_55', 'ch10session4_56', 'ch10session4_57', 'ch10session4_58', 'ch10session4_59', 'ch10session4_5a', 'ch10session4_5b', 'ch10session4_5c', 'ch10session4_5d', 'ch10session4_5e', 'ch10session4_5f', 'ch10session4_60', 'ch10session4_61', 'ch10session4_62', 'ch10session4_63', 'ch10session4_64', 'ch10session4_65', 'ch10session4_66', 'ch10session4_67', 'ch10session4_68', 'ch10session4_69', 'ch10session4_6a', 'ch10session4_6b', 'ch10session4_6c', 'ch10session4_6d', 'ch10session4_6e', 'ch10session4_6f', 'ch10session4_70', 'ch10session4_71', 'ch10session4_72', 'ch10session4_73', 'ch10session4_74', 'ch10session4_75', 'ch10session4_76', 'ch10session4_77', 'ch10session4_78', 'ch10session4_79', 'ch10session4_7a', 'ch10session4_7b', 'ch10session4_7c', 'ch10session4_7d', 'ch10session4_7e', 'ch10session4_7f', 'ch10session4_80', 'ch10session4_81', 'ch10session4_82', 'ch10session4_83', 'ch10session4_84', 'ch10session4_85', 'ch10session4_86', 'ch10session4_87', 'ch10session4_88', 'ch10session4_89', 'ch10session4_8a', 'ch10session4_8b', 'ch10session4_8c', 'ch10session4_8d', 'ch10session4_8e', 'ch10session4_8f', 'ch10session4_90', 'ch10session4_91', 'ch10session4_92', 'ch10session4_93', 'ch10session4_94', 'ch10session4_95', 'ch10session4_96', 'ch10session4_97', 'ch10session4_98', 'ch10session4_99', 'ch10session4_9a', 'ch10session4_9b', 'ch10session4_9c', 'ch10session4_9d', 'ch10session4_9e', 'ch10session4_9f', 'ch10session4_a0', 'ch10session4_a1', 'ch10session4_a2', 'ch10session4_a3', 'ch10session4_a4', 'ch10session4_a5', 'ch10session4_a6', 'ch10session4_a7', 'ch10session4_a8', 'ch10session4_a9', 'ch10session4_aa', 'ch10session4_ab', 'ch10session4_ac', 'ch10session4_ad', 'ch10session4_ae', 'ch10session4_af', 'ch10session4_b0', 'ch10session4_b1', 'ch10session4_b2', 'ch10session4_b3', 'ch10session4_b4', 'ch10session4_b5', 'ch10session4_b6', 'ch10session4_b7', 'ch10session4_b8', 'ch10session4_b9', 'ch10session4_ba', 'ch10session4_bb', 'ch10session4_bc', 'ch10session4_bd', 'ch10session4_be', 'ch10session4_bf', 'ch10session4_c0', 'ch10session4_c1', 'ch10session4_c2', 'ch10session4_c3', 'ch10session4_c4', 'ch10session4_c5', 'ch10session4_c6', 'ch10session4_c7', 'ch10session4_c8', 'ch10session4_c9', 'ch10session4_ca', 'ch10session4_cb', 'ch10session4_cc', 'ch10session4_cd', 'ch10session4_ce', 'ch10session4_cf', 'ch10session4_d0', 'ch10session4_d1', 'ch10session4_d2', 'ch10session4_d3', 'ch10session4_d4', 'ch10session4_d5', 'ch10session4_d6', 'ch10session4_d7', 'ch10session4_d8', 'ch10session4_d9', 'ch10session4_da', 'ch10session4_db', 'ch10session4_dc', 'ch10session4_dd', 'ch10session4_de', 'ch10session4_df', 'ch10session4_e0', 'ch10session4_e1', 'ch10session4_e2', 'ch10session4_e3', 'ch10session4_e4', 'ch10session4_e5', 'ch10session4_e6', 'ch10session4_e7', 'ch10session4_e8', 'ch10session4_e9', 'ch10session4_ea', 'ch10session4_eb', 'ch10session4_ec', 'ch10session4_ed', 'ch10session4_ee', 'ch10session4_ef', 'ch10session4_f0', 'ch10session4_f1', 'ch10session4_f2', 'ch10session4_f3', 'ch10session4_f4', 'ch10session4_f5', 'ch10session4_f6', 'ch10session4_f7', 'ch10session4_f8', 'ch10session4_f9', 'ch10session4_fa', 'ch10session4_fb', 'ch10session4_fc', 'ch10session4_fd', 'ch10session4_fe', 'ch10session4_ff', 'ch10session4_100', 'ch10session4_101', 'ch10session4_102', 'ch10session4_103', 'ch10session4_104', 'ch10session4_105', 'ch10session4_106', 'ch10session4_107', 'ch10session4_108', 'ch10session4_109', 'ch10session4_10a', 'ch10session4_10b', 'ch10session4_10c', 'ch10session4_10d', 'ch10session4_10e', 'ch10session4_10f', 'ch10session4_110', 'ch10session4_111', 'ch10session4_112', 'ch10session4_113', 'ch10session4_114', 'ch10session4_115', 'ch10session4_116', 'ch10session4_117', 'ch10session4_118', 'ch10session4_119', 'ch10session4_11a', 'ch10session4_11b', 'ch10session4_11c', 'ch10session4_11d', 'ch10session4_11e', 'ch10session4_11f', 'ch10session4_120', 'ch10session4_121', 'ch10session4_122', 'ch10session4_123', 'ch10session4_124', 'ch10session4_125', 'ch10session4_126', 'ch10session4_127', 'ch10session4_128', 'ch10session4_129', 'ch10session4_12a', 'ch10session4_12b', 'ch10session4_12c', 'ch10session4_12d', 'ch10session4_12e', 'ch10session4_12f', 'ch10session4_130', 'ch10session4_131', 'ch10session4_132', 'ch10session4_133', 'ch10session4_134', 'ch10session4_135', 'ch10session4_136', 'ch10session4_137', 'ch10session4_138', 'ch10session4_139', 'ch10session4_13a', 'ch10session4_13b', 'ch10session4_13c', 'ch10session4_13d', 'ch10session4_13e', 'ch10session4_13f', 'ch10session4_140', 'ch10session4_141', 'ch10session4_142', 'ch10session4_143', 'ch10session4_144', 'ch10session4_145', 'ch10session4_146', 'ch10session4_147', 'ch10session4_148', 'ch10session4_149', 'ch10session4_14a', 'ch10session4_14b', 'ch10session4_14c', 'ch10session4_14d', 'ch10session4_14e', 'ch10session4_14f', 'ch10session4_150', 'ch10session4_151', 'ch10session4_152', 'ch10session4_153', 'ch10session4_154', 'ch10session4_155', 'ch10session4_156', 'ch10session4_157', 'ch10session4_158', 'ch10session4_159', 'ch10session4_15a', 'ch10session4_15b', 'ch10session4_15c', 'ch10session4_15d', 'ch10session4_15e', 'ch10session4_15f', 'ch10session4_160', 'ch10session4_161', 'ch10session4_162', 'ch10session4_163', 'ch10session4_164', 'ch10session4_165', 'ch10session4_166', 'ch10session4_167', 'ch10session4_168', 'ch10session4_169', 'ch10session4_16a', 'ch10session4_16b', 'ch10session4_16c', 'ch10session4_16d', 'ch10session4_16e', 'ch10session4_16f', 'ch10session4_170', 'ch10session4_171', 'ch10session4_172', 'ch10session4_173', 'ch10session4_174', 'ch10session4_175', 'ch10session4_176', 'ch10session4_177', 'ch10session4_178', 'ch10session4_179', 'ch10session4_17a', 'ch10session4_17b', 'ch10session4_17c', 'ch10session4_17d', 'ch10session4_17e', 'ch10session4_17f', 'ch10session4_180', 'ch10session4_181', 'ch10session4_182', 'ch10session4_183', 'ch10session4_184', 'ch10session4_185', 'ch10session4_186', 'ch10session4_187', 'ch10session4_188', 'ch10session4_189', 'ch10session4_18a', 'ch10session4_18b', 'ch10session4_18c', 'ch10session4_18d', 'ch10session4_18e', 'ch10session4_18f', 'ch10session4_190', 'ch10session4_191', 'ch10session4_192', 'ch10session4_193', 'ch10session4_194', 'ch10session4_195', 'ch10session4_196', 'ch10session4_197', 'ch10session4_198', 'ch10session4_199', 'ch10session4_19a', 'ch10session4_19b', 'ch10session4_19c', 'ch10session4_19d', 'ch10session4_19e', 'ch10session4_19f', 'ch10session4_1a0', 'ch10session4_1a1', 'ch10session4_1a2', 'ch10session4_1a3', 'ch10session4_1a4', 'ch10session4_1a5', 'ch10session4_1a6', 'ch10session4_1a7', 'ch10session4_1a8', 'ch10session4_1a9', 'ch10session4_1aa', 'ch10session4_1ab', 'ch10session4_1ac', 'ch10session4_1ad', 'ch10session4_1ae', 'ch10session4_1af', 'ch10session4_1b0', 'ch10session4_1b1', 'ch10session4_1b2', 'ch10session4_1b3', 'ch10session4_1b4', 'ch10session4_1b5', 'ch10session4_1b6', 'ch10session4_1b7', 'ch10session4_1b8', 'ch10session4_1b9', 'ch10session4_1ba', 'ch10session4_1bb', 'ch10session4_1bc', 'ch10session4_1bd', 'ch10session4_1be', 'ch10session4_1bf', 'ch10session4_1c0', 'ch10session4_1c1', 'ch10session4_1c2', 'ch10session4_1c3', 'ch10session4_1c4', 'ch10session4_1c5', 'ch10session4_1c6', 'ch10session4_1c7', 'ch10session4_1c8', 'ch10session4_1c9', 'ch10session4_1ca', 'ch10session4_1cb', 'ch10session4_1cc', 'ch10session4_1cd', 'ch10session4_1ce', 'ch10session4_1cf', 'ch10session4_1d0', 'ch10session4_1d1', 'ch10session4_1d2', 'ch10session4_1d3', 'ch10session4_1d4', 'ch10session4_1d5', 'ch10session4_1d6', 'ch10session4_1d7', 'ch10session4_1d8', 'ch10session4_1d9', 'ch10session4_1da', 'ch10session4_1db', 'ch10session4_1dc', 'ch10session4_1dd', 'ch10session4_1de', 'ch10session4_1df', 'ch10session4_1e0', 'ch10session4_1e1']} \ No newline at end of file diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/p_list.pyc Binary file SEESenv/web/hgbook/comments/p_list.pyc has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/p_list.py~ --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/comments/p_list.py~ Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1 @@ +p_list={'chap_intro':['x_38', 'x_39', 'x_3d', 'x_3e', 'x_3f', 'x_40', 'x_41', 'x_42', 'x_46', 'x_4c', 'x_4d', 'x_4e', 'x_4f', 'x_50', 'x_51', 'x_52', 'x_53', 'x_54', 'x_55', 'x_56', 'x_57', 'x_58', 'x_59', 'x_5a', 'x_5b', 'x_5c', 'x_5d', 'x_5e', 'x_5f', 'x_60', 'x_61', 'x_62', 'x_63', 'x_64', 'x_65', 'x_66', 'x_67', 'x_68', 'x_69', 'x_6d', 'x_6e', 'x_6f', 'x_70', 'x_71', 'x_72', 'x_73', 'x_74', 'x_75', 'x_79', 'x_7a', 'x_7b', 'x_7c', 'x_7d', 'x_7e', 'x_7f', 'x_80', 'x_81', 'x_82', 'x_83', 'x_84', 'x_85', 'x_86', 'x_87', 'x_88', 'x_89', 'x_8a', 'x_8b', 'x_8c', 'x_8d', 'x_8e', 'x_8f', 'x_90', 'x_91', 'x_92', 'x_93', 'x_94', 'x_95', 'x_96', 'x_97', 'x_98', 'x_99', 'x_9a', 'x_9b', 'x_9c', 'x_9d'],'chap01_intro':['x_1', 'x_2'],'chap02_basic':['x_197', 'x_198', 'x_19c', 'x_19d', 'x_19e', 'x_19f', 'x_1a0', 'x_1a1', 'x_1a5', 'x_1ab', 'x_1ac', 'x_1ad', 'x_1ae', 'x_1af', 'x_1b0', 'x_1b1', 'x_1b2', 'x_1b3', 'x_1b4', 'x_1b5', 'x_1b6', 'x_1b7', 'x_1b8', 'x_1b9', 'x_1ba', 'x_1bb', 'x_1bc', 'x_1bd', 'x_1be', 'x_1bf', 'x_1c0', 'x_1c1', 'x_1c2', 'x_1c3', 'x_1c4', 'x_1c5', 'x_1c6', 'x_1c7', 'x_1c8', 'x_1cc', 'x_1cd', 'x_1ce', 'x_1cf', 'x_1d0', 'x_1d1', 'x_1d2', 'x_1d3', 'x_1d4', 'x_1d8', 'x_1d9', 'x_1da', 'x_1db', 'x_1dc', 'x_1dd', 'x_1de', 'x_1df', 'x_1e0', 'x_1e1', 'x_1e2', 'x_1e3', 'x_1e4', 'x_1e5', 'x_1e6', 'x_1e7', 'x_1e8', 'x_1e9', 'x_1ea', 'x_1eb', 'x_1ec', 'x_1ed', 'x_1ee', 'x_1ef', 'x_1f0', 'x_1f1', 'x_1f2', 'x_1f3', 'x_1f4', 'x_1f5', 'x_1f6', 'x_1f7', 'x_1f8', 'x_1f9', 'x_1fa', 'x_1fb','x_1fc'],'chap02_basicoop':['x_c2', 'x_c3', 'x_c4', 'x_c5', 'x_c6', 'x_c7'],'chap03-session4':['x_f', 'x_10', 'x_11', 'x_12', 'x_13', 'x_14', 'x_15', 'x_16', 'x_17', 'x_18', 'x_19', 'x_1a', 'x_1b', 'x_1c', 'x_1d', 'x_1e', 'x_1f', 'x_23', 'x_24', 'x_25', 'x_26', 'x_27', 'x_28', 'x_29', 'x_2a', 'x_2b', 'x_2c', 'x_2d', 'x_2e', 'x_2f', 'x_30', 'x_31', 'x_32', 'x_33', 'x_34', 'x_35', 'x_66', 'x_67', 'x_68', 'x_71', 'x_72', 'x_7c', 'x_7d', 'x_7e', 'x_9b', 'x_9c', 'x_9d', 'x_9e', 'x_9f', 'x_a0', 'x_a1', 'x_a2', 'x_a3', 'x_a4', 'x_a5', 'x_a6', 'x_b8', 'x_b9', 'x_ba', 'x_bb', 'x_bc', 'x_c0', 'x_c1','chap04_handout':['x_204']} diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/sql/comment.mysql.sql --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/comments/sql/comment.mysql.sql Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,2 @@ +alter table comments_comment convert to character set utf8 collate utf8_bin; +alter table comments_comment default character set utf8 collate utf8_bin; diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/sql/element.mysql.sql --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/comments/sql/element.mysql.sql Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,2 @@ +alter table comments_element convert to character set utf8 collate utf8_bin; +alter table comments_element default character set utf8 collate utf8_bin; diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/test --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/comments/test Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,5 @@ +Chapter 9. Finding and fixing mistakes

Chapter 9. Finding and fixing mistakes

+ diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/test~ --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/comments/test~ Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,5 @@ +html>Chapter 9. Finding and fixing mistakes

Chapter 9. Finding and fixing mistakes

+ diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/urls.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/comments/urls.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,8 @@ +from django.conf.urls.defaults import * + +urlpatterns = patterns('', + (r'chapter/(?P[^/]+)/?$', 'hgbook.comments.views.chapter'), + (r'chapter/(?P[^/]+)/count/?$', 'hgbook.comments.views.single_com'), + (r'single/(?P[^/]+)/?$', 'hgbook.comments.views.single'), + (r'submit/(?P[^/]+)/?$', 'hgbook.comments.views.submit') +) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/urls.pyc Binary file SEESenv/web/hgbook/comments/urls.pyc has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/urls.py~ --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/comments/urls.py~ Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,8 @@ +from django.conf.urls.defaults import * + +urlpatterns = patterns('', + (r'chapter/(?P[^/]+)/?$', 'hgbook.comments.views.chapter'), + (r'chapter/(?P[^/]+)/count/?$', 'hgbook.comments.views.single_com'), + (r'single/(?P[^/]+)/?$', 'hgbook.comments.views.single'), + (r'submit/(?P[^/]+)/?$', 'hgbook.comments.views.submit') +) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/views.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/comments/views.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,239 @@ +import sys +import django.forms as forms +from django.db import connection +from django.http import HttpResponse +from hgbook.comments.models import Comment, Element +from django.shortcuts import get_object_or_404, render_to_response +from django.template import Context ,RequestContext +from django.template.loader import get_template +from django.utils.simplejson import dumps +from p_list import * +from BeautifulSoup import BeautifulSoup +import glob +html_folder='/home/amit/SEES-hacks/web/html/' + +def sort_dict(dict): + new_dict = {} + sorted_keys = dict.keys() + sorted_keys.sort() + print >> sys.stderr , sorted_keys + for key in sorted_keys: + new_dict[key] = dict[key] + + return new_dict + +def dump_queries(): + # requires settings.DEBUG to be set to True in order to work + if len(connection.queries) == 1: + print connection.queries + else: + qs = {} + for q in connection.queries: + qs[q['sql']] = qs.setdefault(q['sql'], 0) + 1 + for q in sorted(qs.items(), key=lambda x: x[1], reverse=True): + print q + print len(connection.queries) + +class CommentForm(forms.Form): + id = forms.CharField(widget=forms.HiddenInput) + name = forms.CharField(max_length=64) + url = forms.URLField(max_length=128, required=False) + comment = forms.CharField(widget=forms.Textarea(attrs={ + 'rows': 8, 'cols': 60 + })) + remember = forms.BooleanField(initial=True, required=False) + +def search(request): + print request + + +def index(request): + html_files = glob.glob(html_folder+'ch*.html') + print >> sys.stderr ,html_files + print >> sys.stderr ,"just checking whether i got here" + print >> sys.stderr , html_files + html_files.sort() + link_list=['/review/html/'+a.split('/')[-1] for a in html_files] +# print >> sys.stderr , html_files + title_list=[] + for html_file in html_files: + file_str=open(html_file,'r').read() + soup_obj=BeautifulSoup(''.join(file_str)) + title_str=soup_obj.html.head.title.string + title_str=unicode(title_str) + title_str.encode('ascii','ignore') + title_list.append(title_str) + + resp_dict=zip(link_list,title_list) + print >>sys.stderr ,resp_dict + return render_to_response('index.html', {'resp_dict': resp_dict }) + + + + + + + +def comments_by_chapter(id): + objs = {} + for c in Comment.objects.filter(element=id, hidden=False).order_by('date'): + objs.setdefault(c.element_id, []).append(c) + return objs + +def chapter(request, id): + template = get_template('comment.html') + resp = {} + for elt, comments in comments_by_chapter(id).iteritems(): + + form = CommentForm(initial={ + 'id': elt, + 'name': request.session.get('name', ''), + }) + + + + resp[elt] = template.render(RequestContext({ + 'id': elt, + 'form': form, + 'length': len(comments), + 'query': comments, + })) + return HttpResponse(dumps(resp), mimetype='application/json') + +# queryset = Comment.objects.filter(element=id, hidden=False).order_by('date') +# if form is None: +# form = CommentForm(initial={ +# 'id': id, +# 'name': request.session.get('name', ''), +# }) +# try: +# error = form.errors[0] +# except: +# error = '' +# return render_to_response('comment.html', { +# 'id': id, +# 'form': form, +# 'length': len(queryset), +# 'query': queryset, +# 'newid': newid or True, +# 'error': error, +# }) + + + + + +def single_com(request,id): + +# template = get_template('comment.html') + resp = {} + print >> sys.stderr ,id + for i in p_list[id]: +# form=None + queryset = Comment.objects.filter(element=i, hidden=False).order_by('date') + +# if form is None: +# form = CommentForm(initial={ +# 'id': id, +# 'name': request.session.get('name', ''), +# }) + if len(queryset)==0: + resp[i] = "No" + elif len(queryset)==1: + resp[i]="One" + else: + resp[i] = len(queryset) +# 'id': i, +# 'form': form, +# 'length': len(queryset), +# 'query': queryset, + + return HttpResponse(dumps(resp), mimetype='application/json') +# template = get_template('comment.html') +# for i in p_list[id]: +# form=None +# queryset = Comment.objects.filter(element=i, hidden=False).order_by('date') +# print queryset +# if form is None: +# form = CommentForm(initial={ +# 'id': id, +# 'name': request.session.get('name', ''), +# }) +# try: +# error = form.errors[0] +# except: +# error = '' + +# return HttpResponse(response_list) + + + + + + +def chapter_count(request, id): + + resp = comments_by_chapter(id) + + for elt, comments in resp.iteritems(): + + resp[elt] = len(comments) + + return HttpResponse(dumps(resp), mimetype='application/json') + +def single(request, id, form=None, newid=None ): + queryset = Comment.objects.filter(element=id, hidden=False).order_by('date') + if form is None: + form = CommentForm(initial={ + 'id': id, + 'name': request.session.get('name', ''), + }) + try: + error = form.errors[0] + except: + error = '' + return render_to_response('comment.html', { + 'id': id, + 'form': form, + 'length': len(queryset), + 'query': queryset, + 'newid': newid or True, + 'error': error, + }) + +def submit(request, id): +# print request +# print id + try : + element=Element.objects.get(id=id,chapter='chap_intro') + except Element.DoesNotExist: + element=Element(id=id) + element.save() + + + if request.method == 'POST': + + form = CommentForm(request.POST) + + if form.is_valid(): + data = form.cleaned_data + + if data.get('remember'): + + request.session['name'] = data['name'] + request.session['url'] = data['url'] + else: + request.session.pop('name', None) + request.session.pop('url', None) + c = Comment(element=element, + comment=data['comment'], + submitter_name=data['name'], + submitter_url=data['url'], + ip=request.META.get('REMOTE_ADDR')) + + + c.save() + newid=c.id + form = None + return single(request, id, form, newid) + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/views.pyc Binary file SEESenv/web/hgbook/comments/views.pyc has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/comments/views.py~ --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/comments/views.py~ Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,239 @@ +import sys +import django.forms as forms +from django.db import connection +from django.http import HttpResponse +from hgbook.comments.models import Comment, Element +from django.shortcuts import get_object_or_404, render_to_response +from django.template import Context ,RequestContext +from django.template.loader import get_template +from django.utils.simplejson import dumps +from p_list import * +from BeautifulSoup import BeautifulSoup +import glob +html_folder='/home/amit/SEES-hacks/web/html' + +def sort_dict(dict): + new_dict = {} + sorted_keys = dict.keys() + sorted_keys.sort() + print >> sys.stderr , sorted_keys + for key in sorted_keys: + new_dict[key] = dict[key] + + return new_dict + +def dump_queries(): + # requires settings.DEBUG to be set to True in order to work + if len(connection.queries) == 1: + print connection.queries + else: + qs = {} + for q in connection.queries: + qs[q['sql']] = qs.setdefault(q['sql'], 0) + 1 + for q in sorted(qs.items(), key=lambda x: x[1], reverse=True): + print q + print len(connection.queries) + +class CommentForm(forms.Form): + id = forms.CharField(widget=forms.HiddenInput) + name = forms.CharField(max_length=64) + url = forms.URLField(max_length=128, required=False) + comment = forms.CharField(widget=forms.Textarea(attrs={ + 'rows': 8, 'cols': 60 + })) + remember = forms.BooleanField(initial=True, required=False) + +def search(request): + print request + + +def index(request): + html_files = glob.glob(html_folder+'ch*.html') + print >> sys.stderr ,html_files + print >> sys.stderr ,"just checking whether i got here" + print >> sys.stderr , html_files + html_files.sort() + link_list=['/review/html/'+a.split('/')[-1] for a in html_files] +# print >> sys.stderr , html_files + title_list=[] + for html_file in html_files: + file_str=open(html_file,'r').read() + soup_obj=BeautifulSoup(''.join(file_str)) + title_str=soup_obj.html.head.title.string + title_str=unicode(title_str) + title_str.encode('ascii','ignore') + title_list.append(title_str) + + resp_dict=zip(link_list,title_list) + print >>sys.stderr ,resp_dict + return render_to_response('index.html', {'resp_dict': resp_dict }) + + + + + + + +def comments_by_chapter(id): + objs = {} + for c in Comment.objects.filter(element=id, hidden=False).order_by('date'): + objs.setdefault(c.element_id, []).append(c) + return objs + +def chapter(request, id): + template = get_template('comment.html') + resp = {} + for elt, comments in comments_by_chapter(id).iteritems(): + + form = CommentForm(initial={ + 'id': elt, + 'name': request.session.get('name', ''), + }) + + + + resp[elt] = template.render(RequestContext({ + 'id': elt, + 'form': form, + 'length': len(comments), + 'query': comments, + })) + return HttpResponse(dumps(resp), mimetype='application/json') + +# queryset = Comment.objects.filter(element=id, hidden=False).order_by('date') +# if form is None: +# form = CommentForm(initial={ +# 'id': id, +# 'name': request.session.get('name', ''), +# }) +# try: +# error = form.errors[0] +# except: +# error = '' +# return render_to_response('comment.html', { +# 'id': id, +# 'form': form, +# 'length': len(queryset), +# 'query': queryset, +# 'newid': newid or True, +# 'error': error, +# }) + + + + + +def single_com(request,id): + +# template = get_template('comment.html') + resp = {} + print >> sys.stderr ,id + for i in p_list[id]: +# form=None + queryset = Comment.objects.filter(element=i, hidden=False).order_by('date') + +# if form is None: +# form = CommentForm(initial={ +# 'id': id, +# 'name': request.session.get('name', ''), +# }) + if len(queryset)==0: + resp[i] = "No" + elif len(queryset)==1: + resp[i]="One" + else: + resp[i] = len(queryset) +# 'id': i, +# 'form': form, +# 'length': len(queryset), +# 'query': queryset, + + return HttpResponse(dumps(resp), mimetype='application/json') +# template = get_template('comment.html') +# for i in p_list[id]: +# form=None +# queryset = Comment.objects.filter(element=i, hidden=False).order_by('date') +# print queryset +# if form is None: +# form = CommentForm(initial={ +# 'id': id, +# 'name': request.session.get('name', ''), +# }) +# try: +# error = form.errors[0] +# except: +# error = '' + +# return HttpResponse(response_list) + + + + + + +def chapter_count(request, id): + + resp = comments_by_chapter(id) + + for elt, comments in resp.iteritems(): + + resp[elt] = len(comments) + + return HttpResponse(dumps(resp), mimetype='application/json') + +def single(request, id, form=None, newid=None ): + queryset = Comment.objects.filter(element=id, hidden=False).order_by('date') + if form is None: + form = CommentForm(initial={ + 'id': id, + 'name': request.session.get('name', ''), + }) + try: + error = form.errors[0] + except: + error = '' + return render_to_response('comment.html', { + 'id': id, + 'form': form, + 'length': len(queryset), + 'query': queryset, + 'newid': newid or True, + 'error': error, + }) + +def submit(request, id): +# print request +# print id + try : + element=Element.objects.get(id=id,chapter='chap_intro') + except Element.DoesNotExist: + element=Element(id=id) + element.save() + + + if request.method == 'POST': + + form = CommentForm(request.POST) + + if form.is_valid(): + data = form.cleaned_data + + if data.get('remember'): + + request.session['name'] = data['name'] + request.session['url'] = data['url'] + else: + request.session.pop('name', None) + request.session.pop('url', None) + c = Comment(element=element, + comment=data['comment'], + submitter_name=data['name'], + submitter_url=data['url'], + ip=request.META.get('REMOTE_ADDR')) + + + c.save() + newid=c.id + form = None + return single(request, id, form, newid) + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/converter.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/converter.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,52 @@ +from lxml import etree +from lxml import html +from lxml.cssselect import CSSSelector +import md5 +import sys + + +args = sys.argv[1:] + +# django stuff +from django.core.management import setup_environ +import settings # Assumed to be in the same directory. +setup_environ(settings) # ugly django collateral effects :( +from comments.models import Element + +doc_id = 'MMSC' +sel = CSSSelector('div.chapter p, pre, h1, table.equation') +chapter_sel = CSSSelector('div.chapter') + +try: + filename = args[0] +except IndexError: + raise IndexError("Usage: %s " % __file__) + +tree = etree.parse(filename, html.HTMLParser(remove_blank_text=True)) +root = tree.getroot() + +chapter = chapter_sel(root)[0] +chapter_title = chapter.get('id').split(':')[1] +chapter_hash = md5.new(chapter.get('id').encode('utf8')).hexdigest() + +chapter.set('id', chapter_hash) + +for element in sel(root): + hsh_source = element.text or element.get('alt') or etree.tostring(element) + + if hsh_source: + hsh_source_encoded = hsh_source.encode('utf8') + hsh = md5.new(hsh_source_encoded).hexdigest() + element.set('id', '%s-%s' % (chapter_hash, hsh)) + + # create the commentable element in the DB + e = Element() + e.id = '%s-%s' % (chapter_hash, hsh) + e.chapter = chapter_hash + e.title = chapter_title + e.save() + + + +print etree.tostring(root) # pipe to a file if you wish + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/dbutil.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/dbutil.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,32 @@ +import MySQLdb as mysql +import sys + +def connect(): + try: + import secrets + except ImportError: + print >> sys.stderr, 'Decrypt secrets.py.gpg or create a new copy!' + sys.exit(1) + + if secrets.DATABASE_ENGINE != 'mysql': + print >> sys.stderr, ('You are using a %s database' % + secrets.DATABASE_ENGINE) + sys.exit(1) + + kwargs = { + 'charset': 'utf8', + 'use_unicode': True, + } + if secrets.DATABASE_USER: + kwargs['user'] = secrets.DATABASE_USER + if secrets.DATABASE_NAME: + kwargs['db'] = secrets.DATABASE_NAME + if secrets.DATABASE_PASSWORD: + kwargs['passwd'] = secrets.DATABASE_PASSWORD + if secrets.DATABASE_HOST.startswith('/'): + kwargs['unix_socket'] = secrets.DATABASE_HOST + elif secrets.DATABASE_HOST: + kwargs['host'] = secrets.DATABASE_HOST + if secrets.DATABASE_PORT: + kwargs['port'] = int(secrets.DATABASE_PORT) + return mysql.connect(**kwargs) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/flintlock diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/iamflint Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/iamflint has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/position.DB Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/position.DB has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/position.baseA Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/position.baseA has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/position.baseB Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/position.baseB has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/postlist.DB Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/postlist.DB has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/postlist.baseA Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/postlist.baseA has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/postlist.baseB Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/postlist.baseB has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/record.DB Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/record.DB has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/record.baseA Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/record.baseA has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/record.baseB Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/record.baseB has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/spelling.DB Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/spelling.DB has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/spelling.baseA Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/spelling.baseA has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/spelling.baseB Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/spelling.baseB has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/termlist.DB Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/termlist.DB has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/termlist.baseA Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/termlist.baseA has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/termlist.baseB Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/termlist.baseB has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/value.DB Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/value.DB has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/value.baseA Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/value.baseA has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/value.baseB Binary file SEESenv/web/hgbook/flatpages/flatpage/hgbook.comments.models.flatpageindexer/value.baseB has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/load_elements.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/load_elements.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,18 @@ +#!/usr/bin/env python +# +# This script updates the contents of the comments_element table. +# It's fugly, but a lot less painful than trying to use Django's +# fixtures system. + +import os, sys +sys.path.append(os.path.dirname(__file__)) +import dbutil + +os.system('make -C ../../en all-ids.dat') + +conn = dbutil.connect() +c = conn.cursor() +c.execute('''load data local infile "../../en/all-ids.dat" replace + into table comments_element + fields terminated by "|"''') +print 'Database updated' diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/manage.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/manage.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,11 @@ +#!/usr/bin/env python +from django.core.management import execute_manager +try: + import settings # Assumed to be in the same directory. +except ImportError: + import sys + sys.stderr.write("Error: Can't find the file 'settings.py' in the directory containing %r. It appears you've customized things.\nYou'll have to run django-admin.py, passing it your settings module.\n(If the file settings.py does indeed exist, it's causing an ImportError somehow.)\n" % __file__) + sys.exit(1) + +if __name__ == "__main__": + execute_manager(settings) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/reviewers.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/reviewers.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,81 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- + +import os, sys +sys.path.append(os.path.dirname(__file__)) +import dbutil + +conn = dbutil.connect() +c = conn.cursor() + +c.execute('''select submitter_name from comments_comment''') + +reviewers = {} + +mappings = { + u'alejandro "tab-lover" dubrovsky': u'Alejandro Dubrovsky', + u'alex hirzel ': u'Alex Hirzel', + u'anonymous coward': u'Anonymous', + u'arthur van leeuwen': u'Arthur van Leeuwen', + u'augustss': u'Lennart Augustsson', + u'ed t': u'Anonymous', + u'geogre moschovitis': u'George Moschovitis', + u'george m': u'George Moschovitis', + u'haskell newb': u'Anonymous', + u'j. pablo fernandez': u'J. Pablo Fernández', + u'kamal al-marhoobi': u'Kamal Al-Marhubi', + u'kevin w.': u'Kevin Watters', + u'max cantor (#haskell - mxc)': u'Max Cantor', + u'michael campbell': u'Michael Campbell', + u'mike btauwerman': u'Mike Brauwerman', + u'no credit necessary': u'Anonymous', + u'nykänen, matti': u'Matti Nykänen', + u'omar antolin camarena': u'Omar Antolín Camarena', + u'ryan t mulligan': u'Ryan T. Mulligan', + u'sengan baring-gould': u'Sengan Baring-Gould', + u'some guy': u'Anonymous', + u'tomas janousek': u'Tomáš Janoušek', + u'william halchin': u'William N. Halchin', + } + +def fixup(s): + try: + return s.encode('ascii') + except UnicodeEncodeError: + def f(c): + o = ord(c) + if o < 128: + return c + return '&#%d;' % o + return ''.join(map(f, s)) + +total = 0 +for r in c.fetchall(): + r = r[0].decode('utf-8') + if r in ("Bryan O'Sullivan",): + continue + total += 1 + m = mappings.get(r.lower()) + if m: + r = m + elif len(r) < 2 or ' ' not in r: + r = 'Anonymous' + reviewers.setdefault(r, 0) + reviewers[r] += 1 + +reviewers = sorted(reviewers.iteritems(), key=lambda x: x[0]) + +cohorts = [(.01,1),(.002,.01)] + +for (lo,hi) in cohorts: + lo = total * lo + hi = total * hi + for r in [n for n in reviewers if lo <= n[1] < hi]: + if r[1] > 3: + print '%s,' % fixup(r[0]) + print + +lo = total * .002 +for n in reviewers: + if n[1] < lo: + print '%s,' % fixup(n[0]) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/run.wsgi --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/run.wsgi Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,12 @@ +import os +import sys +import django +sys.path.append('/'.join(os.getcwd().split('/')[:-1])') +sys.path.append(os.path.join(os.getcwd(),'comments')) +os.environ['DJANGO_SETTINGS_MODULE'] = 'hgbook.settings' +import django.core.handlers.wsgi +application = django.core.handlers.wsgi.WSGIHandler() + + + + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/run.wsgi~ --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/run.wsgi~ Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,12 @@ +import os +import sys +import django +sys.path.append('/'.join(os.getcwd().split('/')[:-1])') +sys.path.append(os.join(os.getcwd().split('/'),comments)) +os.environ['DJANGO_SETTINGS_MODULE'] = 'hgbook.settings' +import django.core.handlers.wsgi +application = django.core.handlers.wsgi.WSGIHandler() + + + + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/secrets.py.gpg Binary file SEESenv/web/hgbook/secrets.py.gpg has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/settings.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/settings.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,99 @@ +# Django settings for hgbook project. + +import os, sys + +DEBUG = True +TEMPLATE_DEBUG = DEBUG + +ADMINS = ( + ("Bryan O'Sullivan", 'bos@serpentine.com'), +) + +MANAGERS = ADMINS + +ROOT = os.path.dirname(sys.modules[__name__].__file__) + +DJAPIAN_DATABASE_PATH = '/home/amit/hgbook-alqua/web/hgbook' +DJAPIAN_STEMMING_LANG = 'en' + + +try: + from secrets import DATABASE_ENGINE, DATABASE_NAME, DATABASE_USER, \ + DATABASE_PASSWORD, DATABASE_HOST, DATABASE_PORT, SECRET_KEY +except ImportError: + print >> sys.stderr, 'Faking up some database configuration for you' + DATABASE_ENGINE = 'sqlite3' + DATABASE_NAME = os.path.join(ROOT, '.database.sqlite3') + DATABASE_USER = '' + DATABASE_PASSWORD = '' + DATABASE_HOST = '' + DATABASE_PORT = '' + SECRET_KEY = '' + +# Local time zone for this installation. Choices can be found here: +# http://en.wikipedia.org/wiki/List_of_tz_zones_by_name +# although not all choices may be avilable on all operating systems. +# If running in a Windows environment this must be set to the same as your +# system time zone. +TIME_ZONE = 'America/Los_Angeles' + +# Language code for this installation. All choices can be found here: +# http://www.w3.org/TR/REC-html40/struct/dirlang.html#langcodes +LANGUAGE_CODE = 'en-us' + +SITE_ID = 1 + +# If you set this to False, Django will make some optimizations so as not +# to load the internationalization machinery. +USE_I18N = True + +# Absolute path to the directory that holds media. +# Example: "/home/media/media.lawrence.com/" +MEDIA_ROOT = '' + +# URL that handles the media served from MEDIA_ROOT. Make sure to use a +# trailing slash if there is a path component (optional in other cases). +# Examples: "http://media.lawrence.com", "http://example.com/media/" +MEDIA_URL = '' + +# URL prefix for admin media -- CSS, JavaScript and images. Make sure to use a +# trailing slash. +# Examples: "http://foo.com/media/", "/media/". +ADMIN_MEDIA_PREFIX = '/media/' + + +#Index Directories for searching static html files +DD_SEARCH_INDEX_DIR = '../html/' + + + +# List of callables that know how to import templates from various sources. +TEMPLATE_LOADERS = ( + 'django.template.loaders.filesystem.load_template_source', + 'django.template.loaders.app_directories.load_template_source', +# 'django.template.loaders.eggs.load_template_source', +) + +MIDDLEWARE_CLASSES = ( + 'django.middleware.common.CommonMiddleware', + 'django.contrib.sessions.middleware.SessionMiddleware', + 'django.contrib.auth.middleware.AuthenticationMiddleware', + 'django.middleware.doc.XViewMiddleware', + 'django.contrib.flatpages.middleware.FlatpageFallbackMiddleware' +) + +ROOT_URLCONF = 'hgbook.urls' + +TEMPLATE_DIRS = ( + os.path.join(ROOT, 'templates') +) + +INSTALLED_APPS = ( + 'django.contrib.admin', + 'django.contrib.auth', + 'django.contrib.contenttypes', + 'django.contrib.sessions', + 'django.contrib.sites', + 'django.contrib.flatpages', + 'hgbook.comments', +) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/settings.pyc Binary file SEESenv/web/hgbook/settings.pyc has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/settings.py~ --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/settings.py~ Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,99 @@ +# Django settings for hgbook project. + +import os, sys + +DEBUG = True +TEMPLATE_DEBUG = DEBUG + +ADMINS = ( + ("Bryan O'Sullivan", 'bos@serpentine.com'), +) + +MANAGERS = ADMINS + +ROOT = os.path.dirname(sys.modules[__name__].__file__) + +DJAPIAN_DATABASE_PATH = '/path/to/my/project/data/djapian/' +DJAPIAN_STEMMING_LANG = 'en' + + +try: + from secrets import DATABASE_ENGINE, DATABASE_NAME, DATABASE_USER, \ + DATABASE_PASSWORD, DATABASE_HOST, DATABASE_PORT, SECRET_KEY +except ImportError: + print >> sys.stderr, 'Faking up some database configuration for you' + DATABASE_ENGINE = 'sqlite3' + DATABASE_NAME = os.path.join(ROOT, '.database.sqlite3') + DATABASE_USER = '' + DATABASE_PASSWORD = '' + DATABASE_HOST = '' + DATABASE_PORT = '' + SECRET_KEY = '' + +# Local time zone for this installation. Choices can be found here: +# http://en.wikipedia.org/wiki/List_of_tz_zones_by_name +# although not all choices may be avilable on all operating systems. +# If running in a Windows environment this must be set to the same as your +# system time zone. +TIME_ZONE = 'America/Los_Angeles' + +# Language code for this installation. All choices can be found here: +# http://www.w3.org/TR/REC-html40/struct/dirlang.html#langcodes +LANGUAGE_CODE = 'en-us' + +SITE_ID = 1 + +# If you set this to False, Django will make some optimizations so as not +# to load the internationalization machinery. +USE_I18N = True + +# Absolute path to the directory that holds media. +# Example: "/home/media/media.lawrence.com/" +MEDIA_ROOT = '' + +# URL that handles the media served from MEDIA_ROOT. Make sure to use a +# trailing slash if there is a path component (optional in other cases). +# Examples: "http://media.lawrence.com", "http://example.com/media/" +MEDIA_URL = '' + +# URL prefix for admin media -- CSS, JavaScript and images. Make sure to use a +# trailing slash. +# Examples: "http://foo.com/media/", "/media/". +ADMIN_MEDIA_PREFIX = '/media/' + + +#Index Directories for searching static html files +DD_SEARCH_INDEX_DIR = '../html/' + + + +# List of callables that know how to import templates from various sources. +TEMPLATE_LOADERS = ( + 'django.template.loaders.filesystem.load_template_source', + 'django.template.loaders.app_directories.load_template_source', +# 'django.template.loaders.eggs.load_template_source', +) + +MIDDLEWARE_CLASSES = ( + 'django.middleware.common.CommonMiddleware', + 'django.contrib.sessions.middleware.SessionMiddleware', + 'django.contrib.auth.middleware.AuthenticationMiddleware', + 'django.middleware.doc.XViewMiddleware', + 'django.contrib.flatpages.middleware.FlatpageFallbackMiddleware' +) + +ROOT_URLCONF = 'hgbook.urls' + +TEMPLATE_DIRS = ( + os.path.join(ROOT, 'templates') +) + +INSTALLED_APPS = ( + 'django.contrib.admin', + 'django.contrib.auth', + 'django.contrib.contenttypes', + 'django.contrib.sessions', + 'django.contrib.sites', + 'django.contrib.flatpages', + 'hgbook.comments', +) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/templates/404.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/templates/404.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,8 @@ +{% extends "simple.html" %} + +{% block title %}Page Not Found{% endblock %} + +{% block body %} +

Sorry, we hit when trying to find the +page you requested.

+{% endblock %} diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/templates/500.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/templates/500.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,11 @@ +{% extends "simple.html" %} + +{% block title %}Internal Server Error{% endblock %} + +{% block body %} +

Sorry, we hit when +trying to process your request. If possible, please let Bryan know that this problem happened, +and what you were doing when it occurred.

+{% endblock %} diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/templates/boilerplate.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/templates/boilerplate.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,37 @@ + + + + {% block pagetitle %}Mercurial: The Definitive Guide{% endblock %} + + + + + + + + + + {% block bodycontent %}{% endblock %} + +

Want to stay + up to date? Subscribe to comment feeds for any chapter, or + the entire book.

Copyright + 2006, 2007, 2008, 2009 Bryan O'Sullivan. + Icons by + Paul Davey aka Mattahan.

+
+ + + + + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/templates/comment.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/templates/comment.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,62 @@ +{% ifequal length 1 %} + One comment +{% else %} + {% if length %} + {{ length }} comments + {% else %} + No comments + + {% endif %} +{% endifequal %} +{% for c in query %} + +{% endfor %} + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/templates/comment.html~ --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/templates/comment.html~ Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,62 @@ +{% ifequal length 1 %} + One comment +{% else %} + {% if length %} + {{ length }} comments + {% else %} + No comments + + {% endif %} +{% endifequal %} +{% for c in query %} + +{% endfor %} + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/templates/feeds/comments_description.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/templates/feeds/comments_description.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,12 @@ +

On {{ obj.date|date:"Y-m-d" }}, + {% if obj.submitter_url %} + {{ obj.submitter_name|escape }} + {% else %} + {{ obj.submitter_name|escape }} + {% endif %} +commented on “{{ obj.element.title|escape }}”:

+
+{{ obj.comment|escape|linebreaks }} +
+

To see this comment in context or to respond, visit {{ site.domain }}

diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/templates/feeds/comments_title.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/templates/feeds/comments_title.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1 @@ +Comment on “{{ obj.element.title|escape }}” diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/templates/index.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/templates/index.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,42 @@ + + + Software Engineering for Engineers and Scientists + + + + + + + + + + + + +
+{% if resp_dict %} +{% for key,value in resp_dict %} +
  • +{{ value }} +
  • +{% endfor %} +{% endif %} + + + + +
    + + + + + + + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/templates/index.html~ --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/templates/index.html~ Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,42 @@ + + + Software Engineering for Engineers and Scientists + + + + + + + + + + + + +
    +{% if resp_dict %} +{% for item in resp_dict %} +
  • +{{ item(1) }} +
  • +{% endfor %} +{% endif %} + + + + +
    + + + + + + + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/templates/simple.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/templates/simple.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,7 @@ +{% extends "boilerplate.html" %} + +{% block bodycontent %} + + +
    {% block body %}{% endblock %}
    +{% endblock %} diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/templates/single_com.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/templates/single_com.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,19 @@ +{% ifequal length 1 %} + One comment +{% else %} + {% if length %} + {{ length }} comments + {% else %} + No comments + + {% endif %} +{% endifequal %} + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/templates/single_com.html~ --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/templates/single_com.html~ Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,19 @@ +{% ifequal length 1 %} + One comment +{% else %} + {% if length %} + {{ length }} comments + {% else %} + No comments + + {% endif %} +{% endifequal %} +{% for c in query %} diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/urls.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/urls.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,31 @@ +import os, sys +from django.conf.urls.defaults import * +import hgbook.comments.feeds as feeds +from django.contrib import admin + +admin.autodiscover() + +feeds = { + 'comments': feeds.Comments, + } + +urlpatterns = patterns('', + (r'^comments/', include('hgbook.comments.urls')), + + (r'^feeds/(?P.*)/$', 'django.contrib.syndication.views.feed', + {'feed_dict': feeds}), + (r'^$', 'hgbook.comments.views.index'), + + (r'^search/', 'hgbook.comments.views.search'), + + # Only uncomment this for local testing without Apache. + (r'^html/(?P.*)$', 'django.views.static.serve', + {'document_root': os.path.realpath(os.path.dirname( + sys.modules[__name__].__file__) + '/../html/')}), + (r'^support/(?P.*)$', 'django.views.static.serve', + {'document_root': os.path.realpath(os.path.dirname( + sys.modules[__name__].__file__) + '/../support/')}), + + # Uncomment this for admin: + (r'^admin/(.*)', admin.site.root), +) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/urls.pyc Binary file SEESenv/web/hgbook/urls.pyc has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/hgbook/urls.py~ --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/hgbook/urls.py~ Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,31 @@ +import os, sys +from django.conf.urls.defaults import * +import hgbook.comments.feeds as feeds +from django.contrib import admin + +admin.autodiscover() + +feeds = { + 'comments': feeds.Comments, + } + +urlpatterns = patterns('', + (r'^comments/', include('hgbook.comments.urls')), + + (r'^feeds/(?P.*)/$', 'django.contrib.syndication.views.feed', + {'feed_dict': feeds}), + (r'^/$', 'hgbook.comments.views.index'), + + (r'^search/', 'hgbook.comments.views.search'), + + # Only uncomment this for local testing without Apache. + (r'^html/(?P.*)$', 'django.views.static.serve', + {'document_root': os.path.realpath(os.path.dirname( + sys.modules[__name__].__file__) + '/../html/')}), + (r'^support/(?P.*)$', 'django.views.static.serve', + {'document_root': os.path.realpath(os.path.dirname( + sys.modules[__name__].__file__) + '/../support/')}), + + # Uncomment this for admin: + (r'^admin/(.*)', admin.site.root), +) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/backup/abc.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/backup/abc.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,530 @@ + +Chapter 9. Finding and fixing mistakes + + + + + + + + + + + +
    +

    +Chapter 14. Adding functionality with extensions

    + +

    While the core of Mercurial is quite complete from a + functionality standpoint, it's deliberately shorn of fancy + features. This approach of preserving simplicity keeps the + software easy to deal with for both maintainers and users.

    +

    However, Mercurial doesn't box you in with an inflexible + command set: you can add features to it as + extensions (sometimes known as + plugins). We've already discussed a few of + these extensions in earlier chapters.

    +

    When you provide a directory name, Mercurial will interpret + this as “operate on every file in this directory and its + subdirectories”. Mercurial traverses the files and + subdirectories in a directory in alphabetical order. When it + encounters a subdirectory, it will traverse that subdirectory + before continuing with the current directory.

    + + + +
    +

    In this chapter, we'll cover some of the other extensions that + are available for Mercurial, and briefly touch on some of the + machinery you'll need to know about if you want to write an + extension of your own.

    +
    +
    +

    +14.1. Improve performance with the inotify extension

    +

    Are you interested in having some of the most common + Mercurial operations run as much as a hundred times faster? + Read on!

    +

    Mercurial has great performance under normal circumstances. + For example, when you run the hg + status command, Mercurial has to scan almost every + directory and file in your repository so that it can display + file status. Many other Mercurial commands need to do the same + work behind the scenes; for example, the hg diff command uses the status + machinery to avoid doing an expensive comparison operation on + files that obviously haven't changed.

    +

    Because obtaining file status is crucial to good + performance, the authors of Mercurial have optimised this code + to within an inch of its life. However, there's no avoiding the + fact that when you run hg + status, Mercurial is going to have to perform at + least one expensive system call for each managed file to + determine whether it's changed since the last time Mercurial + checked. For a sufficiently large repository, this can take a + long time.

    +

    To put a number on the magnitude of this effect, I created a + repository containing 150,000 managed files. I timed hg status as taking ten seconds to + run, even when none of those files had been + modified.

    +

    Many modern operating systems contain a file notification + facility. If a program signs up to an appropriate service, the + operating system will notify it every time a file of interest is + created, modified, or deleted. On Linux systems, the kernel + component that does this is called + inotify.

    +

    Mercurial's inotify + extension talks to the kernel's inotify + component to optimise hg status + commands. The extension has two components. A daemon sits in + the background and receives notifications from the + inotify subsystem. It also listens for + connections from a regular Mercurial command. The extension + modifies Mercurial's behavior so that instead of scanning the + filesystem, it queries the daemon. Since the daemon has perfect + information about the state of the repository, it can respond + with a result instantaneously, avoiding the need to scan every + directory and file in the repository.

    +

    Recall the ten seconds that I measured plain Mercurial as + taking to run hg status on a + 150,000 file repository. With the inotify extension enabled, the time + dropped to 0.1 seconds, a factor of one + hundred faster.

    +

    Before we continue, please pay attention to some + caveats.

    +
      +
    • The inotify + extension is Linux-specific. Because it interfaces directly + to the Linux kernel's inotify subsystem, + it does not work on other operating systems.

    • +
    • It should work on any Linux distribution that + was released after early 2005. Older distributions are + likely to have a kernel that lacks + inotify, or a version of + glibc that does not have the necessary + interfacing support.

    • +
    • Not all filesystems are suitable for use with + the inotify extension. + Network filesystems such as NFS are a non-starter, for + example, particularly if you're running Mercurial on several + systems, all mounting the same network filesystem. The + kernel's inotify system has no way of + knowing about changes made on another system. Most local + filesystems (e.g. ext3, XFS, ReiserFS) should work + fine.

    • +
    +

    The inotify extension is + not yet shipped with Mercurial as of May 2007, so it's a little + more involved to set up than other extensions. But the + performance improvement is worth it!

    +

    The extension currently comes in two parts: a set of patches + to the Mercurial source code, and a library of Python bindings + to the inotify subsystem.

    +
    + + + + + +
    [Note]Note

    There are two Python + inotify binding libraries. One of them is + called pyinotify, and is packaged by some + Linux distributions as python-inotify. + This is not the one you'll need, as it is + too buggy and inefficient to be practical.

    +

    To get going, it's best to already have a functioning copy + of Mercurial installed.

    +
    + + + + + +
    [Note]Note

    If you follow the instructions below, you'll be + replacing and overwriting any existing + installation of Mercurial that you might already have, using + the latest bleeding edge Mercurial code. Don't + say you weren't warned!

    +
      +
    1. +

      Clone the Python inotify + binding repository. Build and install it.

      +
      hg clone http://hg.kublai.com/python/inotify
      +cd inotify
      +python setup.py build --force
      +sudo python setup.py install --skip-build
      +
    2. +
    3. +

      Clone the crew Mercurial repository. + Clone the inotify patch + repository so that Mercurial Queues will be able to apply + patches to your cope of the crew repository.

      +
      hg clone http://hg.intevation.org/mercurial/crew
      +hg clone crew inotify
      +hg clone http://hg.kublai.com/mercurial/patches/inotify inotify/.hg/patches
      +
    4. +
    5. Make sure that you have the Mercurial Queues + extension, mq, enabled. If + you've never used MQ, read Section 12.5, “Getting started with Mercurial Queues” to get started + quickly.

    6. +
    7. +

      Go into the inotify repo, and apply all + of the inotify patches + using the hg + -a option to the qpush command.

      +
      cd inotify
      +hg qpush -a
      +
    8. +
    9. If you get an error message from qpush, you should not continue. + Instead, ask for help.

    10. +
    11. +

      Build and install the patched version of + Mercurial.

      +
      python setup.py build --force
      +sudo python setup.py install --skip-build
      +
    12. +
    +

    Once you've build a suitably patched version of Mercurial, + all you need to do to enable the inotify extension is add an entry to + your ~/.hgrc.

    +
    [extensions] inotify =
    +

    When the inotify extension + is enabled, Mercurial will automatically and transparently start + the status daemon the first time you run a command that needs + status in a repository. It runs one status daemon per + repository.

    +

    The status daemon is started silently, and runs in the + background. If you look at a list of running processes after + you've enabled the inotify + extension and run a few commands in different repositories, + you'll thus see a few hg processes sitting + around, waiting for updates from the kernel and queries from + Mercurial.

    +

    The first time you run a Mercurial command in a repository + when you have the inotify + extension enabled, it will run with about the same performance + as a normal Mercurial command. This is because the status + daemon needs to perform a normal status scan so that it has a + baseline against which to apply later updates from the kernel. + However, every subsequent command that does + any kind of status check should be noticeably faster on + repositories of even fairly modest size. Better yet, the bigger + your repository is, the greater a performance advantage you'll + see. The inotify daemon makes + status operations almost instantaneous on repositories of all + sizes!

    +

    If you like, you can manually start a status daemon using + the inserve command. + This gives you slightly finer control over how the daemon ought + to run. This command will of course only be available when the + inotify extension is + enabled.

    +

    When you're using the inotify extension, you should notice + no difference at all in Mercurial's + behavior, with the sole exception of status-related commands + running a whole lot faster than they used to. You should + specifically expect that commands will not print different + output; neither should they give different results. If either of + these situations occurs, please report a bug.

    +
    +
    +

    +14.2. Flexible diff support with the extdiff extension

    +

    Mercurial's built-in hg + diff command outputs plaintext unified diffs.

    +
    $ hg diff
    +diff -r 80997726a0ea myfile
    +--- a/myfile	Wed Jan 06 06:50:18 2010 +0000
    ++++ b/myfile	Wed Jan 06 06:50:18 2010 +0000
    +@@ -1,1 +1,2 @@
    + The first line.
    ++The second line.
    +
    +

    If you would like to use an external tool to display + modifications, you'll want to use the extdiff extension. This will let you + use, for example, a graphical diff tool.

    +

    The extdiff extension is + bundled with Mercurial, so it's easy to set up. In the extensions section of your + ~/.hgrc, simply add a + one-line entry to enable the extension.

    +
    [extensions]
    +extdiff =
    +

    This introduces a command named extdiff, which by default uses + your system's diff command to generate a + unified diff in the same form as the built-in hg diff command.

    +
    $ hg extdiff
    +--- a.80997726a0ea/myfile	2010-01-06 06:50:18.613674526 +0000
    ++++ /tmp/extdiffNErQlu/a/myfile	2010-01-06 06:50:18.437687076 +0000
    +@@ -1 +1,2 @@
    + The first line.
    ++The second line.
    +
    +

    The result won't be exactly the same as with the built-in + hg diff variations, because the + output of diff varies from one system to + another, even when passed the same options.

    +

    As the making snapshot + lines of output above imply, the extdiff command works by + creating two snapshots of your source tree. The first snapshot + is of the source revision; the second, of the target revision or + working directory. The extdiff command generates + these snapshots in a temporary directory, passes the name of + each directory to an external diff viewer, then deletes the + temporary directory. For efficiency, it only snapshots the + directories and files that have changed between the two + revisions.

    +

    Snapshot directory names have the same base name as your + repository. If your repository path is /quux/bar/foo, then foo will be the name of each + snapshot directory. Each snapshot directory name has its + changeset ID appended, if appropriate. If a snapshot is of + revision a631aca1083f, the directory will be + named foo.a631aca1083f. + A snapshot of the working directory won't have a changeset ID + appended, so it would just be foo in this example. To see what + this looks like in practice, look again at the extdiff example above. Notice + that the diff has the snapshot directory names embedded in its + header.

    +

    The extdiff command + accepts two important options. The hg -p option + lets you choose a program to view differences with, instead of + diff. With the hg -o option, + you can change the options that extdiff passes to the program + (by default, these options are + -Npru, which only make sense + if you're running diff). In other respects, + the extdiff command + acts similarly to the built-in hg + diff command: you use the same option names, syntax, + and arguments to specify the revisions you want, the files you + want, and so on.

    +

    As an example, here's how to run the normal system + diff command, getting it to generate context + diffs (using the -c option) + instead of unified diffs, and five lines of context instead of + the default three (passing 5 as the argument + to the -C option).

    +
    $ hg extdiff -o -NprcC5
    +*** a.80997726a0ea/myfile	Wed Jan  6 06:50:18 2010
    +--- /tmp/extdiffNErQlu/a/myfile	Wed Jan  6 06:50:18 2010
    +***************
    +*** 1 ****
    +--- 1,2 ----
    +  The first line.
    ++ The second line.
    +
    +

    Launching a visual diff tool is just as easy. Here's how to + launch the kdiff3 viewer.

    +
    hg extdiff -p kdiff3 -o
    +

    If your diff viewing command can't deal with directories, + you can easily work around this with a little scripting. For an + example of such scripting in action with the mq extension and the + interdiff command, see Section 13.9.2, “Viewing the history of a patch”.

    +
    +

    +14.2.1. Defining command aliases

    +

    It can be cumbersome to remember the options to both the + extdiff command and + the diff viewer you want to use, so the extdiff extension lets you define + new commands that will invoke your diff + viewer with exactly the right options.

    +

    All you need to do is edit your ~/.hgrc, and add a section named + extdiff. Inside this + section, you can define multiple commands. Here's how to add + a kdiff3 command. Once you've defined + this, you can type hg kdiff3 + and the extdiff extension + will run kdiff3 for you.

    +
    [extdiff]
    +cmd.kdiff3 =
    +

    If you leave the right hand side of the definition empty, + as above, the extdiff + extension uses the name of the command you defined as the name + of the external program to run. But these names don't have to + be the same. Here, we define a command named + hg wibble, which runs + kdiff3.

    +
    [extdiff]
    + cmd.wibble = kdiff3
    +

    You can also specify the default options that you want to + invoke your diff viewing program with. The prefix to use is + opts., followed by the name + of the command to which the options apply. This example + defines a hg vimdiff command + that runs the vim editor's + DirDiff extension.

    +
    [extdiff]
    + cmd.vimdiff = vim
    +opts.vimdiff = -f '+next' '+execute "DirDiff" argv(0) argv(1)'
    +
    +
    +
    +

    +14.3. Cherrypicking changes with the transplant extension

    +

    Need to have a long chat with Brendan about this.

    +
    +
    +

    +14.4. Send changes via email with the patchbomb extension

    +

    Many projects have a culture of change + review, in which people send their modifications to a + mailing list for others to read and comment on before they + commit the final version to a shared repository. Some projects + have people who act as gatekeepers; they apply changes from + other people to a repository to which those others don't have + access.

    +

    Mercurial makes it easy to send changes over email for + review or application, via its patchbomb extension. The extension is + so named because changes are formatted as patches, and it's usual + to send one changeset per email message. Sending a long series + of changes by email is thus much like bombing the + recipient's inbox, hence patchbomb.

    +

    As usual, the basic configuration of the patchbomb extension takes just one or + two lines in your + /.hgrc.

    +
    [extensions]
    +patchbomb =
    +

    Once you've enabled the extension, you will have a new + command available, named email.

    +

    The safest and best way to invoke the email command is to + always run it first with the hg -n option. + This will show you what the command would + send, without actually sending anything. Once you've had a + quick glance over the changes and verified that you are sending + the right ones, you can rerun the same command, with the hg -n option + removed.

    +

    The email command + accepts the same kind of revision syntax as every other + Mercurial command. For example, this command will send every + revision between 7 and tip, inclusive.

    +
    hg email -n 7:tip
    +

    You can also specify a repository to + compare with. If you provide a repository but no revisions, the + email command will + send all revisions in the local repository that are not present + in the remote repository. If you additionally specify revisions + or a branch name (the latter using the hg -b option), + this will constrain the revisions sent.

    +

    It's perfectly safe to run the email command without the + names of the people you want to send to: if you do this, it will + just prompt you for those values interactively. (If you're + using a Linux or Unix-like system, you should have enhanced + readline-style editing capabilities when + entering those headers, too, which is useful.)

    +

    When you are sending just one revision, the email command will by + default use the first line of the changeset description as the + subject of the single email message it sends.

    +

    If you send multiple revisions, the email command will usually + send one message per changeset. It will preface the series with + an introductory message, in which you should describe the + purpose of the series of changes you're sending.

    +
    +

    +14.4.1. Changing the behavior of patchbombs

    +

    Not every project has exactly the same conventions for + sending changes in email; the patchbomb extension tries to + accommodate a number of variations through command line + options.

    +
      +
    • You can write a subject for the introductory + message on the command line using the hg -s + option. This takes one argument, the text of the subject + to use.

    • +
    • To change the email address from which the + messages originate, use the hg -f + option. This takes one argument, the email address to + use.

    • +
    • The default behavior is to send unified diffs + (see Section 12.4, “Understanding patches” for a + description of the + format), one per message. You can send a binary bundle + instead with the hg -b + option.

    • +
    • Unified diffs are normally prefaced with a + metadata header. You can omit this, and send unadorned + diffs, with the hg + --plain option.

    • +
    • Diffs are normally sent inline, + in the same body part as the description of a patch. This + makes it easiest for the largest number of readers to + quote and respond to parts of a diff, as some mail clients + will only quote the first MIME body part in a message. If + you'd prefer to send the description and the diff in + separate body parts, use the hg -a + option.

    • +
    • Instead of sending mail messages, you can + write them to an mbox-format mail + folder using the hg -m + option. That option takes one argument, the name of the + file to write to.

    • +
    • If you would like to add a + diffstat-format summary to each patch, + and one to the introductory message, use the hg -d + option. The diffstat command displays + a table containing the name of each file patched, the + number of lines affected, and a histogram showing how much + each file is modified. This gives readers a qualitative + glance at how complex a patch is.

    • +
    +
    +
    +
    + + + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/backup/abc.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/backup/abc.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,2 @@ +pid_list=['x_4fe', 'x_4ff', 'x_546', 'x_503', 'x_505', 'x_506', 'x_507', 'x_508', 'x_509', 'x_50a', 'x_50b', 'x_50c', 'x_510', 'x_511', 'x_513', 'x_515', 'x_516', 'x_518', 'x_51a', 'x_51b', 'x_51c', 'x_51d', 'x_51e', 'x_51f', 'x_520', 'x_521', 'x_522', 'x_523', 'x_524', 'x_525', 'x_526', 'x_527', 'x_528', 'x_529', 'x_52a', 'x_52b', 'x_52c', 'x_52d', 'x_52e', 'x_52f', 'x_530', 'x_531', 'x_532', 'x_533', 'x_534', 'x_535', 'x_536', 'x_537', 'x_538', 'x_539', 'x_53a', 'x_53b'] + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/backup/abcd.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/backup/abcd.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,929 @@ +Chapter 9. Finding and fixing mistakes

    Chapter 9. Finding and fixing mistakes

    To err might be human, but to really handle the consequences + well takes a top-notch revision control system. In this chapter, + we'll discuss some of the techniques you can use when you find + that a problem has crept into your project. Mercurial has some + highly capable features that will help you to isolate the sources + of problems, and to handle them appropriately.

    Erasing local history

    The accidental commit

    I have the occasional but persistent problem of typing + rather more quickly than I can think, which sometimes results + in me committing a changeset that is either incomplete or + plain wrong. In my case, the usual kind of incomplete + changeset is one in which I've created a new source file, but + forgotten to hg add it. A + “plain wrong” changeset is not as common, but no + less annoying.

    Rolling back a transaction

    In the section called “Safe operation”, I + mentioned that Mercurial treats each modification of a + repository as a transaction. Every time + you commit a changeset or pull changes from another + repository, Mercurial remembers what you did. You can undo, + or roll back, exactly one of these + actions using the hg rollback + + command. (See the section called “Rolling back is useless once you've pushed” + for an important caveat about the use of this command.)

    Here's a mistake that I often find myself making: + committing a change in which I've created a new file, but + forgotten to hg add + it.

    $ hg status
    +M a
    +$ echo b > b
    +
    +$ hg commit -m 'Add file b'
    +

    Looking at the output of hg + status after the commit immediately confirms the + error.

    $ hg status
    +? b
    +$ hg tip
    +
    +changeset:   1:f2db1de2ba4f
    +tag:         tip
    +user:        Bryan O'Sullivan <bos@serpentine.com>
    +date:        Tue May 05 06:55:44 2009 +0000
    +summary:     Add file b
    +
    +

    The commit captured the changes to the file + a, but not the new file + b. If I were to push this changeset to a + repository that I shared with a colleague, the chances are + high that something in a would refer to + b, which would not be present in their + repository when they pulled my changes. I would thus become + the object of some indignation.

    However, luck is with me—I've caught my error + before I pushed the changeset. I use the hg rollback command, and Mercurial + makes that last changeset vanish.

    $ hg rollback
    +
    +rolling back last transaction
    +$ hg tip
    +changeset:   0:cde70bc943e1
    +tag:         tip
    +user:        Bryan O'Sullivan <bos@serpentine.com>
    +date:        Tue May 05 06:55:44 2009 +0000
    +summary:     First commit
    +
    +$ hg status
    +M a
    +? b
    +

    Notice that the changeset is no longer present in the + repository's history, and the working directory once again + thinks that the file a is modified. The + commit and rollback have left the working directory exactly as + it was prior to the commit; the changeset has been completely + erased. I can now safely hg + add the file b, and rerun my + commit.

    $ hg add b
    +
    +$ hg commit -m 'Add file b, this time for real'
    +

    The erroneous pull

    It's common practice with Mercurial to maintain separate + development branches of a project in different repositories. + Your development team might have one shared repository for + your project's “0.9” release, and another, + containing different changes, for the “1.0” + release.

    Given this, you can imagine that the consequences could be + messy if you had a local “0.9” repository, and + accidentally pulled changes from the shared “1.0” + repository into it. At worst, you could be paying + insufficient attention, and push those changes into the shared + “0.9” tree, confusing your entire team (but don't + worry, we'll return to this horror scenario later). However, + it's more likely that you'll notice immediately, because + Mercurial will display the URL it's pulling from, or you will + see it pull a suspiciously large number of changes into the + repository.

    The hg rollback command + will work nicely to expunge all of the changesets that you + just pulled. Mercurial groups all changes from one hg pull into a single transaction, + so one hg rollback is all you + need to undo this mistake.

    Rolling back is useless once you've pushed

    The value of the hg + rollback command drops to zero once you've pushed + your changes to another repository. Rolling back a change + makes it disappear entirely, but only in + the repository in which you perform the hg rollback. Because a rollback + eliminates history, there's no way for the disappearance of a + change to propagate between repositories.

    If you've pushed a change to another + repository—particularly if it's a shared + repository—it has essentially “escaped into the + wild,” and you'll have to recover from your mistake + in a different way. If you push a changeset somewhere, then + roll it back, then pull from the repository you pushed to, the + changeset you thought you'd gotten rid of will simply reappear + in your repository.

    (If you absolutely know for sure that the change + you want to roll back is the most recent change in the + repository that you pushed to, and you + know that nobody else could have pulled it from that + repository, you can roll back the changeset there, too, but + you really should not expect this to work reliably. Sooner or + later a change really will make it into a repository that you + don't directly control (or have forgotten about), and come + back to bite you.)

    You can only roll back once

    Mercurial stores exactly one transaction in its + transaction log; that transaction is the most recent one that + occurred in the repository. This means that you can only roll + back one transaction. If you expect to be able to roll back + one transaction, then its predecessor, this is not the + behavior you will get.

    $ hg rollback
    +
    +rolling back last transaction
    +$ hg rollback
    +no rollback information available
    +

    Once you've rolled back one transaction in a repository, + you can't roll back again in that repository until you perform + another commit or pull.

    Reverting the mistaken change

    If you make a modification to a file, and decide that you + really didn't want to change the file at all, and you haven't + yet committed your changes, the hg + revert command is the one you'll need. It looks at + the changeset that's the parent of the working directory, and + restores the contents of the file to their state as of that + changeset. (That's a long-winded way of saying that, in the + normal case, it undoes your modifications.)

    Let's illustrate how the hg + revert command works with yet another small example. + We'll begin by modifying a file that Mercurial is already + tracking.

    $ cat file
    +
    +original content
    +$ echo unwanted change >> file
    +$ hg diff file
    +diff -r b52afd4afc59 file
    +--- a/file	Tue May 05 06:55:32 2009 +0000
    ++++ b/file	Tue May 05 06:55:32 2009 +0000
    +@@ -1,1 +1,2 @@
    + original content
    ++unwanted change
    +

    If we don't + want that change, we can simply hg + revert the file.

    $ hg status
    +
    +M file
    +$ hg revert file
    +$ cat file
    +original content
    +

    The hg revert command + provides us with an extra degree of safety by saving our + modified file with a .orig + extension.

    $ hg status
    +
    +? file.orig
    +$ cat file.orig
    +original content
    +unwanted change
    +
    [Tip]Be careful with .orig files

    It's extremely unlikely that you are either using + Mercurial to manage files with .orig + extensions or that you even care about the contents of such + files. Just in case, though, it's useful to remember that + hg revert will + unconditionally overwrite an existing file with a + .orig extension. For instance, if you + already have a file named foo.orig when + you revert foo, the contents of + foo.orig will be clobbered.

    Here is a summary of the cases that the hg revert command can deal with. We + will describe each of these in more detail in the section that + follows.

    • If you modify a file, it will restore the file + to its unmodified state.

    • If you hg add a + file, it will undo the “added” state of the + file, but leave the file itself untouched.

    • If you delete a file without telling Mercurial, + it will restore the file to its unmodified contents.

    • If you use the hg + remove command to remove a file, it will undo + the “removed” state of the file, and restore + the file to its unmodified contents.

    File management errors

    The hg revert command is + useful for more than just modified files. It lets you reverse + the results of all of Mercurial's file management + commands—hg add, + hg remove, and so on.

    If you hg add a file, + then decide that in fact you don't want Mercurial to track it, + use hg revert to undo the + add. Don't worry; Mercurial will not modify the file in any + way. It will just “unmark” the file.

    $ echo oops > oops
    +
    +$ hg add oops
    +$ hg status oops
    +A oops
    +$ hg revert oops
    +$ hg status
    +
    +? oops
    +

    Similarly, if you ask Mercurial to hg remove a file, you can use + hg revert to restore it to + the contents it had as of the parent of the working directory. + +

    $ hg remove file
    +$ hg status
    +R file
    +
    +$ hg revert file
    +$ hg status
    +$ ls file
    +file
    +

    + + This works just as + well for a file that you deleted by hand, without telling + Mercurial (recall that in Mercurial terminology, this kind of + file is called “missing”).

    $ rm file
    +
    +$ hg status
    +! file
    +$ hg revert file
    +$ ls file
    +file
    +

    If you revert a hg copy, + the copied-to file remains in your working directory + afterwards, untracked. Since a copy doesn't affect the + copied-from file in any way, Mercurial doesn't do anything + with the copied-from file.

    $ hg copy file new-file
    +
    +$ hg revert new-file
    +$ hg status
    +? new-file
    +

    Dealing with committed changes

    Consider a case where you have committed a change + a, and another change + b on top of it; you then realise that + change a was incorrect. Mercurial lets you + “back out” an entire changeset automatically, and + building blocks that let you reverse part of a changeset by + hand.

    Before you read this section, here's something to + keep in mind: the hg backout + + command undoes the effect of a change by + adding to your repository's history, not by + modifying or erasing it. It's the right tool to use if you're + fixing bugs, but not if you're trying to undo some change that + has catastrophic consequences. To deal with those, see + the section called “Changes that should never have been”.

    Backing out a changeset

    The hg backout command + lets you “undo” the effects of an entire + changeset in an automated fashion. Because Mercurial's + history is immutable, this command does + not get rid of the changeset you want to undo. + Instead, it creates a new changeset that + reverses the effect of the to-be-undone + changeset.

    The operation of the hg + backout command is a little intricate, so let's + illustrate it with some examples. First, we'll create a + repository with some simple changes.

    $ hg init myrepo
    +
    +$ cd myrepo
    +$ echo first change >> myfile
    +$ hg add myfile
    +$ hg commit -m 'first change'
    +
    +$ echo second change >> myfile
    +$ hg commit -m 'second change'
    +

    The hg backout command + takes a single changeset ID as its argument; this is the + changeset to back out. Normally, hg + backout will drop you into a text editor to write + a commit message, so you can record why you're backing the + change out. In this example, we provide a commit message on + the command line using the -m option.

    Backing out the tip changeset

    We're going to start by backing out the last changeset we + committed.

    $ hg backout -m 'back out second change' tip
    +
    +reverting myfile
    +changeset 2:01adc4672142 backs out changeset 1:7e341ee3be7a
    +$ cat myfile
    +first change
    +

    You can see that the second line from + myfile is no longer present. Taking a + look at the output of hg log + gives us an idea of what the hg + backout command has done. + + +

    $ hg log --style compact
    +2[tip]   01adc4672142   2009-05-05 06:55 +0000   bos
    +  back out second change
    +
    +1   7e341ee3be7a   2009-05-05 06:55 +0000   bos
    +  second change
    +
    +0   56b97fc928f2   2009-05-05 06:55 +0000   bos
    +  first change
    +
    +

    + + Notice that the new changeset + that hg backout has created + is a child of the changeset we backed out. It's easier to see + this in Figure 9.1, “Backing out a change using the hg backout command”, which presents a + graphical view of the change history. As you can see, the + history is nice and linear.

    Figure 9.1. Backing out a change using the hg backout command

    XXX add text

    Backing out a non-tip change

    If you want to back out a change other than the last one + you committed, pass the --merge option to the + hg backout command.

    $ cd ..
    +
    +$ hg clone -r1 myrepo non-tip-repo
    +requesting all changes
    +adding changesets
    +adding manifests
    +adding file changes
    +added 2 changesets with 2 changes to 1 files
    +updating working directory
    +1 files updated, 0 files merged, 0 files removed, 0 files unresolved
    +$ cd non-tip-repo
    +

    This makes backing out any changeset a + “one-shot” operation that's usually simple and + fast.

    $ echo third change >> myfile
    +
    +$ hg commit -m 'third change'
    +$ hg backout --merge -m 'back out second change' 1
    +reverting myfile
    +created new head
    +changeset 3:abc7fd860049 backs out changeset 1:7e341ee3be7a
    +merging with changeset 3:abc7fd860049
    +merging myfile
    +0 files updated, 1 files merged, 0 files removed, 0 files unresolved
    +(branch merge, don't forget to commit)
    +

    If you take a look at the contents of + myfile after the backout finishes, you'll + see that the first and third changes are present, but not the + second.

    $ cat myfile
    +
    +first change
    +third change
    +

    As the graphical history in Figure 9.2, “Automated backout of a non-tip change using the + hg backout command” illustrates, Mercurial + still commits one change in this kind of situation (the + box-shaped node is the ones that Mercurial commits + automatically), but the revision graph now looks different. + Before Mercurial begins the backout process, it first + remembers what the current parent of the working directory is. + It then backs out the target changeset, and commits that as a + changeset. Finally, it merges back to the previous parent of + the working directory, but notice that it does not + commit the result of the merge. The repository + now contains two heads, and the working directory is in a + merge state.

    Figure 9.2. Automated backout of a non-tip change using the + hg backout command

    XXX add text

    The result is that you end up “back where you + were”, only with some extra history that undoes the + effect of the changeset you wanted to back out.

    You might wonder why Mercurial does not commit the result + of the merge that it performed. The reason lies in Mercurial + behaving conservatively: a merge naturally has more scope for + error than simply undoing the effect of the tip changeset, + so your work will be safest if you first inspect (and test!) + the result of the merge, then commit + it.

    Always use the --merge option

    In fact, since the --merge option will do the + “right thing” whether or not the changeset + you're backing out is the tip (i.e. it won't try to merge if + it's backing out the tip, since there's no need), you should + always use this option when you run the + hg backout command.

    Gaining more control of the backout process

    While I've recommended that you always use the --merge option when backing + out a change, the hg backout + + command lets you decide how to merge a backout changeset. + Taking control of the backout process by hand is something you + will rarely need to do, but it can be useful to understand + what the hg backout command + is doing for you automatically. To illustrate this, let's + clone our first repository, but omit the backout change that + it contains.

    $ cd ..
    +$ hg clone -r1 myrepo newrepo
    +requesting all changes
    +adding changesets
    +adding manifests
    +adding file changes
    +added 2 changesets with 2 changes to 1 files
    +updating working directory
    +1 files updated, 0 files merged, 0 files removed, 0 files unresolved
    +$ cd newrepo
    +
    +

    As with our + earlier example, We'll commit a third changeset, then back out + its parent, and see what happens.

    $ echo third change >> myfile
    +$ hg commit -m 'third change'
    +$ hg backout -m 'back out second change' 1
    +reverting myfile
    +created new head
    +changeset 3:abc7fd860049 backs out changeset 1:7e341ee3be7a
    +the backout changeset is a new head - do not forget to merge
    +(use "backout --merge" if you want to auto-merge)
    +

    Our new changeset is again a descendant of the changeset + we backout out; it's thus a new head, not + + a descendant of the changeset that was the tip. The hg backout command was quite + explicit in telling us this.

    $ hg log --style compact
    +3[tip]:1   abc7fd860049   2009-05-05 06:55 +0000   bos
    +  back out second change
    +
    +2   bae4005ddac4   2009-05-05 06:55 +0000   bos
    +  third change
    +
    +1   7e341ee3be7a   2009-05-05 06:55 +0000   bos
    +  second change
    +
    +0   56b97fc928f2   2009-05-05 06:55 +0000   bos
    +  first change
    +
    +

    Again, it's easier to see what has happened by looking at + a graph of the revision history, in Figure 9.3, “Backing out a change using the hg backout command”. This makes it clear + that when we use hg backout + to back out a change other than the tip, Mercurial adds a new + head to the repository (the change it committed is + box-shaped).

    Figure 9.3. Backing out a change using the hg backout command

    XXX add text

    After the hg backout + + command has completed, it leaves the new + “backout” changeset as the parent of the working + directory.

    $ hg parents
    +changeset:   2:bae4005ddac4
    +user:        Bryan O'Sullivan <bos@serpentine.com>
    +date:        Tue May 05 06:55:12 2009 +0000
    +summary:     third change
    +
    +

    Now we have two isolated sets of changes.

    $ hg heads
    +
    +changeset:   3:abc7fd860049
    +tag:         tip
    +parent:      1:7e341ee3be7a
    +user:        Bryan O'Sullivan <bos@serpentine.com>
    +date:        Tue May 05 06:55:12 2009 +0000
    +summary:     back out second change
    +
    +changeset:   2:bae4005ddac4
    +user:        Bryan O'Sullivan <bos@serpentine.com>
    +date:        Tue May 05 06:55:12 2009 +0000
    +summary:     third change
    +
    +

    Let's think about what we expect to see as the contents of + myfile now. The first change should be + present, because we've never backed it out. The second change + should be missing, as that's the change we backed out. Since + the history graph shows the third change as a separate head, + we don't expect to see the third change + present in myfile.

    $ cat myfile
    +
    +first change
    +

    To get the third change back into the file, we just do a + normal merge of our two heads.

    $ hg merge
    +abort: outstanding uncommitted changes
    +$ hg commit -m 'merged backout with previous tip'
    +$ cat myfile
    +first change
    +

    Afterwards, the graphical history of our + repository looks like + Figure 9.4, “Manually merging a backout change”.

    Figure 9.4. Manually merging a backout change

    XXX add text

    Why hg backout works as + it does

    Here's a brief description of how the hg backout command works.

    1. It ensures that the working directory is + “clean”, i.e. that the output of hg status would be empty.

    2. It remembers the current parent of the working + directory. Let's call this changeset + orig.

    3. It does the equivalent of a hg update to sync the working + directory to the changeset you want to back out. Let's + call this changeset backout.

    4. It finds the parent of that changeset. Let's + call that changeset parent.

    5. For each file that the + backout changeset affected, it does the + equivalent of a hg revert -r + parent on that file, to restore it to the + contents it had before that changeset was + committed.

    6. It commits the result as a new changeset. + This changeset has backout as its + parent.

    7. If you specify --merge on the command + line, it merges with orig, and commits + the result of the merge.

    An alternative way to implement the hg backout command would be to + hg export the + to-be-backed-out changeset as a diff, then use the --reverse option to the + patch command to reverse the effect of the + change without fiddling with the working directory. This + sounds much simpler, but it would not work nearly as + well.

    The reason that hg + backout does an update, a commit, a merge, and + another commit is to give the merge machinery the best chance + to do a good job when dealing with all the changes + between the change you're backing out and + the current tip.

    If you're backing out a changeset that's 100 revisions + back in your project's history, the chances that the + patch command will be able to apply a + reverse diff cleanly are not good, because intervening changes + are likely to have “broken the context” that + patch uses to determine whether it can + apply a patch (if this sounds like gibberish, see the section called “Understanding patches” for a + discussion of the patch command). Also, + Mercurial's merge machinery will handle files and directories + being renamed, permission changes, and modifications to binary + files, none of which patch can deal + with.

    Changes that should never have been

    Most of the time, the hg + backout command is exactly what you need if you want + to undo the effects of a change. It leaves a permanent record + of exactly what you did, both when committing the original + changeset and when you cleaned up after it.

    On rare occasions, though, you may find that you've + committed a change that really should not be present in the + repository at all. For example, it would be very unusual, and + usually considered a mistake, to commit a software project's + object files as well as its source files. Object files have + almost no intrinsic value, and they're big, + so they increase the size of the repository and the amount of + time it takes to clone or pull changes.

    Before I discuss the options that you have if you commit a + “brown paper bag” change (the kind that's so bad + that you want to pull a brown paper bag over your head), let me + first discuss some approaches that probably won't work.

    Since Mercurial treats history as + accumulative—every change builds on top of all changes + that preceded it—you generally can't just make disastrous + changes disappear. The one exception is when you've just + committed a change, and it hasn't been pushed or pulled into + another repository. That's when you can safely use the hg rollback command, as I detailed in + the section called “Rolling back a transaction”.

    After you've pushed a bad change to another repository, you + could still use hg + rollback to make your local copy of the change + disappear, but it won't have the consequences you want. The + change will still be present in the remote repository, so it + will reappear in your local repository the next time you + pull.

    If a situation like this arises, and you know which + repositories your bad change has propagated into, you can + try to get rid of the change from + every one of those repositories. This is, + of course, not a satisfactory solution: if you miss even a + single repository while you're expunging, the change is still + “in the wild”, and could propagate further.

    If you've committed one or more changes + after the change that you'd like to see + disappear, your options are further reduced. Mercurial doesn't + provide a way to “punch a hole” in history, leaving + changesets intact.

    Backing out a merge

    Since merges are often complicated, it is not unheard of + for a merge to be mangled badly, but committed erroneously. + Mercurial provides an important safeguard against bad merges + by refusing to commit unresolved files, but human ingenuity + guarantees that it is still possible to mess a merge up and + commit it.

    Given a bad merge that has been committed, usually the + best way to approach it is to simply try to repair the damage + by hand. A complete disaster that cannot be easily fixed up + by hand ought to be very rare, but the hg backout command may help in + making the cleanup easier. It offers a --parent option, which lets + you specify which parent to revert to when backing out a + merge.

    Figure 9.5. A bad merge

    XXX add text

    Suppose we have a revision graph like that in Figure 9.5, “A bad merge”. What we'd like is to + redo the merge of revisions 2 and + 3.

    One way to do so would be as follows.

    1. Call hg backout --rev=4 + --parent=2. This tells hg backout to back out revision + 4, which is the bad merge, and to when deciding which + revision to prefer, to choose parent 2, one of the parents + of the merge. The effect can be seen in Figure 9.6, “Backing out the merge, favoring one parent”.

      Figure 9.6. Backing out the merge, favoring one parent

      XXX add text

    2. Call hg backout --rev=4 + --parent=3. This tells hg backout to back out revision + 4 again, but this time to choose parent 3, the other + parent of the merge. The result is visible in Figure 9.7, “Backing out the merge, favoring the other + parent”, in which the repository + now contains three heads.

      Figure 9.7. Backing out the merge, favoring the other + parent

      XXX add text

    3. Redo the bad merge by merging the two backout heads, + which reduces the number of heads in the repository to + two, as can be seen in Figure 9.8, “Merging the backouts”.

      Figure 9.8. Merging the backouts

      XXX add text

    4. Merge with the commit that was made after the bad + merge, as shown in Figure 9.9, “Merging the backouts”.

      Figure 9.9. Merging the backouts

      XXX add text

    Protect yourself from “escaped” + changes

    If you've committed some changes to your local repository + and they've been pushed or pulled somewhere else, this isn't + necessarily a disaster. You can protect yourself ahead of + time against some classes of bad changeset. This is + particularly easy if your team usually pulls changes from a + central repository.

    By configuring some hooks on that repository to validate + incoming changesets (see chapter Chapter 10, Handling repository events with hooks), + you can + automatically prevent some kinds of bad changeset from being + pushed to the central repository at all. With such a + configuration in place, some kinds of bad changeset will + naturally tend to “die out” because they can't + propagate into the central repository. Better yet, this + happens without any need for explicit intervention.

    For instance, an incoming change hook that + verifies that a changeset will actually compile can prevent + people from inadvertently “breaking the + build”.

    What to do about sensitive changes that escape

    Even a carefully run project can suffer an unfortunate + event such as the committing and uncontrolled propagation of a + file that contains important passwords.

    If something like this happens to you, and the information + that gets accidentally propagated is truly sensitive, your + first step should be to mitigate the effect of the leak + without trying to control the leak itself. If you are not 100% + certain that you know exactly who could have seen the changes, + you should immediately change passwords, cancel credit cards, + or find some other way to make sure that the information that + has leaked is no longer useful. In other words, assume that + the change has propagated far and wide, and that there's + nothing more you can do.

    You might hope that there would be mechanisms you could + use to either figure out who has seen a change or to erase the + change permanently everywhere, but there are good reasons why + these are not possible.

    Mercurial does not provide an audit trail of who has + pulled changes from a repository, because it is usually either + impossible to record such information or trivial to spoof it. + In a multi-user or networked environment, you should thus be + extremely skeptical of yourself if you think that you have + identified every place that a sensitive changeset has + propagated to. Don't forget that people can and will send + bundles by email, have their backup software save data + offsite, carry repositories on USB sticks, and find other + completely innocent ways to confound your attempts to track + down every copy of a problematic change.

    Mercurial also does not provide a way to make a file or + changeset completely disappear from history, because there is + no way to enforce its disappearance; someone could easily + modify their copy of Mercurial to ignore such directives. In + addition, even if Mercurial provided such a capability, + someone who simply hadn't pulled a “make this file + disappear” changeset wouldn't be affected by it, nor + would web crawlers visiting at the wrong time, disk backups, + or other mechanisms. Indeed, no distributed revision control + system can make data reliably vanish. Providing the illusion + of such control could easily give a false sense of security, + and be worse than not providing it at all.

    Finding the source of a bug

    While it's all very well to be able to back out a changeset + that introduced a bug, this requires that you know which + changeset to back out. Mercurial provides an invaluable + command, called hg bisect, that + helps you to automate this process and accomplish it very + efficiently.

    The idea behind the hg + bisect command is that a changeset has introduced + some change of behavior that you can identify with a simple + pass/fail test. You don't know which piece of code introduced the + change, but you know how to test for the presence of the bug. + The hg bisect command uses your + test to direct its search for the changeset that introduced the + code that caused the bug.

    Here are a few scenarios to help you understand how you + might apply this command.

    • The most recent version of your software has a + bug that you remember wasn't present a few weeks ago, but + you don't know when it was introduced. Here, your binary + test checks for the presence of that bug.

    • You fixed a bug in a rush, and now it's time to + close the entry in your team's bug database. The bug + database requires a changeset ID when you close an entry, + but you don't remember which changeset you fixed the bug in. + Once again, your binary test checks for the presence of the + bug.

    • Your software works correctly, but runs 15% + slower than the last time you measured it. You want to know + which changeset introduced the performance regression. In + this case, your binary test measures the performance of your + software, to see whether it's “fast” or + “slow”.

    • The sizes of the components of your project that + you ship exploded recently, and you suspect that something + changed in the way you build your project.

    From these examples, it should be clear that the hg bisect command is not useful only + for finding the sources of bugs. You can use it to find any + “emergent property” of a repository (anything that + you can't find from a simple text search of the files in the + tree) for which you can write a binary test.

    We'll introduce a little bit of terminology here, just to + make it clear which parts of the search process are your + responsibility, and which are Mercurial's. A + test is something that + you run when hg + bisect chooses a changeset. A + probe is what hg + bisect runs to tell whether a revision is good. + Finally, we'll use the word “bisect”, as both a + noun and a verb, to stand in for the phrase “search using + the hg bisect + + command”.

    One simple way to automate the searching process would be + simply to probe every changeset. However, this scales poorly. + If it took ten minutes to test a single changeset, and you had + 10,000 changesets in your repository, the exhaustive approach + would take on average 35 days to find the + changeset that introduced a bug. Even if you knew that the bug + was introduced by one of the last 500 changesets, and limited + your search to those, you'd still be looking at over 40 hours to + find the changeset that introduced your bug.

    What the hg bisect command + does is use its knowledge of the “shape” of your + project's revision history to perform a search in time + proportional to the logarithm of the number + of changesets to check (the kind of search it performs is called + a dichotomic search). With this approach, searching through + 10,000 changesets will take less than three hours, even at ten + minutes per test (the search will require about 14 tests). + Limit your search to the last hundred changesets, and it will + take only about an hour (roughly seven tests).

    The hg bisect command is + aware of the “branchy” nature of a Mercurial + project's revision history, so it has no problems dealing with + branches, merges, or multiple heads in a repository. It can + prune entire branches of history with a single probe, which is + how it operates so efficiently.

    Using the hg bisect + + command

    Here's an example of hg + bisect in action.

    [Note]Note

    In versions 0.9.5 and earlier of Mercurial, hg bisect was not a core command: + it was distributed with Mercurial as an extension. This + section describes the built-in command, not the old + extension.

    Now let's create a repository, so that we can try out the + hg bisect command in + isolation.

    $ hg init mybug
    +
    +$ cd mybug
    +

    We'll simulate a project that has a bug in it in a + simple-minded way: create trivial changes in a loop, and + nominate one specific change that will have the + “bug”. This loop creates 35 changesets, each + adding a single file to the repository. We'll represent our + “bug” with a file that contains the text “i + have a gub”.

    $ buggy_change=22
    +$ for (( i = 0; i < 35; i++ )); do
    +
    +>   if [[ $i = $buggy_change ]]; then
    +>     echo 'i have a gub' > myfile$i
    +>     hg commit -q -A -m 'buggy changeset'
    +>   else
    +
    +>     echo 'nothing to see here, move along' > myfile$i
    +>     hg commit -q -A -m 'normal changeset'
    +>   fi
    +> done
    +
    +

    The next thing that we'd like to do is figure out how to + use the hg bisect command. + We can use Mercurial's normal built-in help mechanism for + this.

    $ hg help bisect
    +hg bisect [-gbsr] [-c CMD] [REV]
    +
    +subdivision search of changesets
    +
    +    This command helps to find changesets which introduce problems.
    +    To use, mark the earliest changeset you know exhibits the problem
    +    as bad, then mark the latest changeset which is free from the
    +    problem as good. Bisect will update your working directory to a
    +    revision for testing (unless the --noupdate option is specified).
    +    Once you have performed tests, mark the working directory as bad
    +    or good and bisect will either update to another candidate changeset
    +    or announce that it has found the bad revision.
    +
    +    As a shortcut, you can also use the revision argument to mark a
    +    revision as good or bad without checking it out first.
    +
    +    If you supply a command it will be used for automatic bisection. Its exit
    +    status will be used as flag to mark revision as bad or good. In case exit
    +    status is 0 the revision is marked as good, 125 - skipped, 127 (command not
    +    found) - bisection will be aborted; any other status bigger than 0 will
    +    mark revision as bad.
    +
    +options:
    +
    + -r --reset     reset bisect state
    + -g --good      mark changeset good
    + -b --bad       mark changeset bad
    + -s --skip      skip testing changeset
    + -c --command   use command to check changeset state
    + -U --noupdate  do not update to target
    +
    +use "hg -v help bisect" to show global options
    +

    The hg bisect command + works in steps. Each step proceeds as follows.

    1. You run your binary test.

      • If the test succeeded, you tell hg bisect by running the + hg bisect --good + + command.

      • If it failed, run the hg bisect --bad + command.

    2. The command uses your information to decide + which changeset to test next.

    3. It updates the working directory to that + changeset, and the process begins again.

    The process ends when hg + bisect identifies a unique changeset that marks + the point where your test transitioned from + “succeeding” to “failing”.

    To start the search, we must run the hg bisect --reset command.

    $ hg bisect --reset
    +
    +

    In our case, the binary test we use is simple: we check to + see if any file in the repository contains the string “i + have a gub”. If it does, this changeset contains the + change that “caused the bug”. By convention, a + changeset that has the property we're searching for is + “bad”, while one that doesn't is + “good”.

    Most of the time, the revision to which the working + directory is synced (usually the tip) already exhibits the + problem introduced by the buggy change, so we'll mark it as + “bad”.

    $ hg bisect --bad
    +

    Our next task is to nominate a changeset that we know + doesn't have the bug; the hg bisect command will + “bracket” its search between the first pair of + good and bad changesets. In our case, we know that revision + 10 didn't have the bug. (I'll have more words about choosing + the first “good” changeset later.)

    $ hg bisect --good 10
    +
    +Testing changeset 22:b8789808fc48 (24 changesets remaining, ~4 tests)
    +0 files updated, 0 files merged, 12 files removed, 0 files unresolved
    +

    Notice that this command printed some output.

    • It told us how many changesets it must + consider before it can identify the one that introduced + the bug, and how many tests that will require.

    • It updated the working directory to the next + changeset to test, and told us which changeset it's + testing.

    We now run our test in the working directory. We use the + grep command to see if our + “bad” file is present in the working directory. + If it is, this revision is bad; if not, this revision is good. + +

    $ if grep -q 'i have a gub' *
    +> then
    +
    +>   result=bad
    +> else
    +>   result=good
    +> fi
    +$ echo this revision is $result
    +
    +this revision is bad
    +$ hg bisect --$result
    +Testing changeset 16:e61fdddff53e (12 changesets remaining, ~3 tests)
    +0 files updated, 0 files merged, 6 files removed, 0 files unresolved
    +

    + +

    This test looks like a perfect candidate for automation, + so let's turn it into a shell function.

    $ mytest() {
    +>   if grep -q 'i have a gub' *
    +
    +>   then
    +>     result=bad
    +>   else
    +>     result=good
    +>   fi
    +
    +>   echo this revision is $result
    +>   hg bisect --$result
    +> }
    +

    We can now run an entire test step with a single command, + mytest.

    $ mytest
    +
    +this revision is good
    +Testing changeset 19:706df39b003b (6 changesets remaining, ~2 tests)
    +3 files updated, 0 files merged, 0 files removed, 0 files unresolved
    +

    A few more invocations of our canned test step command, + and we're done.

    $ mytest
    +this revision is good
    +Testing changeset 20:bf7ea9a054e6 (3 changesets remaining, ~1 tests)
    +1 files updated, 0 files merged, 0 files removed, 0 files unresolved
    +$ mytest
    +this revision is good
    +Testing changeset 21:921391dd45c1 (2 changesets remaining, ~1 tests)
    +1 files updated, 0 files merged, 0 files removed, 0 files unresolved
    +$ mytest
    +this revision is good
    +The first bad revision is:
    +changeset:   22:b8789808fc48
    +user:        Bryan O'Sullivan <bos@serpentine.com>
    +
    +date:        Tue May 05 06:55:14 2009 +0000
    +summary:     buggy changeset
    +
    +

    Even though we had 40 changesets to search through, the + hg bisect command let us find + the changeset that introduced our “bug” with only + five tests. Because the number of tests that the hg bisect command performs grows + logarithmically with the number of changesets to search, the + advantage that it has over the “brute force” + search approach increases with every changeset you add.

    Cleaning up after your search

    When you're finished using the hg + bisect command in a repository, you can use the + hg bisect --reset command to + drop the information it was using to drive your search. The + command doesn't use much space, so it doesn't matter if you + forget to run this command. However, hg bisect won't let you start a new + search in that repository until you do a hg bisect --reset.

    $ hg bisect --reset
    +
    +

    Tips for finding bugs effectively

    Give consistent input

    The hg bisect command + requires that you correctly report the result of every test + you perform. If you tell it that a test failed when it really + succeeded, it might be able to detect the + inconsistency. If it can identify an inconsistency in your + reports, it will tell you that a particular changeset is both + good and bad. However, it can't do this perfectly; it's about + as likely to report the wrong changeset as the source of the + bug.

    Automate as much as possible

    When I started using the hg + bisect command, I tried a few times to run my + tests by hand, on the command line. This is an approach that + I, at least, am not suited to. After a few tries, I found + that I was making enough mistakes that I was having to restart + my searches several times before finally getting correct + results.

    My initial problems with driving the hg bisect command by hand occurred + even with simple searches on small repositories; if the + problem you're looking for is more subtle, or the number of + tests that hg bisect must + perform increases, the likelihood of operator error ruining + the search is much higher. Once I started automating my + tests, I had much better results.

    The key to automated testing is twofold:

    • always test for the same symptom, and

    • always feed consistent input to the hg bisect command.

    In my tutorial example above, the grep + + command tests for the symptom, and the if + statement takes the result of this check and ensures that we + always feed the same input to the hg + bisect command. The mytest + function marries these together in a reproducible way, so that + every test is uniform and consistent.

    Check your results

    Because the output of a hg + bisect search is only as good as the input you + give it, don't take the changeset it reports as the absolute + truth. A simple way to cross-check its report is to manually + run your test at each of the following changesets:

    • The changeset that it reports as the first bad + revision. Your test should still report this as + bad.

    • The parent of that changeset (either parent, + if it's a merge). Your test should report this changeset + as good.

    • A child of that changeset. Your test should + report this changeset as bad.

    Beware interference between bugs

    It's possible that your search for one bug could be + disrupted by the presence of another. For example, let's say + your software crashes at revision 100, and worked correctly at + revision 50. Unknown to you, someone else introduced a + different crashing bug at revision 60, and fixed it at + revision 80. This could distort your results in one of + several ways.

    It is possible that this other bug completely + “masks” yours, which is to say that it occurs + before your bug has a chance to manifest itself. If you can't + avoid that other bug (for example, it prevents your project + from building), and so can't tell whether your bug is present + in a particular changeset, the hg + bisect command cannot help you directly. Instead, + you can mark a changeset as untested by running hg bisect --skip.

    A different problem could arise if your test for a bug's + presence is not specific enough. If you check for “my + program crashes”, then both your crashing bug and an + unrelated crashing bug that masks it will look like the same + thing, and mislead hg + bisect.

    Another useful situation in which to use hg bisect --skip is if you can't + test a revision because your project was in a broken and hence + untestable state at that revision, perhaps because someone + checked in a change that prevented the project from + building.

    Bracket your search lazily

    Choosing the first “good” and + “bad” changesets that will mark the end points of + your search is often easy, but it bears a little discussion + nevertheless. From the perspective of hg bisect, the “newest” + changeset is conventionally “bad”, and the older + changeset is “good”.

    If you're having trouble remembering when a suitable + “good” change was, so that you can tell hg bisect, you could do worse than + testing changesets at random. Just remember to eliminate + contenders that can't possibly exhibit the bug (perhaps + because the feature with the bug isn't present yet) and those + where another problem masks the bug (as I discussed + above).

    Even if you end up “early” by thousands of + changesets or months of history, you will only add a handful + of tests to the total number that hg + bisect must perform, thanks to its logarithmic + behavior.

    Want to stay up to date? Subscribe to the comment feed for this chapter, or the entire book.

    Copyright 2006, 2007, 2008, 2009 Bryan O'Sullivan. + Icons by Paul Davey aka Mattahan.

    + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/backup/ar01.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/backup/ar01.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,387 @@ + + + +Functional Approach + + + + + + + + + + + + +
    +
    +

    +Functional Approach

    +
    +
    + +

    + Functions allow us to enclose a set of statements and call the function again and again instead of repeating the group of statements everytime. Functions also allow us to isolate a piece of code from all the other code and provides the convenience of not polluting the global variables. +

    +

    + Function in python is defined with the keyword + def followed by the name of the function, in turn followed by a pair of parenthesis which encloses the list of parameters to the function. The definition line ends with a ':'. The definition line is followed by the body of the function intended by one block. The + Function must return a value: +

    +
    def factorial(n):
    +  fact = 1
    +  for i in range(2, n):
    +    fact *= i
    +
    +  return fact
    +
    +
    +

    The code snippet above defines a function with the name factorial, takes the number for which the factorial must be computed, computes the factorial and returns the value.

    +

    A + Function once defined can be used or called anywhere else in the program. We call a fucntion with its name followed by a pair of parenthesis which encloses the arguments to the function. +

    +

    The value that function returns can be assigned to a variable. Let's call the above function and store the factorial in a variable:

    +
    fact5 = factorial(5)
    +
    +
    +

    The value of fact5 will now be 120, which is the factorial of 5. Note that we passed 5 as the argument to the function.

    +

    It may be necessary to document what the function does, for each of the function to help the person who reads our code to understand it better. In order to do this Python allows the first line of the function body to be a string. This string is called as + Documentation String or + docstring. + docstrings prove to be very handy since there are number of tools which can pull out all the docstrings from Python functions and generate the documentation automatically from it. + docstrings for functions can be written as follows: +

    +
    def factorial(n):
    +  'Returns the factorial for the number n.'
    +  fact = 1
    +  for i in range(2, n):
    +    fact *= i
    +
    +  return fact
    +
    +
    +

    An important point to note at this point is that, a function can return any Python value or a Python object, which also includes a + Tuple. A + Tuple is just a collection of values and those values themselves can be of any other valid Python datatypes, including + Lists, + Tuples, + Dictionaries among other things. So effectively, if a function can return a tuple, it can return any number of values through a tuple +

    +

    Let us write a small function to swap two values:

    +
    def swap(a, b):
    +  return b, a
    +
    +c, d = swap(a, b)
    +
    +
    +

    Function scope --------------- The variables used inside the function are confined to the function's scope and doesn't pollute the variables of the same name outside the scope of the function. Also the arguments passed to the function are passed by-value if it is of basic Python data type:

    +
    def cant_change(n):
    +  n = 10
    +
    +n = 5
    +cant_change(n)
    +
    +
    +

    Upon running this code, what do you think would have happened to value of n which was assigned 5 before the function call? If you have already tried out that snippet on the interpreter you already know that the value of n is not changed. This is true of any immutable types of Python like + Numbers, + Strings and + Tuples. But when you pass mutable objects like + Lists and + Dictionaries the values are manipulated even outside the function: +

    +
    >>> def can_change(n):
    +...   n[1] = James
    +...
    +
    +>>> name = ['Mr.', 'Steve', 'Gosling']
    +>>> can_change(name)
    +>>> name
    +['Mr.', 'James', 'Gosling']
    +
    +
    +

    If nothing is returned by the function explicitly, Python takes care to return None when the funnction is called.

    +
    +

    +1. Default Arguments

    +

    There may be situations where we need to allow the functions to take the arguments optionally. Python allows us to define function this way by providing a facility called + Default Arguments. For example, we need to write a function that returns a list of fibonacci numbers. Since our function cannot generate an infinite list of fibonacci numbers, we need to specify the number of elements that the fibonacci sequence must contain. Suppose, additionally, we want to the function to return 10 numbers in the sequence if no option is specified we can define the function as follows: +

    +
    def fib(n=10):
    +  fib_list = [0, 1]
    +  for i in range(n - 2):
    +    next = fib_list[-2] + fib_list[-1]
    +    fib_list.append(next)
    +  return fib_list
    +
    +
    +

    When we call this function, we can optionally specify the value for the parameter n, during the call as an argument. Calling with no argument and argument with n=5 returns the following fibonacci sequences:

    +
    fib()
    +[0, 1, 1, 2, 3, 5, 8, 13, 21, 34]
    +fib(5)
    +[0, 1, 1, 2, 3]
    +
    +
    +
    +
    +

    +2. Keyword Arguments

    +

    When a function takes a large number of arguments, it may be difficult to remember the order of the parameters in the function definition or it may be necessary to pass values to only certain parameters since others take the default value. In either of these cases, Python provides the facility of passing arguments by specifying the name of the parameter as defined in the function definition. This is known as + Keyword Arguments. +

    +

    In a function call, + Keyword arguments can be used for each argument, in the following fashion: +

    +
    argument_name=argument_value
    +Also denoted as: keyword=argument
    +
    +def wish(name='World', greetings='Hello'):
    +  print "%s, %s!" % (greetings, name)
    +
    +
    +

    This function can be called in one of the following ways. It is important to note that no restriction is imposed in the order in which + Keyword arguments can be specified. Also note, that we have combined + Keyword arguments with + Default arguments in this example, however it is not necessary: +

    +
    wish(name='Guido', greetings='Hey')
    +wish(greetings='Hey', name='Guido')
    +
    +
    +

    Calling functions by specifying arguments in the order of parameters specified in the function definition is called as + Positional arguments, as opposed to + Keyword arguments. It is possible to use both + Positional arguments and + Keyword arguments in a single function call. But Python doesn't allow us to bungle up both of them. The arguments to the function, in the call, must always start with + Positional arguments which is in turn followed by + Keyword arguments: +

    +
    def my_func(x, y, z, u, v, w):
    +  # initialize variables.
    +  ...
    +  # do some stuff 
    +  ...
    +  # return the value
    +
    +
    +

    It is valid to call the above functions in the following ways:

    +
    my_func(10, 20, 30, u=1.0, v=2.0, w=3.0)
    +my_func(10, 20, 30, 1.0, 2.0, w=3.0)
    +my_func(10, 20, z=30, u=1.0, v=2.0, w=3.0)
    +my_func(x=10, y=20, z=30, u=1.0, v=2.0, w=3.0)
    +
    +
    +

    Following lists some of the invalid calls:

    +
    my_func(10, 20, z=30, 1.0, 2.0, 3.0)
    +my_func(x=10, 20, z=30, 1.0, 2.0, 3.0)
    +my_func(x=10, y=20, z=30, u=1.0, v=2.0, 3.0)
    +
    +
    +
    +
    +

    +3. Parameter Packing and Unpacking

    +

    The positional arguments passed to a function can be collected in a tuple parameter and keyword arguments can be collected in a dictionary. Since keyword arguments must always be the last set of arguments passed to a function, the keyword dictionary parameter must be the last parameter. The function definition must include a list explicit parameters, followed by tuple paramter collecting parameter, whose name is preceded by a *****, for collecting positional parameters, in turn followed by the dictionary collecting parameter, whose name is preceded by a ****** :

    +
    def print_report(title, *args, **name):
    +  """Structure of *args*
    +  (age, email-id)
    +  Structure of *name*
    +  {
    +      'first': First Name
    +      'middle': Middle Name
    +      'last': Last Name
    +  }
    +  """
    +
    +  print "Title: %s" % (title)
    +  print "Full name: %(first)s %(middle)s %(last)s" % name
    +  print "Age: %d\nEmail-ID: %s" % args
    +
    +
    +

    The above function can be called as. Note, the order of keyword parameters can be interchanged:

    +
    >>> print_report('Employee Report', 29, 'johny@example.com', first='Johny',
    +                 last='Charles', middle='Douglas')
    +Title: Employee Report
    +Full name: Johny Douglas Charles
    +Age: 29
    +Email-ID: johny@example.com
    +
    +
    +

    The reverse of this can also be achieved by using a very identical syntax while calling the function. A tuple or a dictionary can be passed as arguments in place of a list of *Positional arguments* or *Keyword arguments* respectively using ***** or ****** :

    +
    def print_report(title, age, email, first, middle, last):
    +  print "Title: %s" % (title)
    +  print "Full name: %s %s %s" % (first, middle, last)
    +  print "Age: %d\nEmail-ID: %s" % (age, email)
    +
    +>>> args = (29, 'johny@example.com')
    +>>> name = {
    +        'first': 'Johny',
    +        'middle': 'Charles',
    +        'last': 'Douglas'
    +        }
    +>>> print_report('Employee Report', *args, **name)
    +Title: Employee Report
    +Full name: Johny Charles Douglas
    +Age: 29
    +Email-ID: johny@example.com
    +
    +
    +
    +
    +

    +4. Nested Functions and Scopes

    +

    Python allows nesting one function inside another. This style of programming turns out to be extremely flexible and powerful features when we use + Python decorators. We will not talk about decorators is beyond the scope of this course. If you are interested in knowing more about + decorator programming in Python you are suggested to read: +

    +


    +      http://avinashv.net/2008/04/python-decorators-syntactic-sugar/
    +      http://personalpages.tds.net/~kent37/kk/00001.html
    +    

    +

    However, the following is an example for nested functions in Python:

    +
    def outer():
    +  print "Outer..."
    +  def inner():
    +    print "Inner..."
    +  print "Outer..."
    +  inner()
    +
    +>>> outer()
    +
    +
    +
    +
    +

    +5. map, reduce and filter functions

    +

    Python provides several built-in functions for convenience. The + map(), + reduce() and + filter() functions prove to be very useful with sequences like + Lists. +

    +

    The + map ( + function, + sequence) function takes two arguments: + function and a + sequence argument. The + function argument must be the name of the function which in turn takes a single argument, the individual element of the + sequence. The + map function calls + function(item), for each item in the sequence and returns a list of values, where each value is the value returned by each call to + function(item). + map() function allows to pass more than one sequence. In this case, the first argument, + function must take as many arguments as the number of sequences passed. This function is called with each corresponding element in the each of the sequences, or + None if one of the sequence is exhausted: +

    +
    def square(x):
    +  return x*x
    +
    +>>> map(square, [1, 2, 3, 4])
    +[1, 4, 9, 16]
    +
    +def mul(x, y):
    +  return x*y
    +
    +>>> map(mul, [1, 2, 3, 4], [6, 7, 8, 9])
    +
    +
    +

    The + filter ( + function, + sequence) function takes two arguments, similar to the + map() function. The + filter function calls + function(item), for each item in the sequence and returns all the elements in the sequence for which + function(item) returned True: +

    +
    def even(x):
    +  if x % 2:
    +    return True
    +  else:
    +    return False
    +
    +>>> filter(even, range(1, 10))
    +[1, 3, 5, 7, 9]
    +
    +
    +

    The + reduce ( + function, + sequence) function takes two arguments, similar to + map function, however multiple sequences are not allowed. The + reduce function calls + function with first two consecutive elements in the sequence, obtains the result, calls + function with the result and the subsequent element in the sequence and so on until the end of the list and returns the final result: +

    +
    def mul(x, y):
    +  return x*y
    +
    +>>> reduce(mul, [1, 2, 3, 4])
    +24
    +
    +
    +
    +

    +5.1. List Comprehensions

    +

    List Comprehension is a convenvience utility provided by Python. It is a syntatic sugar to create + Lists. Using + List Comprehensions one can create + Lists from other type of sequential data structures or other + Lists itself. The syntax of + List Comprehensions consists of a square brackets to indicate the result is a + List within which we include at least one + for clause and multiple + if clauses. It will be more clear with an example: +

    +
    >>> num = [1, 2, 3]
    +>>> sq = [x*x for x in num]
    +>>> sq
    +[1, 4, 9]
    +>>> all_num = [1, 2, 3, 4, 5, 6, 7, 8, 9]
    +>>> even = [x for x in all_num if x%2 == 0]
    +
    +
    +

    The syntax used here is very clear from the way it is written. It can be translated into english as, "for each element x in the list all_num, if remainder of x divided by 2 is 0, add x to the list."

    +
    +
    +
    + + + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/backup/autoid.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/backup/autoid.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,47 @@ +#!/usr/bin/env python +# +# Add unique ID attributes to para tags. This script should only be +# run by one person, since otherwise it introduces the possibility of +# chaotic conflicts among tags. + +import glob, os, re, sys + +tagged = re.compile(']* id="x_([0-9a-f]+)"[^>]*>', re.M) +untagged = re.compile('') + +names = glob.glob('ch*.xml') + glob.glob('app*.xml') + +# First pass: find the highest-numbered paragraph ID. + +biggest_id = 0 +seen = set() +errs = 0 + +for name in names: + for m in tagged.finditer(open(name).read()): + i = int(m.group(1),16) + if i in seen: + print >> sys.stderr, '%s: duplication of ID %s' % (name, i) + errs += 1 + seen.add(i) + if i > biggest_id: + biggest_id = i + +def retag(s): + global biggest_id + biggest_id += 1 + return '' % biggest_id + +# Second pass: add IDs to paragraphs that currently lack them. + +for name in names: + f = open(name).read() + f1 = untagged.sub(retag, f) + if f1 != f: + tmpname = name + '.tmp' + fp = open(tmpname, 'w') + fp.write(f1) + fp.close() + os.rename(tmpname, name) + +sys.exit(errs) diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/backup/ch01-intro.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/backup/ch01-intro.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,686 @@ + + + +Basic Python + + + + + + + + + + +
    +
    +
    +

    +Basic Python

    +
    +
    + +

    This document is intended to be handed out at the end of the workshop. It has been designed for Engineering students who are Python beginners and have basic programming skills. The focus is on basic numerics and plotting using Python.

    +

    The system requirements:

    +
      +
    • Python - version 2.5.x or newer.

    • +
    • IPython

    • +
    • Text editor - scite, vim, emacs or whatever you are comfortable with.

    • +
    +
    +

    +1. Introduction

    +

    The Python programming language was created by a dutch named Guido van Rossum. The idea of Python was conceived in December 1989. The name Python has nothing to do with the reptilian, but its been named after the 70s comedy series "Monty Python's Flying Circus", since it happens to be Guido's favourite TV series.

    +

    Current stable version of Python is 2.6.x, although Python 3.0 is also the stable version, it is not backwards compatible with the previous versions and is hence not entirely popular at the moment. This material will focus on the 2.6.x series.

    +

    Python is licensed under the Python Software Foundation License (PSF License) which is GPL compatible Free Software license (excepting license version 1.6 and 2.0) It is a no strings attached license, which means the source code is free to modify and redistribute.

    +

    The Python docs define Python as "Python is an interpreted, object-oriented, high-level programming language with dynamic semantics." A more detailed summary can be found at + http://www.python.org/doc/essays/blurb.html. Python is a language that has been designed to help the programmer concentrate on solving the problem at hand and not worry about the programming language idiosyncrasies. +

    +

    Python is a highly cross platform compatible language on account of it being an interpreted language. It is highly scalable and hence has been adapted to run on the Nokia 60 series phones. Python has been designed to be readable and easy to use

    +

    + Resources available for reference +

    +
    +

    + Advantages of Python - Why Python?? +

    +
      +
    • Python has been designed for readability and ease of use. Its been designed in such a fashion that it imposes readability on the programmer. Python does away with the braces and the semicolons and instead implements code blocks based on indentation, thus enhancing readability.

    • +
    • Python is a high level, interpreted, modular and object oriented language. Python performs memory management on its own, thus the programmer need not bother about allocating and deallocating memory to variables. Python provides extensibility by providing modules which can be easily imported similar to headers in C and packages in Java. Python is object oriented and hence provides all the object oriented characteristics such as inheritance, encapsulation and polymorphism.

    • +
    • Python offers a highly powerful interactive programming interface in the form of the 'Interactive Interpreter' which will be discussed in more detail in the following sections.

    • +
    • Python provides a rich standard library and an extensive set of modules. The power of Python modules can be seen in this slightly exaggerated cartoon + http://xkcd.com/353/ +

    • +
    • Python interfaces well with most other programming languages such as C, C++ and FORTRAN.

    • +
    +

    Although, Python has one setback. Python is not fast as some of the compiled languages like C or C++. Yet, the amount of flexibility and power more than make up for this setback.

    +
    +
    +

    +2. The Python Interpreter

    +
    +

    +2.1. The Interactive Interpreter

    +

    Typing + python at the shell prompt on any standard Unix/Gnu-Linux system and hitting the enter key fires up the Python 'Interactive Interpreter'. The Python interpreter is one of the most integral features of Python. The prompt obtained when the interactive interpreter is similar to what is shown below. The exact appearance might differ based on the version of Python being used. The + >>> thing shown is the python prompt. When something is typed at the prompt and the enter key is hit, the python interpreter interprets the command entered and performs the appropriate action. All the examples presented in this document are to be tried hands on, on the interactive interpreter. +

    +
    Python 2.5.2 (r252:60911, Oct  5 2008, 19:24:49) 
    +[GCC 4.3.2] on linux2
    +Type "help", "copyright", "credits" or "license" for more information.
    +>>> 
    +
    +
    +

    Lets try with an example, type + print 'Hello, World!' at the prompt and hit the enter key. +

    +
    >>> print 'Hello, World!'
    +Hello, World!
    +
    +
    +

    This example was quite straight forward, and thus we have written our first line of Python code. Now let us try typing something arbitrary at the prompt. For example:

    +
    >>> arbit word
    +  File "<stdin>", line 1
    +    arbit word
    +            ^
    +SyntaxError: invalid syntax
    +>>>
    +
    +
    +

    The interpreter gave an error message saying that 'arbit word' was invalid syntax which is valid. The interpreter is an amazing tool when learning to program in Python. The interpreter provides a help function that provides the necessary documentation regarding all Python syntax, constructs, modules and objects. Typing + help() at the prompt gives the following output: +

    +
    >>> help()
    +
    +Welcome to Python 2.5!  This is the online help utility.
    +
    +If this is your first time using Python, you should definitely check out
    +the tutorial on the Internet at http://www.python.org/doc/tut/.
    +
    +Enter the name of any module, keyword, or topic to get help on writing
    +Python programs and using Python modules.  To quit this help utility and
    +return to the interpreter, just type "quit".
    +
    +To get a list of available modules, keywords, or topics, type "modules",
    +"keywords", or "topics".  Each module also comes with a one-line summary
    +of what it does; to list the modules whose summaries contain a given word
    +such as "spam", type "modules spam".
    +
    +help> 
    +
    +
    +
    +

    As mentioned in the output, entering the name of any module, keyword or topic will provide the documentation and help regarding the same through the online help utility. Pressing + Ctrl+d exits the help prompt and returns to the python prompt. +

    +

    Let us now try a few examples at the python interpreter.

    +

    Eg 1:

    +
    >>> print 'Hello, python!'
    +Hello, python!
    +>>>
    +
    +
    +

    Eg 2:

    +
    >>> print 4321*567890
    +2453852690
    +>>> 
    +
    +
    +

    Eg 3:

    +
    >>> 4321*567890
    +2453852690L
    +>>>
    +
    +
    +
    Note: Notice the 'L' at the end of the output. The 'L' signifies that the
    +output of the operation is of type *long*. It was absent in the previous
    +example because we used the print statement. This is because *print* formats
    +the output before displaying.
    +
    +
    +

    Eg 4:

    +
    >>> big = 12345678901234567890 ** 3
    +>>> print big
    +1881676372353657772490265749424677022198701224860897069000
    +>>> 
    +
    +
    +
    This example is to show that unlike in C or C++ there is no limit on the
    +value of an integer.
    +
    +
    +

    Try this on the interactive interpreter: + import this +

    +

    + Hint: The output gives an idea of Power of Python +

    +
    +
    +

    +2.2.  + ipython - An enhanced interactive Python interpreter +

    +

    The power and the importance of the interactive interpreter was the highlight of the previous section. This section provides insight into the enhanced interpreter with more advanced set of features called + ipython. Entering + ipython at the shell prompt fires up the interactive interpreter. +

    +
    $ ipython
    +Python 2.5.2 (r252:60911, Oct  5 2008, 19:24:49) 
    +Type "copyright", "credits" or "license" for more information.
    +
    +IPython 0.8.4 -- An enhanced Interactive Python.
    +?         -> Introduction and overview of IPython's features.
    +%quickref -> Quick reference.
    +help      -> Python's own help system.
    +object?   -> Details about 'object'. ?object also works, ?? prints more.
    +
    +In [1]: 
    +
    +
    +

    This is the output obtained upon firing ipython. The exact appearance may change based on the Python version installed. The following are some of the various features provided by + ipython: +

    +
    +
    +
    +

    +

    Suggestions - ipython provides suggestions of the possible methods and operations available for the given python object.

    +
    +
    +

    Eg 5:

    +
    In [4]: a = 6
    +
    +In [5]: a.
    +a.__abs__           a.__divmod__        a.__index__         a.__neg__          a.__rand__          a.__rmod__          a.__rxor__
    +a.__add__           a.__doc__           a.__init__          a.__new__          a.__rdiv__          a.__rmul__          a.__setattr__
    +a.__and__           a.__float__         a.__int__           a.__nonzero__      a.__rdivmod__       a.__ror__           a.__str__
    +a.__class__         a.__floordiv__      a.__invert__        a.__oct__          a.__reduce__        a.__rpow__          a.__sub__
    +a.__cmp__           a.__getattribute__  a.__long__          a.__or__           a.__reduce_ex__     a.__rrshift__       a.__truediv__
    +a.__coerce__        a.__getnewargs__    a.__lshift__        a.__pos__          a.__repr__          a.__rshift__        a.__xor__
    +a.__delattr__       a.__hash__          a.__mod__           a.__pow__          a.__rfloordiv__     a.__rsub__          
    +a.__div__           a.__hex__           a.__mul__           a.__radd__         a.__rlshift__       a.__rtruediv__      
    +
    +
    +

    In this example, we initialized 'a' (a variable - a concept that will be discussed in the subsequent sections.) to 6. In the next line when the + tab key is pressed after typing ' + a.' ipython displays the set of all possible methods that are applicable on the object 'a' (an integer in this context). Ipython provides many such datatype specific features which will be presented in the further sections as and when the datatypes are introduced. +

    +
    +
    +
    +

    +3. Editing and running a python file

    +

    The previous sections focused on the use of the interpreter to run python code. While the interpeter is an excellent tool to test simple solutions and experiment with small code snippets, its main disadvantage is that everything written in the interpreter is lost once its quit. Most of the times a program is used by people other than the author. So the programs have to be available in some form suitable for distribution, and hence they are written in files. This section will focus on editing and running python files. Start by opening a text editor ( it is recommended you choose one from the list at the top of this page ). In the editor type down python code and save the file with an extension + .py (python files have an extension of .py). Once done with the editing, save the file and exit the editor. +

    +

    Let us look at a simple example of calculating the gcd of 2 numbers using Python:

    +

    + Creating the first python script(file) : +

    +
    $ emacs gcd.py
    +  def gcd(x,y):
    +    if x % y == 0:
    +      return y
    +    return gcd(y, x%y)
    +
    +  print gcd(72, 92)
    +
    +
    +

    To run the script, open the shell prompt, navigate to the directory that contains the python file and run + python <filename.py> at the prompt ( in this case filename is gcd.py ) +

    +

    + Running the python script : +

    +
    $ python gcd.py
    +4
    +$ 
    +
    +
    +

    Another method to run a python script would be to include the line

    +

    + #! /usr/bin/python +

    +

    at the beginning of the python file and then make the file executable by

    +

    $ chmod a+x + filename.py +

    +

    Once this is done, the script can be run as a standalone program as follows:

    +

    $ ./ + filename.py +

    +
    +
    +

    +4. Basic Datatypes and operators in Python

    +

    Python provides the following set of basic datatypes.

    +
      +
    • Numbers: int, float, long, complex

    • +
    • Strings

    • +
    • Boolean

    • +
    +
    +

    +4.1. Numbers

    +

    Numbers were introduced in the examples presented in the interactive interpreter section. Numbers include types as mentioned earlier viz., int (integers), float (floating point numbers), long (large integers), complex (complex numbers with real and imaginary parts). Python is not a strongly typed language, which means the type of a variable need not mentioned during its initialization. Let us look at a few examples.

    +

    Eg 6:

    +
    >>> a = 1 #here a is an integer variable
    +
    +
    +

    Eg 7:

    +
    >>> lng = 122333444455555666666777777788888888999999999 #here lng is a variable of type long
    +>>> lng
    +122333444455555666666777777788888888999999999L #notice the trailing 'L'
    +>>> print lng
    +122333444455555666666777777788888888999999999 #notice the absence of the trailing 'L'
    +>>> lng+1
    +122333444455555666666777777788888889000000000L
    +
    +
    +
    +

    Long numbers are the same as integers in almost all aspects. They can be used in operations just like integers and along with integers without any distinction. The only distinction comes during type checking (which is not a healthy practice). Long numbers are tucked with a trailing 'L' just to signify that they are long. Notice that in the example just lng at the prompt displays the value of the variable with the 'L' whereas + print lng displays without the 'L'. This is because print formats the output before printing. Also in the example, notice that adding an integer to a long does not give any errors and the result is as expected. So for all practical purposes longs can be treated as ints. +

    +

    Eg 8:

    +
    >>> fl = 3.14159 #fl is a float variable
    +>>> e = 1.234e-4 #e is also a float variable, specified in the exponential form
    +>>> a = 1
    +>>> b = 2
    +>>> a/b #integer division
    +0
    +>>> a/fl #floating point division
    +0.31831015504887655
    +>>> e/fl
    +3.9279473133031364e-05
    +
    +
    +
    +

    Floating point numbers, simply called floats are real numbers with a decimal point. The example above shows the initialization of a float variable. Shown also in this example is the difference between integer division and floating point division. 'a' and 'b' here are integer variables and hence the division gives 0 as the quotient. When either of the operands is a float, the operation is a floating point division, and the result is also a float as illustrated.

    +

    Eg 9:

    +
    >>> cplx = 3 + 4j #cplx is a complex variable
    +>>> cplx
    +(3+4j)
    +>>> print cplx.real #prints the real part of the complex number
    +3.0
    +>>> print cplx.imag #prints the imaginary part of the complex number
    +4.0
    +>>> print cplx*fl  #multiplies the real and imag parts of the complex number with the multiplier
    +(9.42477+12.56636j)
    +>>> abs(cplx) #returns the absolute value of the complex number
    +5.0
    +
    +
    +

    Python provides a datatype for complex numbers. Complex numbers are initialized as shown in the example above. The + real and + imag operators return the real and imaginary parts of the complex number as shown. The + abs() returns the absolute value of the complex number. +

    +
    +
    +

    +4.2. Variables

    +

    Variables are just names that represent a value. Variables have already been introduced in the various examples from the previous sections. Certain rules about using variables:

    +
      +
    • Variables have to be initialized or assigned a value before being used.

    • +
    • Variable names can consist of letters, digits and + underscores . +

    • +
    • Variable names cannot begin with digits, but can contain digits in them.

    • +
    +

    In reference to the previous section examples, 'a', 'b', 'lng', 'fl', 'e' and 'cplx' are all variables of various datatypes.

    +
    Note: Python is not a strongly typed language and hence an integer variable can at a
    +later stage be used as a float variable as well.
    +
    +
    +
    +
    +

    +4.3. Strings

    +

    Strings are one of the essential data structures of any programming language. The + print "Hello, World!" program was introduced in the earlier section, and the + "Hello, World!" in the print statement is a string. A string is basically a set of characters. Strings can be represented in various ways shown below: +

    +
    s = 'this is a string'              # a string variable can be represented using single quotes
    +s = 'This one has "quotes" inside!' # The string can have quotes inside it as shown
    +s = "I have 'single-quotes' inside!"
    +l = "A string spanning many lines\
    +one more line\
    +yet another"                        # a string can span more than a single line.
    +t = """A triple quoted string does  # another way of representing multiline strings.
    +not need to be escaped at the end and
    +"can have nested quotes" etc."""
    +
    +
    +

    Try the following on the interpreter: + s = 'this is a string with 'quotes' of similar kind' +

    +

    + Exercise: How to use single quotes within single quotes in a string as shown in the above example without getting an error? +

    +
    +

    +4.3.1. String operations

    +

    A few basic string operations are presented here.

    +

    + String concatenation String concatenation is done by simple addition of two strings. +

    +
    >>> x = 'Hello'
    +>>> y = ' Python'
    +>>> print x+y
    +Hello Python
    +
    +
    +

    Try this yourself: +

    +
    >>> somenum = 13
    +>>> print x+somenum
    +
    +
    +

    The problem with the above example is that here a string variable and an integer variable are trying to be concantenated. To obtain the desired result from the above example the str(), repr() and the `` can be used.

    +

    + str() simply converts a value to a string in a reasonable form. + repr() creates a string that is a representation of the value. +

    +

    The difference can be seen in the example shown below:

    +
    >>> str(1000000000000000000000000000000000000000000000000L)
    +'1000000000000000000000000000000000000000000000000'
    +>>> repr(1000000000000000000000000000000000000000000000000L)
    +'1000000000000000000000000000000000000000000000000L'
    +
    +
    +

    It can be observed that the 'L' in the long value shown was omitted by str(), whereas repr() converted that into a string too. An alternative way of using repr(value) is ` + value`. +

    +

    A few more examples:

    +
    >>> x = "Let's go \nto Pycon"
    +>>> print x
    +Let's go 
    +to Pycon
    +
    +
    +

    In the above example, notice that the '\n'(newline) character is formatted and the string is printed on two lines. The strings discussed until now were normal strings. Other than these there are two other types of strings namely, raw strings and unicode strings.

    +

    + Raw strings are strings which are unformatted, that is the backslashes(\) are not parsed and are left as it is in the string. Raw strings are represented with an 'r' at the start of a string. Let us look at an example +

    +
    >>> x = r"Let's go \nto Pycon"
    +>>> print x
    +Let's go \nto Pycon
    +
    +
    +

    Note: The '\n' is not being parsed into a new line and is left as it is.

    +

    + Try this yourself: +

    +
    >>> x = r"Let's go to Pycon\"
    +
    +
    +

    Unicode strings are strings where the characters are Unicode characters as opposed to ASCII characters. Unicode strings are represented with a 'u' at the start of the string. Let us look at an example: +

    +
    >>> x = u"Let's go to Pycon!"
    +>>> print x
    +Let's go to Pycon!
    +
    +
    +
    +
    +
    +

    +4.4. Boolean

    +

    Python also provides special Boolean datatype. A boolean variable can assume a value of either + True or + False (Note the capitalizations). +

    +

    Let us look at examples:

    +
    >>> t = True
    +>>> f = not t
    +>>> print f
    +False
    +>>> f or t
    +True
    +>>> f and t
    +False
    +
    +
    +
    +
    +
    +

    +5. The + while loop +

    +

    The Python + while loop is similar to the C/C++ while loop. The syntax is as follows: +

    +
    statement 0
    +while condition:
    +  statement 1 #while block
    +  statement 2 #while block
    +statement 3 #outside the while block.
    +
    +
    +

    Let us look at an example:

    +
    >>> x = 1  
    +>>> while x <= 5:
    +...   print x
    +...   x += 1
    +... 
    +1
    +2
    +3
    +4
    +5
    +
    +
    +
    +
    +

    +6. The + if conditional +

    +

    The Python + if block provides the conditional execution of statements. If the condition evaluates as true the block of statements defined under the if block are executed. +

    +

    If the first block is not executed on account of the condition not being satisfied, the set of statements in the + else block are executed. +

    +

    The + elif block provides the functionality of evaluation of multiple conditions as shown in the example. +

    +

    The syntax is as follows:

    +
    if condition :
    +    statement_1
    +    statement_2
    +
    +elif condition:
    +    statement_3
    +    statement_4
    +else:
    +    statement_5
    +    statement_6
    +
    +
    +

    Let us look at an example:

    +
    >>> n = raw_input("Input a number:")
    +>>> if n < 0:
    +      print n," is negative"
    +      elif n > 0:
    +      print n," is positive"
    +      else:
    +      print n, " is 0"
    +
    +
    +
    +
    +

    +7.  raw_input() +

    +

    In the previous example we saw the call to the raw_input() subroutine. The + raw_input() method is used to take user inputs through the console. Unlike + input() which assumes the data entered by the user as a standard python expression, + raw_input() treats all the input data as raw data and converts everything into a string. To illustrate this let us look at an example. +

    +
    >>> input("Enter a number thats a palindrome:")
    +Enter a number thats a palindrome:121
    +121
    +
    +>>> input("Enter your name:")
    +Enter your name:PythonFreak
    +Traceback (most recent call last):
    +  File "<stdin>", line 1, in <module>
    +  File "<string>", line 1, in <module>
    +NameError: name 'PythonFreak' is not defined
    +
    +
    +

    As shown above the + input() assumes that the data entered is a valid Python expression. In the first call it prompts for an integer input and when entered it accepts the integer as an integer, whereas in the second call, when the string is entered without the quotes, + input() assumes that the entered data is a valid Python expression and hence it raises and exception saying PythonFreak is not defined. +

    +
    >>> input("Enter your name:")
    +Enter your name:'PythonFreak'
    +'PythonFreak'
    +>>> 
    +
    +
    +

    Here the name is accepted because its entered as a string (within quotes). But its unreasonable to go on using quotes each time a string is entered. Hence the alternative is to use + raw_input(). +

    +

    Let us now look at how + raw_input() operates with an example. +

    +
    >>> raw_input("Enter your name:")
    +Enter your name:PythonFreak
    +'PythonFreak'
    +
    +
    +

    Observe that the + raw_input() is converting it into a string all by itself. +

    +
    >>> pal = raw_input("Enter a number thats a palindrome:")
    +Enter a number thats a palindrome:121
    +'121'
    +
    +
    +

    Observe that + raw_input() is converting the integer 121 also to a string as '121'. Let us look at another example: +

    +
    >>> pal = raw_input("Enter a number thats a palindrome:")
    +Enter a number thats a palindrome:121
    +>>> pal + 2
    +Traceback (most recent call last):
    +  File "<stdin>", line 1, in <module>
    +TypeError: cannot concatenate 'str' and 'int' objects
    +>>> pal
    +'121'
    +
    +
    +

    Observe here that the variable + pal is a string and hence integer operations cannot be performed on it. Hence the exception is raised. +

    +
    +
    +

    +8.  + int() method +

    +

    Generally for computing purposes, the data used is not strings or raw data but on integers, floats and similar mathematical data structures. The data obtained from + raw_input() is raw data in the form of strings. In order to obtain integers from strings we use the method + int(). +

    +

    Let us look at an example.

    +
    >>> intpal = int(pal)
    +>>> intpal
    +121
    +
    +
    +

    In the previous example it was observed that + pal was a string variable. Here using the + int() method the string + pal was converted to an integer variable. +

    +

    + Try This Yourself: +

    +
    >>> stringvar = raw_input("Enter a name:")
    +Enter a name:Guido Van Rossum
    +>>> stringvar
    +'Guido Van Rossum'
    +>>> numvar = int(stringvar)
    +
    +
    +
    +
    + + + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/backup/ch02-list_tuples.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/backup/ch02-list_tuples.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,652 @@ + + + +Lists and Tuples + + + + + + + + +
    +
    +

    +Lists and Tuples

    +
    +
    +Lists +

    Python provides an intuitive way to represent a group items, called + Lists. The items of a + List are called its elements. Unlike C/C++, elements can be of any type. A + List is represented as a list of comma-sepated elements with square brackets around them: +

    +
    >>> a = [10, 'Python programming', 20.3523, 23, 3534534L]
    +>>> a
    +[10, 'Python programming', 20.3523, 23, 3534534L]
    +
    +
    +
    +
    +

    +1. Common List Operations

    +

    The following are some of the most commonly used operations on + Lists. +

    +
    +

    +1.1. Indexing

    +

    Individual elements of a + List can be accessed using an index to the element. The indices start at 0. One can also access the elements of the + List in reverse using negative indices.: +

    +
    >>> a[1]
    +'Python programming'
    +>>> a[-1]
    +3534534L
    +
    +
    +

    It is important to note here that the last element of the + List has an index of -1. +

    +
    +
    +

    +1.2. Concatenating

    +

    Two or more + Lists can be concatenated using the + operator: +

    +
    >>> a + ['foo', 12, 23.3432, 54]
    +[10, 'Python programming', 20.3523, 'foo', 12, 23.3432, 54]
    +>>> [54, 75, 23] + ['write', 67, 'read']
    +[54, 75, 23, 'write', 67, 'read']
    +
    +
    +
    +
    +
    +

    +1.3. Slicing

    +

    A + List can be sliced off to contain a subset of elements of the + List. Slicing can be done by using two indices separated by a colon, where the first index is inclusive and the second index is exclusive. The resulting slice is also a + List.: +

    +
    >>> num = [1, 2, 3, 4, 5, 6, 7, 8, 9]
    +>>> num[3:6]
    +[4, 5, 6]
    +>>> num[0:1]
    +[1]
    +>>> num[7:10]
    +[7, 8, 9]
    +
    +
    +

    The last example showed how to access last 3 elements of the + List. There is a small catch here. The second index 10 actually refers to the 11th element of the + List which is still valid, even though it doesn't exist because the second index is exclusive and tells the Python interpreter to get the last element of the + List. But this can also be done in a much easier way using negative indices: +

    +
    >>> num[-3:-1]
    +[7, 8, 9]
    +
    +
    +

    Excluding the first index implies that the slice must start at the beginning of the + List, while excluding the second index includes all the elements till the end of the + List. A third parameter to a slice, which is implicitly taken as 1 is the step of the slice. It is specified as a value which follows a colon after the second index: +

    +
    >>> num[:4]
    +[1, 2, 3, 4]
    +>>> num[7:]
    +[8, 9]
    +>>> num[-3:]
    +[7, 8, 9]
    +>>> num[:]
    +[1, 2, 3, 4, 5, 6, 7, 8, 9]
    +>>> num[4:9:3]
    +[5, 8]
    +>>> num[3::2]
    +[4, 6, 8]
    +>>> num[::4]
    +[1, 5, 9]
    +
    +
    +
    +
    +
    +

    +1.4. Multiplication

    +

    A + List can be multiplied with an integer to repeat itself: +

    +
    >>> [20] * 5
    +[20, 20, 20, 20, 20]
    +>>> [42, 'Python', 54] * 3
    +[42, 'Python', 54, 42, 'Python', 54, 42, 'Python', 54]
    +
    +
    +
    +
    +
    +

    +1.5. Membership

    +

    + in operator is used to find whether an element is part of the + List. It returns + True if the element is present in the + List or + False if it is not present. Since this operator returns a Boolean value it is called a Boolean operator: +

    +
    >>> names = ['Guido', 'Alex', 'Tim']
    +>>> 'Tim' in names
    +True
    +>>> 'Adam' in names
    +False
    +
    +
    +
    +
    +
    +

    +1.6. Length, Maximum and Minimum

    +

    Length of a + List can be found out using the len function. The max function returns the element with the largest value and the min function returns the element with the smallest value: +

    +
    >>> num = [4, 1, 32, 12, 67, 34, 65]
    +>>> len(num)
    +7
    +>>> max(num)
    +67
    +>>> min(num)
    +1
    +
    +
    +
    +
    +
    +

    +1.7. Changing Elements

    +

    Unlike Strings + Lists are mutable, i.e. elements of a + List can be manipulated: +

    +
    >>> a = [1, 3, 5, 7]
    +>>> a[2] = 9
    +>>> a
    +[1, 3, 9, 7]
    +
    +
    +
    +
    +
    +

    +1.8. Deleting Elements

    +

    An element or a slice of a + List can be deleted by using the + del statement: +

    +
    >>> a = [1, 3, 5, 7, 9, 11]
    +>>> del a[-2:]
    +>>> a
    +[1, 3, 5, 7]
    +>>> del a[1]
    +>>> a
    +[1, 5, 7]
    +
    +
    +
    +
    +
    +

    +1.9. Assign to Slices

    +

    In the same way, values can be assigned to individual elements of the + List, a + List of elements can be assigned to a slice: +

    +
    >>> a = [2, 3, 4, 5]
    +>>> a[:2] = [0, 1]
    +[0, 1, 4, 5]
    +>>> a[2:2] = [2, 3]
    +>>> a
    +[0, 1, 2, 3, 4, 5]
    +>>> a[2:4] = []
    +>>> a
    +[0, 1, 4, 5]
    +
    +
    +

    The last two examples should be particularly noted carefully. The last but one example insert elements or a list of elements into a + List and the last example deletes a list of elements from the + List. +

    +
    +
    +
    +

    +2. None, Empty Lists, and Initialization

    +

    An + Empty List is a + List with no elements and is simply represented as []. A + None List is one with all elements in it being + None. It serves the purpose having a container list of some fixed number of elements with no value: +

    +
    >>> a = []
    +>>> a
    +[]
    +>>> n = [None] * 10
    +>>> n
    +[None, None, None, None, None, None, None, None, None, None]
    +
    +
    +
    +
    +
    +

    +3. Nested Lists

    +

    As mentioned earlier, a List can contain elements of any data type. This also implies a + List can have a + Lists themselves as its elements. These are called as + Nested Lists. There is no limit on the depth of the + Nested Lists: +

    +
    >>> a = [1, [1, 2, 3], 3, [1, [1, 2, 3]], 7]
    +
    +
    +
    +
    +
    +

    +4. List Methods

    +

    A method is a function that is coupled to an object. More about objects and its methods are discussed in Advanced Python module. In general, a method is called like:

    +
    object.method(arguments)
    +
    +
    +

    For now, it is enough to know that a list of elements is an object and so + List methods can be called upon them. Also some of the methods change the + List in-place, meaning it modifies the existing list instead of creating a new one, while other methods don't. It must be noted as we run through the + List methods. +

    +

    Some of the most commonly used + List methods are as follows: +

    +
    +

    +4.1. append

    +

    The + append method is used to append an object at the end of the list: +

    +
    >>> prime = [2, 3, 5]
    +>>> prime.append(7)
    +>>> prime
    +[2, 3, 5, 7]
    +
    +
    +

    It is important to note that append changes the + List in-place. +

    +
    +
    +

    +4.2. count

    +

    The + count method returns the number of occurences of a particular element in a list: +

    +
    >>> [1, 4, 4, 9, 9, 9].count(9)
    +3
    +>>> tlst = ['Python', 'is', 'a', 'beautiful', 'language']
    +>>> tlst.count('Python')
    +1
    +
    +
    +
    +
    +
    +

    +4.3. extend

    +

    The + extend method extends the list on which it is called by the list supplied as argument to it: +

    +
    >>> a = [1, 2, 3]
    +>>> b = [4, 5, 6]
    +>>> a.extend(b)
    +[1, 2, 3, 4, 5, 6]
    +
    +
    +

    This is an in-place method. This method is equivalent to using the + operator, but using the + operator returns a new list.

    +
    +
    +

    +4.4. index

    +

    The + index method returns the index position of the element in the list specified as argument: +

    +
    >>> a = [1, 2, 3, ,4, 5]
    +>>> a.index(4)
    +3
    +
    +
    +
    +
    +
    +

    +4.5. insert

    +

    The + insert method is used to insert an element specified as the second argument to the list at the position specified by the first argument: +

    +
    >>> a = ['Python', 'is', 'cool']
    +>>> a.insert(2, 'so')
    +>>> a
    +['Python', 'is', 'so', 'cool']
    +
    +
    +

    The + insert method changes the + List in-place. +

    +
    +
    +

    +4.6. pop

    +

    The + pop method removes an element from the list. The index position of the element to be removed can be specified as an argument to the + pop method, if not it removes the last element by default: +

    +
    >>> a = [1, 2, 3, 4, 5]
    +>>> a.pop()
    +>>> a
    +5
    +>>> a.pop(2)
    +>>> a
    +3
    +
    +
    +

    The + pop method changes the + List in-place. +

    +
    +
    +

    +4.7. remove

    +

    The + remove method removes the first occurence of an element supplied as a parameter: +

    +
    >>> a = [1, 2, 3, 4, 2, 5, 2]
    +>>> a.remove(2)
    +>>> a
    +[1, 3, 4, 2, 5, 2]
    +
    +
    +
    +
    +
    +

    +4.8. reverse

    +

    The + reverse method reverses elements in the list. It is important to note here that + reverse method changes the list in-place and doesn't return any thing: +

    +
    >>> a = ['guido', 'alex', 'tim']
    +>>> a.reverse()
    +>>> a
    +['tim', 'alex', 'guido']
    +
    +
    +
    +
    +
    +

    +4.9. sort

    +

    The + sort method is used to sort the elements of the list. The + sort method also sorts in-place and does not return anything: +

    +
    >>> a = [5, 1, 3, 7, 4]
    +>>> a.sort()
    +>>> a
    +[1, 3, 4, 5, 7]
    +
    +
    +

    In addition to the sort method on a + List object we can also use the built-in + sorted function. This function takes the + List as a parameter and returns a sorted copy of the list. However the original list is left intact: +

    +
    >>> a = [5, 1, 3, 7, 4]
    +>>> b = sorted(a)
    +>>> b
    +[1, 3, 4, 5, 7]
    +>>> a
    +[5, 1, 3, 7, 4]
    +
    +
    +
    +
    +
    +
    +

    +5. Tuples

    +

    + Tuples are sequences just like + Lists, but they are immutable. In other words + Tuples provides a way to represent a group of items, where the group of items cannot be changed in any way. The syntax of a + Tuple is also very similar to + List. A + Tuple is represented with the list of items, called elements of the + Tuple separated by comma, with the entire list being enclosed in parenthesis. It is not compulsory to use parenthesis around a + Tuple but it may be necessary in some of the cases: +

    +
    >>> a = 1, 2, 3
    +>>> a
    +(1, 2, 3)
    +>>> b = 1,
    +>>> b
    +(1,)
    +
    +
    +

    It is interesting to note the second example. Just a value followed by a comma automatically makes that an element of a + Tuple with only one element. It is also important to note that, irrespective of input having a parenthesis, the output always has a parenthesis. +

    +

    The first example is also known as + Tuple packing, because values are being packed into a tuple. It is also possible to do + Tuple unpacking which is more interesting. It is better to understand that by example. Say we have a co-ordinate pair from which we need to separate x and y co-ordinates: +

    +
    >>> a = (1, 2)
    +>>> x, y = a
    +>>> x
    +1
    +>>> y
    +2
    +
    +
    +

    Tuple unpacking also has several other use-cases of which the most interesting one is to swap the values of two variables. Using programming languages like C would require anywhere around 10 lines of code and an extra temporary variable to do this (including all the #include stuff). Python does it in the most intuitive way in just one line. Say we want to swap the co-ordinates in the above example: +

    +
    >>> x, y = y, x
    +>>> x
    +2
    +>>> y
    +1
    +
    +
    +
    +

    +5.1. Common Tuple Operations

    +

    There is no need to introduce all the + Tuple operations again, since + Tuples support the following operations that + List supports in exactly the same way: +

    +
      +
    • Indexing

    • +
    • Concatenating

    • +
    • Slicing

    • +
    • Membership

    • +
    • Multiplication

    • +
    • Length, Maximum, Minimum

    • +
    +

    The following examples illustrate the above operations:

    +
    >>> a = (1, 2, 3, 4, 5, 6)
    +>>> a[5]
    +6
    +>>> b = (7, 8, 9)
    +>>> a + b
    +(1, 2, 3, 4, 5, 6, 7, 8, 9)
    +>>> a[3:5]
    +(4, 5)
    +>>> 5 in a
    +True
    +>>> c = (1,)
    +>>> c * 5
    +(1, 1, 1, 1, 1)
    +>>> len(a)
    +6
    +>>> max(a)
    +6
    +>>> min(a)
    +1
    +
    +
    +

    However the following + List operations are not supported by + Tuples because + Tuples cannot be changed once they are created: +

    +
      +
    • Changing elements

    • +
    • Deleting elements

    • +
    • Assigning to slices

    • +
    +

    Similarity to + Lists leads to the questions like, why not + Lists only? Why do we even want + Tuples? Can we do the same with + Lists? And the answer is + Yes we can do it, but + Tuples are helpful at times, like we can return Tuples from functions. They are also returned by some built-in functions and methods. And also there are some use cases like co-ordinate among other things. So + Tuples are helpful. +

    +
    +
    +
    +

    +6. Additional Syntax

    +

    The following additional syntax are introduced to make it easier to operate on + Lists. +

    +
    +

    +6.1. range()

    +

    The + range function takes at least one argument and 2 additional optional arguments. If two or more arguments are specified, the range function returns a list of natural numbers starting from the first argument passed to it to the second argument. The third argument, if specified is used as a step. Suppose only one argument is specified, then + range function returns a list of natural numbers starting from 0 upto the argument specified: +

    +
    >>> range(5, 10, 2)
    +[5, 7, 9]
    +>>> range(2, 15)
    +[2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14]
    +>>> range(12)
    +[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11]
    +
    +
    +
    +
    +

    +6.2. for

    +

    The + for keyword is used as a part of the looping construct. Unlike for loops in other languages, Python's for is used to iterate through the elements of sequences like + Lists, + Tuples, + Dictionaries, etc. The syntax of the for loop consists of + for, followed by a variable to hold the individual or the current element of the list during iteration and + in, followed by the sequence and a semicolon(':') The next line which is part of the + for loop, i.e the statements that are part of the loop should start with a new intend: +

    +
    >>> names = ['Guido', 'Alex', 'Tim']
    +>>> for name in names:
    +...   print "Name =", name
    +... 
    +Name = Guido
    +Name = Alex
    +Name = Tim
    +
    +
    +
    +
    +
    +
    +

    +7. Conclusion

    +

    This section on + Lists and + Tuples introduces almost all the necessary machinary required to work on + Lists and + Tuples. Topics like how to use these data structures in bigger more useful programs will be introduced in the subsequent chapters. +

    +
    +
    + + + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/backup/ch03-oop.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/backup/ch03-oop.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,109 @@ + + + +Classes and Objects + + + + + + + + +
    +
    +

    +Classes and Objects

    +
    +
    +

    In the previous sections we learnt about functions which provide certain level of abstraction to our code by holding the code which performs one or more specific functionalities. We were able to use this function as many times as we wanted. In addition to functions, Python also higher level of abstractions through + Classes and + Objects. + Objects can be loosely defined as a collection of a set of data items and a set of methods. The data items can be any valid Python variable or any Python object. Functions enclosed within a class are called as + methods. If you are thinking if methods are functions why is there a distinction between the two? The answer to this will be given as we walk through the concepts of + Classes and + Objects. + Classes contain the definition for the + Objects. + Objects are instances of + Classes. +

    +

    A class is defined using the keyword + class followed by the class name, in turn followed by a semicolon. The statements that a + Class encloses are written in a new block, i.e on the next indentation level: +

    +
    class Employee:
    +  def setName(self, name):
    +    self.name = name
    +
    +  def getName(self):
    +    return self.name
    +
    +
    +

    In the above example, we defined a class with the name Employee. We also defined two methods, setName and getName for this class. It is important to note the differences between the normal Python functions and class methods defined above. Each method of the class must take the same instance of the class(object) from which it was called as the first argument. It is conventionally given the name, + self. Note that + self is only a convention. You can use any other name, but the first argument to the method will always be the same object of the class from which the method was called. The data memebers that belong to the class are called as + class attributes. + Class attributes are preceded by the object of the class and a dot. In the above example, + name is a class attribute since it is preceded by the + self object. + Class attributes can be accessed from anywhere within the class. +

    +

    We can create objects of a class outside the class definition by using the same syntax we use to call a function with no parameters. We can assign this object to a variable:

    +
    emp = Employee()
    +
    +
    +

    In the above example, we create an object named + emp of the class + Employee. All the attributes and methods of the class can be accessed by the object of the class using the standard notation + object.attribute or + object.method(). Although the first parameter of a class method is the self object, it must not be passed as an argument when calling the method. The + self object is implicitly passed to the method by the Python interpreter. All other arguments passing rules like default arguments, keyword arguments, argument packing and unpacking follow the same rules as those for ordinary Python functions: +

    +
    >>> emp.setName('John')
    +>>> name = emp.getName()
    +>>> print name
    +John
    +>>> print emp.name
    +John
    +
    +
    +

    If we at all try to access a class attribute before assigning a value to it, i.e before creating it, Python raises the same error as it would raise for the accessing undefined variable:

    +
    >>> emp = Employee()
    +>>> emp.name
    +Traceback (most recent call last):
    +  File "class.py", line 10, in <module>
    +    print e.name
    +AttributeError: Employee instance has no attribute 'name'
    +
    +
    +
    + + + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/backup/chap_intro.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/backup/chap_intro.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,2 @@ +p_list={'chap_intro':['x_38', 'x_39', 'x_3d', 'x_3e', 'x_3f', 'x_40', 'x_41', 'x_42', 'x_46', 'x_4c', 'x_4d', 'x_4e', 'x_4f', 'x_50', 'x_51', 'x_52', 'x_53', 'x_54', 'x_55', 'x_56', 'x_57', 'x_58', 'x_59', 'x_5a', 'x_5b', 'x_5c', 'x_5d', 'x_5e', 'x_5f', 'x_60', 'x_61', 'x_62', 'x_63', 'x_64', 'x_65', 'x_66', 'x_67', 'x_68', 'x_69', 'x_6d', 'x_6e', 'x_6f', 'x_70', 'x_71', 'x_72', 'x_73', 'x_74', 'x_75', 'x_79', 'x_7a', 'x_7b', 'x_7c', 'x_7d', 'x_7e', 'x_7f', 'x_80', 'x_81', 'x_82', 'x_83', 'x_84', 'x_85', 'x_86', 'x_87', 'x_88', 'x_89', 'x_8a', 'x_8b', 'x_8c', 'x_8d', 'x_8e', 'x_8f', 'x_90', 'x_91', 'x_92', 'x_93', 'x_94', 'x_95', 'x_96', 'x_97', 'x_98', 'x_99', 'x_9a', 'x_9b', 'x_9c', 'x_9d']} + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/backup/func.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/backup/func.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,726 @@ + + +Chapter 9. Finding and fixing mistakes + + + + + + + + + + + +
    +

    +Chapter 14. Adding functionality with extensions

    + +

    Table of Contents

    +
    +
    1. Default Arguments
    +
    2. Keyword Arguments
    +
    3. Parameter Packing and Unpacking
    +
    4. Nested Functions and Scopes
    +
    5. map, reduce and filter functions
    +
    5.1. List Comprehensions
    +
    + +

    + Functions allow us to enclose a set of statements and call the function again and again instead of repeating the group of statements everytime. Functions also allow us to isolate a piece of code from all the other code and provides the convenience of not polluting the global variables. +

    +

    + Function in python is defined with the keyword + def followed by the name of the function, in turn followed by a pair of parenthesis which encloses the list of parameters to the function. The definition line ends with a ':'. The definition line is followed by the body of the function intended by one block. The + Function must return a value: +

    +
    def factorial(n):
    +  fact = 1
    +  for i in range(2, n):
    +    fact *= i
    +
    +  return fact
    +
    +
    +

    The code snippet above defines a function with the name factorial, takes the number for which the factorial must be computed, computes the factorial and returns the value.

    +

    A + Function once defined can be used or called anywhere else in the program. We call a fucntion with its name followed by a pair of parenthesis which encloses the arguments to the function. +

    +

    The value that function returns can be assigned to a variable. Let's call the above function and store the factorial in a variable:

    +
    fact5 = factorial(5)
    +
    +
    +

    The value of fact5 will now be 120, which is the factorial of 5. Note that we passed 5 as the argument to the function.

    +

    It may be necessary to document what the function does, for each of the function to help the person who reads our code to understand it better. In order to do this Python allows the first line of the function body to be a string. This string is called as + Documentation String or + docstring. + docstrings prove to be very handy since there are number of tools which can pull out all the docstrings from Python functions and generate the documentation automatically from it. + docstrings for functions can be written as follows: +

    +
    def factorial(n):
    +  'Returns the factorial for the number n.'
    +  fact = 1
    +  for i in range(2, n):
    +    fact *= i
    +
    +  return fact
    +
    +
    +

    An important point to note at this point is that, a function can return any Python value or a Python object, which also includes a + Tuple. A + Tuple is just a collection of values and those values themselves can be of any other valid Python datatypes, including + Lists, + Tuples, + Dictionaries among other things. So effectively, if a function can return a tuple, it can return any number of values through a tuple +

    +

    Let us write a small function to swap two values:

    +
    def swap(a, b):
    +  return b, a
    +
    +c, d = swap(a, b)
    +
    +
    +

    Function scope --------------- The variables used inside the function are confined to the function's scope and doesn't pollute the variables of the same name outside the scope of the function. Also the arguments passed to the function are passed by-value if it is of basic Python data type:

    +
    def cant_change(n):
    +  n = 10
    +
    +n = 5
    +cant_change(n)
    +
    +
    +

    Upon running this code, what do you think would have happened to value of n which was assigned 5 before the function call? If you have already tried out that snippet on the interpreter you already know that the value of n is not changed. This is true of any immutable types of Python like + Numbers, + Strings and + Tuples. But when you pass mutable objects like + Lists and + Dictionaries the values are manipulated even outside the function: +

    +
    >>> def can_change(n):
    +...   n[1] = James
    +...
    +
    +>>> name = ['Mr.', 'Steve', 'Gosling']
    +>>> can_change(name)
    +>>> name
    +['Mr.', 'James', 'Gosling']
    +
    +
    +

    If nothing is returned by the function explicitly, Python takes care to return None when the funnction is called.

    +
    +

    +1. Default Arguments

    +

    There may be situations where we need to allow the functions to take the arguments optionally. Python allows us to define function this way by providing a facility called + Default Arguments. For example, we need to write a function that returns a list of fibonacci numbers. Since our function cannot generate an infinite list of fibonacci numbers, we need to specify the number of elements that the fibonacci sequence must contain. Suppose, additionally, we want to the function to return 10 numbers in the sequence if no option is specified we can define the function as follows: +

    +
    def fib(n=10):
    +  fib_list = [0, 1]
    +  for i in range(n - 2):
    +    next = fib_list[-2] + fib_list[-1]
    +    fib_list.append(next)
    +  return fib_list
    +
    +
    +

    When we call this function, we can optionally specify the value for the parameter n, during the call as an argument. Calling with no argument and argument with n=5 returns the following fibonacci sequences:

    +
    fib()
    +[0, 1, 1, 2, 3, 5, 8, 13, 21, 34]
    +fib(5)
    +[0, 1, 1, 2, 3]
    +
    +
    +
    +
    +

    +2. Keyword Arguments

    +

    When a function takes a large number of arguments, it may be difficult to remember the order of the parameters in the function definition or it may be necessary to pass values to only certain parameters since others take the default value. In either of these cases, Python provides the facility of passing arguments by specifying the name of the parameter as defined in the function definition. This is known as + Keyword Arguments. +

    +

    In a function call, + Keyword arguments can be used for each argument, in the following fashion: +

    +
    argument_name=argument_value
    +Also denoted as: keyword=argument
    +
    +def wish(name='World', greetings='Hello'):
    +  print "%s, %s!" % (greetings, name)
    +
    +
    +

    This function can be called in one of the following ways. It is important to note that no restriction is imposed in the order in which + Keyword arguments can be specified. Also note, that we have combined + Keyword arguments with + Default arguments in this example, however it is not necessary: +

    +
    wish(name='Guido', greetings='Hey')
    +wish(greetings='Hey', name='Guido')
    +
    +
    +

    Calling functions by specifying arguments in the order of parameters specified in the function definition is called as + Positional arguments, as opposed to + Keyword arguments. It is possible to use both + Positional arguments and + Keyword arguments in a single function call. But Python doesn't allow us to bungle up both of them. The arguments to the function, in the call, must always start with + Positional arguments which is in turn followed by + Keyword arguments: +

    +
    def my_func(x, y, z, u, v, w):
    +  # initialize variables.
    +  ...
    +  # do some stuff 
    +  ...
    +  # return the value
    +
    +
    +

    It is valid to call the above functions in the following ways:

    +
    my_func(10, 20, 30, u=1.0, v=2.0, w=3.0)
    +my_func(10, 20, 30, 1.0, 2.0, w=3.0)
    +my_func(10, 20, z=30, u=1.0, v=2.0, w=3.0)
    +my_func(x=10, y=20, z=30, u=1.0, v=2.0, w=3.0)
    +
    +
    +

    Following lists some of the invalid calls:

    +
    my_func(10, 20, z=30, 1.0, 2.0, 3.0)
    +my_func(x=10, 20, z=30, 1.0, 2.0, 3.0)
    +my_func(x=10, y=20, z=30, u=1.0, v=2.0, 3.0)
    +
    +
    +
    +
    +

    +3. Parameter Packing and Unpacking

    +

    The positional arguments passed to a function can be collected in a tuple parameter and keyword arguments can be collected in a dictionary. Since keyword arguments must always be the last set of arguments passed to a function, the keyword dictionary parameter must be the last parameter. The function definition must include a list explicit parameters, followed by tuple paramter collecting parameter, whose name is preceded by a *****, for collecting positional parameters, in turn followed by the dictionary collecting parameter, whose name is preceded by a ****** :

    +
    def print_report(title, *args, **name):
    +  """Structure of *args*
    +  (age, email-id)
    +  Structure of *name*
    +  {
    +      'first': First Name
    +      'middle': Middle Name
    +      'last': Last Name
    +  }
    +  """
    +
    +  print "Title: %s" % (title)
    +  print "Full name: %(first)s %(middle)s %(last)s" % name
    +  print "Age: %d\nEmail-ID: %s" % args
    +
    +
    +

    The above function can be called as. Note, the order of keyword parameters can be interchanged:

    +
    >>> print_report('Employee Report', 29, 'johny@example.com', first='Johny',
    +                 last='Charles', middle='Douglas')
    +Title: Employee Report
    +Full name: Johny Douglas Charles
    +Age: 29
    +Email-ID: johny@example.com
    +
    +
    +

    The reverse of this can also be achieved by using a very identical syntax while calling the function. A tuple or a dictionary can be passed as arguments in place of a list of *Positional arguments* or *Keyword arguments* respectively using ***** or ****** :

    +
    def print_report(title, age, email, first, middle, last):
    +  print "Title: %s" % (title)
    +  print "Full name: %s %s %s" % (first, middle, last)
    +  print "Age: %d\nEmail-ID: %s" % (age, email)
    +
    +>>> args = (29, 'johny@example.com')
    +>>> name = {
    +        'first': 'Johny',
    +        'middle': 'Charles',
    +        'last': 'Douglas'
    +        }
    +>>> print_report('Employee Report', *args, **name)
    +Title: Employee Report
    +Full name: Johny Charles Douglas
    +Age: 29
    +Email-ID: johny@example.com
    +
    +
    +
    +
    +

    +4. Nested Functions and Scopes

    +

    Python allows nesting one function inside another. This style of programming turns out to be extremely flexible and powerful features when we use + Python decorators. We will not talk about decorators is beyond the scope of this course. If you are interested in knowing more about + decorator programming in Python you are suggested to read: +

    +


    +      http://avinashv.net/2008/04/python-decorators-syntactic-sugar/
    +      http://personalpages.tds.net/~kent37/kk/00001.html
    +    

    +

    However, the following is an example for nested functions in Python:

    +
    def outer():
    +  print "Outer..."
    +  def inner():
    +    print "Inner..."
    +  print "Outer..."
    +  inner()
    +
    +>>> outer()
    +
    +
    +
    +
    +

    +5. map, reduce and filter functions

    +

    Python provides several built-in functions for convenience. The + map(), + reduce() and + filter() functions prove to be very useful with sequences like + Lists. +

    +

    The + map ( + function, + sequence) function takes two arguments: + function and a + sequence argument. The + function argument must be the name of the function which in turn takes a single argument, the individual element of the + sequence. The + map function calls + function(item), for each item in the sequence and returns a list of values, where each value is the value returned by each call to + function(item). + map() function allows to pass more than one sequence. In this case, the first argument, + function must take as many arguments as the number of sequences passed. This function is called with each corresponding element in the each of the sequences, or + None if one of the sequence is exhausted: +

    +
    def square(x):
    +  return x*x
    +
    +>>> map(square, [1, 2, 3, 4])
    +[1, 4, 9, 16]
    +
    +def mul(x, y):
    +  return x*y
    +
    +>>> map(mul, [1, 2, 3, 4], [6, 7, 8, 9])
    +
    +
    +

    The + filter ( + function, + sequence) function takes two arguments, similar to the + map() function. The + filter function calls + function(item), for each item in the sequence and returns all the elements in the sequence for which + function(item) returned True: +

    +
    def even(x):
    +  if x % 2:
    +    return True
    +  else:
    +    return False
    +
    +>>> filter(even, range(1, 10))
    +[1, 3, 5, 7, 9]
    +
    +
    +

    The + reduce ( + function, + sequence) function takes two arguments, similar to + map function, however multiple sequences are not allowed. The + reduce function calls + function with first two consecutive elements in the sequence, obtains the result, calls + function with the result and the subsequent element in the sequence and so on until the end of the list and returns the final result: +

    +
    def mul(x, y):
    +  return x*y
    +
    +>>> reduce(mul, [1, 2, 3, 4])
    +24
    +
    +
    +
    +

    +5.1. List Comprehensions

    +

    List Comprehension is a convenvience utility provided by Python. It is a syntatic sugar to create + Lists. Using + List Comprehensions one can create + Lists from other type of sequential data structures or other + Lists itself. The syntax of + List Comprehensions consists of a square brackets to indicate the result is a + List within which we include at least one + for clause and multiple + if clauses. It will be more clear with an example: +

    +
    >>> num = [1, 2, 3]
    +>>> sq = [x*x for x in num]
    +>>> sq
    +[1, 4, 9]
    +>>> all_num = [1, 2, 3, 4, 5, 6, 7, 8, 9]
    +>>> even = [x for x in all_num if x%2 == 0]
    +
    +
    +

    The syntax used here is very clear from the way it is written. It can be translated into english as, "for each element x in the list all_num, if remainder of x divided by 2 is 0, add x to the list."

    +
    +
    +
    + + + + +Functional Approach + + + +
    +
    +

    +Functional Approach

    +
    +
    + +

    + Functions allow us to enclose a set of statements and call the function again and again instead of repeating the group of statements everytime. Functions also allow us to isolate a piece of code from all the other code and provides the convenience of not polluting the global variables. +

    +

    + Function in python is defined with the keyword + def followed by the name of the function, in turn followed by a pair of parenthesis which encloses the list of parameters to the function. The definition line ends with a ':'. The definition line is followed by the body of the function intended by one block. The + Function must return a value: +

    +
    def factorial(n):
    +  fact = 1
    +  for i in range(2, n):
    +    fact *= i
    +
    +  return fact
    +
    +
    +

    The code snippet above defines a function with the name factorial, takes the number for which the factorial must be computed, computes the factorial and returns the value.

    +

    A + Function once defined can be used or called anywhere else in the program. We call a fucntion with its name followed by a pair of parenthesis which encloses the arguments to the function. +

    +

    The value that function returns can be assigned to a variable. Let's call the above function and store the factorial in a variable:

    +
    fact5 = factorial(5)
    +
    +
    +

    The value of fact5 will now be 120, which is the factorial of 5. Note that we passed 5 as the argument to the function.

    +

    It may be necessary to document what the function does, for each of the function to help the person who reads our code to understand it better. In order to do this Python allows the first line of the function body to be a string. This string is called as + Documentation String or + docstring. + docstrings prove to be very handy since there are number of tools which can pull out all the docstrings from Python functions and generate the documentation automatically from it. + docstrings for functions can be written as follows: +

    +
    def factorial(n):
    +  'Returns the factorial for the number n.'
    +  fact = 1
    +  for i in range(2, n):
    +    fact *= i
    +
    +  return fact
    +
    +
    +

    An important point to note at this point is that, a function can return any Python value or a Python object, which also includes a + Tuple. A + Tuple is just a collection of values and those values themselves can be of any other valid Python datatypes, including + Lists, + Tuples, + Dictionaries among other things. So effectively, if a function can return a tuple, it can return any number of values through a tuple +

    +

    Let us write a small function to swap two values:

    +
    def swap(a, b):
    +  return b, a
    +
    +c, d = swap(a, b)
    +
    +
    +

    Function scope --------------- The variables used inside the function are confined to the function's scope and doesn't pollute the variables of the same name outside the scope of the function. Also the arguments passed to the function are passed by-value if it is of basic Python data type:

    +
    def cant_change(n):
    +  n = 10
    +
    +n = 5
    +cant_change(n)
    +
    +
    +

    Upon running this code, what do you think would have happened to value of n which was assigned 5 before the function call? If you have already tried out that snippet on the interpreter you already know that the value of n is not changed. This is true of any immutable types of Python like + Numbers, + Strings and + Tuples. But when you pass mutable objects like + Lists and + Dictionaries the values are manipulated even outside the function: +

    +
    >>> def can_change(n):
    +...   n[1] = James
    +...
    +
    +>>> name = ['Mr.', 'Steve', 'Gosling']
    +>>> can_change(name)
    +>>> name
    +['Mr.', 'James', 'Gosling']
    +
    +
    +

    If nothing is returned by the function explicitly, Python takes care to return None when the funnction is called.

    +
    +

    +1. Default Arguments

    +

    There may be situations where we need to allow the functions to take the arguments optionally. Python allows us to define function this way by providing a facility called + Default Arguments. For example, we need to write a function that returns a list of fibonacci numbers. Since our function cannot generate an infinite list of fibonacci numbers, we need to specify the number of elements that the fibonacci sequence must contain. Suppose, additionally, we want to the function to return 10 numbers in the sequence if no option is specified we can define the function as follows: +

    +
    def fib(n=10):
    +  fib_list = [0, 1]
    +  for i in range(n - 2):
    +    next = fib_list[-2] + fib_list[-1]
    +    fib_list.append(next)
    +  return fib_list
    +
    +
    +

    When we call this function, we can optionally specify the value for the parameter n, during the call as an argument. Calling with no argument and argument with n=5 returns the following fibonacci sequences:

    +
    fib()
    +[0, 1, 1, 2, 3, 5, 8, 13, 21, 34]
    +fib(5)
    +[0, 1, 1, 2, 3]
    +
    +
    +
    +
    +

    +2. Keyword Arguments

    +

    When a function takes a large number of arguments, it may be difficult to remember the order of the parameters in the function definition or it may be necessary to pass values to only certain parameters since others take the default value. In either of these cases, Python provides the facility of passing arguments by specifying the name of the parameter as defined in the function definition. This is known as + Keyword Arguments. +

    +

    In a function call, + Keyword arguments can be used for each argument, in the following fashion: +

    +
    argument_name=argument_value
    +Also denoted as: keyword=argument
    +
    +def wish(name='World', greetings='Hello'):
    +  print "%s, %s!" % (greetings, name)
    +
    +
    +

    This function can be called in one of the following ways. It is important to note that no restriction is imposed in the order in which + Keyword arguments can be specified. Also note, that we have combined + Keyword arguments with + Default arguments in this example, however it is not necessary: +

    +
    wish(name='Guido', greetings='Hey')
    +wish(greetings='Hey', name='Guido')
    +
    +
    +

    Calling functions by specifying arguments in the order of parameters specified in the function definition is called as + Positional arguments, as opposed to + Keyword arguments. It is possible to use both + Positional arguments and + Keyword arguments in a single function call. But Python doesn't allow us to bungle up both of them. The arguments to the function, in the call, must always start with + Positional arguments which is in turn followed by + Keyword arguments: +

    +
    def my_func(x, y, z, u, v, w):
    +  # initialize variables.
    +  ...
    +  # do some stuff 
    +  ...
    +  # return the value
    +
    +
    +

    It is valid to call the above functions in the following ways:

    +
    my_func(10, 20, 30, u=1.0, v=2.0, w=3.0)
    +my_func(10, 20, 30, 1.0, 2.0, w=3.0)
    +my_func(10, 20, z=30, u=1.0, v=2.0, w=3.0)
    +my_func(x=10, y=20, z=30, u=1.0, v=2.0, w=3.0)
    +
    +
    +

    Following lists some of the invalid calls:

    +
    my_func(10, 20, z=30, 1.0, 2.0, 3.0)
    +my_func(x=10, 20, z=30, 1.0, 2.0, 3.0)
    +my_func(x=10, y=20, z=30, u=1.0, v=2.0, 3.0)
    +
    +
    +
    +
    +

    +3. Parameter Packing and Unpacking

    +

    The positional arguments passed to a function can be collected in a tuple parameter and keyword arguments can be collected in a dictionary. Since keyword arguments must always be the last set of arguments passed to a function, the keyword dictionary parameter must be the last parameter. The function definition must include a list explicit parameters, followed by tuple paramter collecting parameter, whose name is preceded by a *****, for collecting positional parameters, in turn followed by the dictionary collecting parameter, whose name is preceded by a ****** :

    +
    def print_report(title, *args, **name):
    +  """Structure of *args*
    +  (age, email-id)
    +  Structure of *name*
    +  {
    +      'first': First Name
    +      'middle': Middle Name
    +      'last': Last Name
    +  }
    +  """
    +
    +  print "Title: %s" % (title)
    +  print "Full name: %(first)s %(middle)s %(last)s" % name
    +  print "Age: %d\nEmail-ID: %s" % args
    +
    +
    +

    The above function can be called as. Note, the order of keyword parameters can be interchanged:

    +
    >>> print_report('Employee Report', 29, 'johny@example.com', first='Johny',
    +                 last='Charles', middle='Douglas')
    +Title: Employee Report
    +Full name: Johny Douglas Charles
    +Age: 29
    +Email-ID: johny@example.com
    +
    +
    +

    The reverse of this can also be achieved by using a very identical syntax while calling the function. A tuple or a dictionary can be passed as arguments in place of a list of *Positional arguments* or *Keyword arguments* respectively using ***** or ****** :

    +
    def print_report(title, age, email, first, middle, last):
    +  print "Title: %s" % (title)
    +  print "Full name: %s %s %s" % (first, middle, last)
    +  print "Age: %d\nEmail-ID: %s" % (age, email)
    +
    +>>> args = (29, 'johny@example.com')
    +>>> name = {
    +        'first': 'Johny',
    +        'middle': 'Charles',
    +        'last': 'Douglas'
    +        }
    +>>> print_report('Employee Report', *args, **name)
    +Title: Employee Report
    +Full name: Johny Charles Douglas
    +Age: 29
    +Email-ID: johny@example.com
    +
    +
    +
    +
    +

    +4. Nested Functions and Scopes

    +

    Python allows nesting one function inside another. This style of programming turns out to be extremely flexible and powerful features when we use + Python decorators. We will not talk about decorators is beyond the scope of this course. If you are interested in knowing more about + decorator programming in Python you are suggested to read: +

    +


    +      http://avinashv.net/2008/04/python-decorators-syntactic-sugar/
    +      http://personalpages.tds.net/~kent37/kk/00001.html
    +    

    +

    However, the following is an example for nested functions in Python:

    +
    def outer():
    +  print "Outer..."
    +  def inner():
    +    print "Inner..."
    +  print "Outer..."
    +  inner()
    +
    +>>> outer()
    +
    +
    +
    +
    +

    +5. map, reduce and filter functions

    +

    Python provides several built-in functions for convenience. The + map(), + reduce() and + filter() functions prove to be very useful with sequences like + Lists. +

    +

    The + map ( + function, + sequence) function takes two arguments: + function and a + sequence argument. The + function argument must be the name of the function which in turn takes a single argument, the individual element of the + sequence. The + map function calls + function(item), for each item in the sequence and returns a list of values, where each value is the value returned by each call to + function(item). + map() function allows to pass more than one sequence. In this case, the first argument, + function must take as many arguments as the number of sequences passed. This function is called with each corresponding element in the each of the sequences, or + None if one of the sequence is exhausted: +

    +
    def square(x):
    +  return x*x
    +
    +>>> map(square, [1, 2, 3, 4])
    +[1, 4, 9, 16]
    +
    +def mul(x, y):
    +  return x*y
    +
    +>>> map(mul, [1, 2, 3, 4], [6, 7, 8, 9])
    +
    +
    +

    The + filter ( + function, + sequence) function takes two arguments, similar to the + map() function. The + filter function calls + function(item), for each item in the sequence and returns all the elements in the sequence for which + function(item) returned True: +

    +
    def even(x):
    +  if x % 2:
    +    return True
    +  else:
    +    return False
    +
    +>>> filter(even, range(1, 10))
    +[1, 3, 5, 7, 9]
    +
    +
    +

    The + reduce ( + function, + sequence) function takes two arguments, similar to + map function, however multiple sequences are not allowed. The + reduce function calls + function with first two consecutive elements in the sequence, obtains the result, calls + function with the result and the subsequent element in the sequence and so on until the end of the list and returns the final result: +

    +
    def mul(x, y):
    +  return x*y
    +
    +>>> reduce(mul, [1, 2, 3, 4])
    +24
    +
    +
    +
    +

    +5.1. List Comprehensions

    +

    List Comprehension is a convenvience utility provided by Python. It is a syntatic sugar to create + Lists. Using + List Comprehensions one can create + Lists from other type of sequential data structures or other + Lists itself. The syntax of + List Comprehensions consists of a square brackets to indicate the result is a + List within which we include at least one + for clause and multiple + if clauses. It will be more clear with an example: +

    +
    >>> num = [1, 2, 3]
    +>>> sq = [x*x for x in num]
    +>>> sq
    +[1, 4, 9]
    +>>> all_num = [1, 2, 3, 4, 5, 6, 7, 8, 9]
    +>>> even = [x for x in all_num if x%2 == 0]
    +
    +
    +

    The syntax used here is very clear from the way it is written. It can be translated into english as, "for each element x in the list all_num, if remainder of x divided by 2 is 0, add x to the list."

    +
    +
    +
    + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/backup/paragraphlist.py --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/backup/paragraphlist.py Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,8 @@ +import sys +f=open(sys.argv[1],'r') +pid_list=[] +for i in f.readlines(): + if i.startswith('

    + + +Introduction + + + + + + + + + +

    +
    + +
    +
    +

    +
    +
    + +
    +

    +1. Introducing Linux

    +

    (Attribution : A significant chunk of the content under this section is based on data from Wikipedia and the Linux Documentation Project)

    +

    Linux (usually pronounced ˈlɪnəks') is a generic term referring to Unix-like computer operating systems based on the Linux kernel, where a kernel is the intermediate layer between the hardware and the applications. The kernel is, on an abstract level, the core of (most) operating systems, that manages the various system resources. The development of the Linux OS is considered the basis for Free and Open Source Software (FOSS) collaboration since typically the underlying source code can be used, modified freely, and redistributed by anyone under the terms of the GNU (a recursive acronym for "GNU's Not Unix!") Global Public License (GPL) and other free software licences. This freedom to access and reuse various components of a system, is one of the primary reasons for the popularity of Linux.

    +

    Linux is installed on a variety of computer hardware, that include mobile phones, embedded devices and supercomputers, but is infamous for its use in servers.

    +

    The name "Linux" comes from the Linux kernel, originally written in 1991 by Linus Torvalds. The rest of the system usually comprises components such as the Apache HTTP Server, the X Window System, the GNOME and KDE desktop environments, and utilities and libraries from the GNU Project (announced in 1983 by Richard Stallman). Commonly-used applications with desktop Linux systems include the Mozilla Firefox web-browser and the OpenOffice.org office application suite. The GNU contribution is the basis for the Free Software Foundation's preferred name GNU/Linux. The kernel's mascot is a penguin named "Tux". Mozilla Firefox and OpenOffice.org are open-source projects which can be run on most Operating Systems, including proprietary ones.

    +
    +

    +1.1. Historical Background

    +
    +

    +1.1.1. Events leading to the creation

    +
      +
    • The Unix operating system was developed in the 1960s and released for public use in 1970. Its accessibility and portability caused it to be widely adopted, copied and modified by academic institutions and businesses. Its design became influential to authors of other systems. Other free operating systems include the Berkeley Software Distribution (BSD), developed at the University of California at Berkeley, and MINIX which was released by Andrew S. Tanenbaum. The development and adoption of BSD and MINIX were limited due to various reasons, and this lack of a widely-adopted and free kernel triggered Linus Torvalds into starting his project.

    • +
    • In 1983, Richard Stallman started the GNU project with the goal of creating a free UNIX-like operating system. As part of this work, he wrote the GNU General Public License (GPL). By the early 1990s there was almost enough available software to create a full operating system. However, the GNU kernel, called Hurd, failed to attract enough attention from developers leaving GNU incomplete.

    • +
    +
    +
    +

    +1.1.2. The Creation of Linux

    +

    In 1991, Linus Torvalds began a project at the University of Helsinki that later became the Linux kernel. It was initially a terminal (command-line) emulator, which Torvalds used to access the large UNIX servers of the university. He wrote the program targeting just the hardware he was using and independent of an operating system because he wanted to use the functions of his computer with an 80386 processor. Development was done on Minix using the GNU C compiler. This application is still the main choice for compiling Linux today (although the code can be built with other compilers, such as the Intel C Compiler).

    +

    Torvalds continues to direct the development of the kernel. Stallman heads the Free Software Foundation, which in turn supports the GNU components. Finally, individuals and corporations develop third-party non-GNU components, which constitute a vast body of work and including kernel modules, and user applications and libraries. Linux vendors and communities combine and distribute the kernel, GNU components, and non-GNU components, with additional package management software in the form of Linux distributions.

    +
    +
    +
    +

    +1.2. Design and Implications

    +

    A Linux-based system is a modular Unix-like operating system, deriving much of its basic design from principles established in Unix earlier. Such a system uses a monolithic kernel, called the Linux kernel, which handles process control, networking, and peripheral and file system access. Device drivers are integrated directly with the kernel. Separate projects that interface with the kernel provide much of the system's higher-level functionality. The GNU userland is an important part of most Linux-based systems, providing the most common implementation of the C library, a popular shell, and many of the common Unix tools which carry out many basic operating system tasks. The graphical user interface (or GUI) used by most Linux systems is based on the "X Window System".

    +
    +

    +1.2.1. User Interface

    +

    Users can control a Linux-based system through a command line interface (or CLI), a graphical user interface (or GUI), or through controls attached to the associated hardware (this is common for embedded systems). For desktop systems, the default mode is usually the GUI. On desktop machines, "KDE", "GNOME" and "Xfce" are the most popular user interfaces,though a variety of additional user interfaces exist. Most popular user interfaces run on top of the "X Window System" (or X), which enables a graphical application running on one machine to be displayed and controlled from another in a network.

    +

    A Linux system also provides a CLI of some sort through a shell, which is the traditional way of interacting with a Unix system. A Linux distribution specialized for servers may use the CLI as its only interface. A “headless system” (system run without even a monitor) can be controlled by the command line via a remote-control protocol such as SSH or telnet. The CLI is particularly suited for automation of repetitive or delayed tasks, and provides very simple inter-process communication. A graphical terminal emulator program is often used to access the CLI from a Linux desktop.

    +
    +
    +

    +1.2.2. Development

    +

    The primary difference between Linux and many other popular contemporary operating systems is that the Linux kernel and other components are free and open source software. Linux is not the only such operating system, although it is by far the most widely used. Some free and open source software licenses are based on the principle of "copyleft", a kind of reciprocity: any work derived from a copyleft piece of software must also be copyleft itself. The most common free software license, the GNU GPL, is a form of copyleft, and is used for the Linux kernel and many of the components from the GNU project.

    +

    Linux based distributions are intended by developers for interoperability with other operating systems and established computing standards. Linux systems adhere to POSIX, SUS, ISO and ANSI standards where possible, although to date only one Linux distribution has been POSIX.1 certified, Linux-FT.Free software projects, although developed in a collaborative fashion, are often produced independently of each other. The fact that the software licenses explicitly permit redistribution, however, provides a basis for larger scale projects that collect the software produced by stand-alone projects and make it available all at once in the form of a Linux distribution.

    +

    A Linux distribution, commonly called a "distro", is a project that manages a remote collection of system software and application software packages available for download and installation through a network connection. This allows the user to adapt the operating system to his/her specific needs. Distributions are maintained by individuals, loose-knit teams, volunteer organizations, and commercial entities. A distribution can be installed using a CD that contains distribution-specific software for initial system installation and configuration. A package manager such as Synaptic or YAST allows later package upgrades and installations. A distribution is responsible for the default configuration of the installed Linux kernel, general system security, and more generally integration of the different software packages into a coherent whole.

    +
    +
    +

    +1.2.3. Community

    +

    A distribution is largely driven by its developer and user communities. Some vendors develop and fund their distributions on a volunteer basis. Examples include Debian and the Debian-based, Ubuntu. Others maintain a community version of their commercial distributions, as Red Hat does with Fedora.

    +

    In many cities and regions, local associations known as Linux Users Groups (LUGs) seek to promote their preferred distribution and by extension free software. They hold meetings and provide free demonstrations, training, technical support, and operating system installation to new users. Many Internet communities also provide support to Linux users and developers. Most distributions and free software / open source projects have IRC (Internet Relay Chat) chatrooms or newsgroups. Online forums are another means for support. Linux distributions host mailing lists; commonly there will be a specific topic such as usage or development for a given list. All these can be found simply by running an appropriate search on Google.

    +

    Although Linux distributions are generally available without charge, several large corporations sell, support, and contribute to the development of the components of the system and of free software. These include Dell, IBM, HP, Oracle, Sun Microsystems, Novell, Nokia. A number of corporations, notably Red Hat, have built their entire business around Linux distributions.

    +
    +
    +

    +1.2.4. Can I make a profit out of running a business involving Linux?

    +

    The answer is, "Yes!". The free software licenses, on which the various software packages of a distribution built on the Linux kernel are based, explicitly accommodate and encourage commercialization; the relationship between a Linux distribution as a whole and individual vendors may be seen as symbiotic. One common business model of commercial suppliers is charging for support, especially for business users. A number of companies also offer a specialized business version of their distribution, which adds proprietary support packages and tools to administer higher numbers of installations or to simplify administrative tasks. Another business model is to give away the software in order to sell hardware. Examples of corporations that are extensively (and sometimes exclusively) open-source and Linux-powered , with successful revenue generation models involving these, are Google, SUN, Mozilla, etc.

    +
    +
    +

    +1.2.5. Programming on Linux

    +

    Most Linux distributions support dozens of programming languages. The most common collection of utilities for building both Linux applications and operating system programs is found within the GNU toolchain, which includes the GNU Compiler Collection (GCC) and the GNU build system. Amongst others, GCC provides compilers for Ada, C, C++, Java, and Fortran. The Linux kernel itself is written to be compiled with GCC. Proprietary compilers for Linux include the Intel C++ Compiler, Sun Studio, and IBM XL C/C++ Compiler.

    +

    Most distributions also include support for PHP, Perl, Ruby, Python and other dynamic languages. Examples of languages that are less common, but still supported, are C# via the Mono project, sponsored by Novell, and Scheme. A number of Java Virtual Machines and development kits run on Linux, including the original Sun Microsystems JVM (HotSpot), and IBM's J2SE RE, as well as many open-source projects like Kaffe.

    +

    The two main frameworks for developing graphical applications are those of GNOME and KDE. These projects are based on the GTK+ and Qt widget toolkits, respectively, which can also be used independently of the larger framework. Both support a wide variety of languages. There are a number of Integrated Development Environments (IDEs) available including Anjuta, Code::Blocks, Eclipse, KDevelop, Lazarus, MonoDevelop, NetBeans, and Omnis Studio while the long-established editors Vim and Emacs remain popular.

    +
    +
    +
    +

    +1.3. Reasons for Using Linux

    +
    • Linux is free:

    +

    As in "free beer". Linux can be downloaded in its entirety from the Internet completely for free. No registration fees, no costs per user, free updates, and freely available source code in case you want to change the behavior of your system. +Most of all, Linux is free as in "free speech": +The license commonly used is the GNU Public License (GPL). The license says that anybody who may want to do so, has the right to change Linux and eventually to redistribute a changed version, on the one condition that the code is still available after redistribution. In practice, you are free to grab a kernel image and sell the new code, as long as your customers can still have a copy of that code.

    +
    • Linux is portable to any hardware platform:

    +

    A vendor, who wants to sell a new type of computer and who does not know what kind of OS his/her new machine will run, can take a Linux kernel and make it work on his/her hardware, because documentation related to this activity is freely available.

    +
    • Linux was made to keep on running:

    +

    As with UNIX, a Linux system expects to run without rebooting all the time. That is why a lot of tasks are being executed at night or scheduled automatically for other times, resulting in higher availability during busier periods and a more balanced use of the hardware. This property allows for Linux to be applicable to environments where people do not have the time or the possibility to control their systems constantly.

    +
    • Linux is secure and versatile:

    +

    The security model used in Linux is based on the UNIX idea of security, which is known to be robust and of proven quality. But Linux is not only safe from attacks from the Internet: it will adapt equally to other situations, utilizing the same high standards for security.

    +
    • Linux is scalable:

    +

    From a Palmtop with 2 MB of memory to a petabyte storage cluster with hundreds of nodes: add or remove the appropriate packages and Linux fits all. One does not need a supercomputer anymore,because you can use Linux to do big things using the building blocks provided with the system. If one wants to do little things, such as making an operating system for an embedded processor or just recycling your old 486, Linux will do that as well.

    +
    • The Linux OS and Linux applications have very short debug−times:

    +

    Because Linux has been developed and tested by thousands of people, both errors and people to fix them are found very quickly. It often happens that there are only a couple of hours between discovery and fixing of a bug.

    +
    +
    +
    +

    +2. Getting Started

    +
    +

    +2.1. Logging in, activating the user interface and logging out

    +

    In order to work on a Linux system directly, one needs to provide a user name and password. You always need to authenticate to the system. Most PC−based Linux systems have two basic modes for a system to run in: either quick and clean in text console mode,which includes with mouse, multitasking and multi−user features, or in graphical console mode, which looks better but eats more system resources.

    +
    +

    +2.1.1. Graphical Mode

    +

    This is the default nowadays on most desktop computers. You know you will be connecting to the system using graphical mode when you are first asked for your user name, and then to type your password.

    +

    To log in, make sure the mouse pointer is in the login window, provide your user name and password to the system and click OK or press Enter. +It is generally considered a bad idea to connect (graphically) using the root user name, the system adminstrator's account, since the use of graphics includes running a lot of extra programs, in root's case with a lot of extra permissions. To keep all risks as low as possible, use a normal user account to connect graphically. But there are enough risks to keep this in mind as a general advice, for all use of the root account: only log in as root when extra privileges are required.

    +

    After entering your user name/password combination, it can take a little while before the graphical environment is started, depending on the CPU speed of your computer, on the software you use and on your personal settings.

    +

    To continue, you will need to open a terminal window or xterm for short (X being the name for the underlying software supporting the graphical environment). This program can be found in the Applications−>Utilities->System Tools or Internet menu, depending on what window manager you are using. There might be icons that you can use as a shortcut to get an xterm window as well, and clicking the right mouse button on the desktop background will usually present you with a menu containing a terminal window application.

    +

    While browsing the menus, you will notice that a lot of things can be done without entering commands via the keyboard. For most users, the good old point−n−click method of dealing with the computer will do. But for those who want to enter the "heart" of the system, a tool stronger than a mouse will be required to handle the various tasks. This tool is the shell, and when in graphical mode, we activate our shell by opening a terminal window.

    +

    A terminal window should always show a command prompt when you open one. This terminal shows a standard prompt, which displays the user's login name, and the current working directory, represented by the twiddle (~)

    +

    Another common form for a prompt is this one: +[

    +
    +

    user@host
    +

    dir]

    +

    In the above example, user will be your login name, hosts the name of the machine you are working on, and dir an indication of your current location in the file system. Prompts can display all kinds of information, but they are not part of the commands you are giving to your system. To disconnect from the system in graphical mode, you need to close all terminal windows and other applications. After that, hit the logout icon or find Log Out in the menu. Closing everything is not really necessary, and the system can do this for you, but session management might put all currently open applications back on your screen when you connect again, which takes longer and is not always the desired effect. However, this behavior is configurable.

    +

    When you see the login screen again, asking to enter user name and password, logout was successful.

    +
    +
    +

    +2.1.2. Text Mode

    +

    One is in text mode when the whole screen is black, showing (in most cases white) characters. A text mode login screen typically shows some information about the machine you are working on, the name of the machine and a prompt waiting for you to log in.

    +

    The login is different from a graphical login, in that you have to hit the Enter key after providing your user name, because there are no buttons on the screen that you can click with the mouse. Then you should type your password, followed by another Enter. You will not see any indication that you are entering something, not even an asterisk, and you won't see the cursor move. But this is normal on Linux and is done for security +reasons.

    +

    When the system has accepted you as a valid user, you may get some more information, called the message of the day, which can be anything. Additionally, it is popular on UNIX systems to display a fortune cookie, which contains some general wise or unwise (this is up to you) thoughts. After that, you will be given a shell, indicated with the same prompt that you would get in graphical mode.

    +

    Also in text mode: log in as root only to do setup and configuration that absolutely requires administrator privileges, such as adding users, installing software packages, and performing network and other system configuration. Once you are finished, immediately leave the special account and resume your work as a non−privileged user.

    +

    Logging out is done by entering the logout command, followed by Enter. You are successfully disconnected from the system when you see the login screen again.Don't power−off the computer after logging out. It is not meant to be shut off without application of the proper procedures for halting the system. Powering it off without going through the halting process might cause severe damage!

    +
    +
    +
    +
    +

    +3. Basic Commands

    +
    +

    +3.1. ls

    +

    When invoked without any arguments, ls lists the files in the current working directory. A directory that is not the current working directory can be specified and ls will list the files there. The user also may specify any list of files and directories. In this case, all files and all contents of specified directories will be listed. The name ls is derived from list segments which was used in earlier systems.

    +

    Files whose names start with "." are not listed, unless the -a flag is specified or the files are specified explicitly.

    +

    Without options, ls displays files in a bare format. This bare format however makes it difficult to establish the type, permissions, and size of the files. The most common options to reveal this information or change the list of files are:

    +<block_quote>
      +
    • -l long format, displaying Unix file types, permissions, number of hard links, owner, group, size, date, and filename

    • +
    • -F appends a character revealing the nature of a file, for example, * for an executable, or / for a directory. Regular files have no suffix.

    • +
    • -a lists all files in the given directory, including those whose names start with "." (which are hidden files in Unix). By default, these files are excluded from the list.

    • +
    • -R recursively lists subdirectories. The command ls -R / would therefore list all files.

    • +
    • -d shows information about a symbolic link or directory, rather than about the link's target or listing the contents of a directory.

    • +
    • -t sort the list of files by modification time.

    • +
    • -h print sizes in human readable format. (e.g., 1K, 234M, 2G, etc.)

    • +
    </block_quote>

    In some environments, providing the option --color (for GNU ls) or -G (FreeBSD ls) causes ls to highlight different types of files with different colors, instead of with characters as -F would. To determine what color to use for a file, GNU ls checks the Unix file type, the file permissions, and the file extension, while FreeBSD ls checks only the Unix file type and file permissions.:

    +
    $ ls
    +jeeves.rst psmith.html blandings.html
    +$ ls -l
    +drwxr--r--   1 plum  editors   4096  jeeves
    +-rw-r--r--   1 plum  editors  30405  psmith
    +-r-xr-xr-x   1 plum  plum      8460  blandings
    +

    Here "$" actually is the beginning of the prompt. This is typical in most Unix-based systems.

    +
    +
    +

    +3.2. date

    +

    The Unix date command displays the time and date. The super-user can use it to set the system clock.

    +

    With no options, the date command displays the current date and time, including the abbreviated day name, abbreviated month name, day of the month, the time separated by colons, the timezone name, and the year. For example:

    +
    $date
    +Tue Sep  8 12:01:45 IST 2009
    +

    On some systems to set the current date and time to September 8, 2004 01:22 you type:

    +
    $date --set="20040908 01:22"
    +

    In order to view the various options for the date command, type:

    +
    $man date
    +

    This will take you to the "Manual" page comprising of all the details on the date command. You can return to the terminal from the "man" page by pressing the Esc key in the keyboard and typing ":q" in that order.

    +
    +
    +

    +3.3. cd

    +

    This stands for "change directory". When one wants to go up to the parent directory, bypassing the tree of directories one has entered, “ cd ..” can be used.

    +

    One dot '.' represents the current directory while two dots '..' represent the parent directory.

    +

    “ cd -” will return you to the previous directory (a bit like an “undo”).

    +

    You can also use cd absolute path or cd relative path (see below):

    +

    Absolute paths:

    +<block_quote>

    An “ absolute path” is easily recognised from the leading forward slash, /. The / means that you start at the top level directory and continue down.

    </block_quote>

    For example to get to /boot/grub you would type:

    +
    $cd /boot/grub
    +

    This is an absolute path because you start at the top of the hierarchy and go downwards from there (it doesn't matter where in the filesystem you were when you typed the command).

    +

    Relative paths:

    +<block_quote>

    A “ relative path” doesn't have a preceding slash. Use a relative path when you start from a directory below the top level directory structure. This is dependent on where you are in the filesystem.

    +

    For example if you are in root's home directory and want to get to /root/music, you type:

    +
    $ cd music
    </block_quote>

    Please note that there is no / using the above cd command. Using a / would cause this to be an absolute path, working from the top of the hierarchy downward.

    +
    +
    +

    +3.4. who

    +

    The standard Unix command who displays a list of users who are currently logged into a computer.

    +

    The who command is related to the command w, which provides the same information but also displays additional data and statistics.:

    +
    $who
    +beeblebrox tty7         2009-09-08 10:50 (:0)
    +beeblebrox pts/0        2009-09-08 11:25 (:0.0)
    +dumbledore pts/1        2009-09-08 18:11 (potter.xyz.in)
    +beeblebrox pts/2        2009-09-08 18:53 (:0.0)
    +

    The command can be invoked with the arguments am i or am I (so it is invoked as who am i or * who am I*), showing information about the current terminal only (see the -m option below, of which this invocation is equivalent).

    +

    In order to find out the various options that can be appended to the who command, check the man page by typing out the following in the terminal:

    +
    $man who
    +

    This will take you to the "Manual" page containing details about the who command

    +
    +
    +

    +3.5. mkdir

    +

    This command is used to make a new directory. Normal usage is as straightforward as follows:

    +
    $mkdir name_of_directory
    +

    Where name_of_directory is the name of the directory one wants to create. When typed as above (ie. normal usage), the new directory would be created within the current directory. On Unix, multiple directories can be specified, and mkdir will try to create all of them.

    +
    +

    +3.5.1. Options

    +

    On Unix-like operating systems, mkdir takes options. Three of the most common options are:

    +<block_quote>
      +
    • -p: will also create all directories leading up to the given directory that do not exist already. If the given directory already exists, ignore the error.

    • +
    • -v: display each directory that mkdir creates. Most often used with -p.

    • +
    • -m: specify the octal permissions of directories created by mkdir.

    • +
    </block_quote>

    -p is most often used when using mkdir to build up complex directory hierarchies, in case a necessary directory is missing or already there. -m is commonly used to lock down temporary directories used by shell scripts.

    +
    +
    +

    +3.5.2. Examples

    +

    An example of -p in action is:

    +
    $mkdir -p /tmp/a/b/c
    +

    If /tmp/a exists but /tmp/a/b does not, mkdir will create /tmp/a/b before creating /tmp/a/b/c.

    +

    And an even more powerful command, creating a full tree at once (this however is a Shell extension, nothing mkdir does itself):

    +
    $mkdir -p tmpdir/{trunk/sources/{includes,docs},branches,tags}
    +

    This will create:

    +<definition_list><definition_list_item><term>tmpdir - branches</term><definition>
      +
    • tag

    • +
    • <definition_list><definition_list_item><term>trunk - sources - includes</term><definition>
      • docs

      </definition>
      </definition_list_item>
      </definition_list>
    • +
    </definition>
    </definition_list_item>
    </definition_list>
    +
    +
    +
    +
    +

    +4. Getting Help

    +
    +

    +4.1. apropos and whatis

    +

    This is a command to search the manual pages files in Unix and Unix-like operating systems.

    +
    $ apropos grep
    +egrep       egrep (1)       Search a file for a pattern using full regular expressions
    +fgrep       fgrep (1)       Search a file for a fixed-character string
    +fmlgrep     fmlgrep (1)     Search a file for a pattern
    +grep        grep (1)        Search a file for a pattern
    +gzgrep      gzgrep (1)      Search a possibly compressed file for a regular expression
    +nisgrep     nismatch (1)    Utilities for searching NIS+ tables
    +pgrep       pgrep (1)       Find or signal a process by name or other attribute
    +zgrep       zgrep (1)       Search a possibly compressed file for a regular expression
    +...
    +

    In this example, the user uses apropos to search for the string "grep", and apropos returns the indicated man pages that include the term "grep".

    +

    A short index of explanations for commands is available using the whatis command, like in the examples below:

    +
    $whatis ls
    +ls (1)           - list directory contents
    +

    This displays short information about a command, and the first section in the collection of man pages that contains an appropriate page.

    +

    If you don't know where to get started and which man page to read, apropos gives more information. Say that you do not know how to start a browser, then you could enter the following command:

    +
    $apropos browser
    +gmusicbrowser (1)    - Jukebox for large collections of audio files
    +infobrowser (1)      - read Info documents
    +libsmbclient (7)     - An extension library for browsers and that               can be used...
    +opera (1)            - a standards-compliant graphical Web browser
    +sensible-browser (1) - sensible editing, paging, and web browsing
    +smbtree (1)          - A text based smb network browser
    +tvtk_doc (1)         - A GUI based TVTK documentation search browser.
    +viewres (1)          - graphical class browser for Xt
    +w3m (1)              - a text based Web browser and pager
    +www-browser (1)      - a text based Web browser and pager
    +...
    +
    +
    +

    +4.2. man

    +

    Man pages (short for "manual pages") are the extensive documentation that comes preinstalled with almost all substantial Unix and Unix-like operating systems. The Unix command used to display them is man. Each page is a self-contained document.

    +

    To read a manual page for a Unix command, one can use:

    +
    $ man <command_name>
    +

    at a shell prompt; for example, "man ftp". In order to simplify navigation through the output, man generally uses the less terminal pager.

    +

    Pages are traditionally referred to using the notation "name(section)"; for example, ftp(1). The same page name may appear in more than one section of the manual, this can occur when the names of system calls, user commands, or macro packages coincide. Two examples are man(1) and man(7), or exit(2) and exit(3). The syntax for accessing the non-default manual section varies between different man implementations. On Linux and BSD, for example, the syntax for reading *printf(3) is:

    +
    $man 3 printf
    +

    Another example:

    +
    $man man
    +

    The previous example will take you to the "Manual" page entry about manual pages!

    +
    +

    +4.2.1. Layout

    +

    All man pages follow a common layout that is optimized for presentation on a simple ASCII text display, possibly without any form of highlighting or font control. Sections present may include:

    +<definition_list><definition_list_item><term>NAME</term><definition>

    The name of the command or function, followed by a one-line description of what it does.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SYNOPSIS</term><definition>

    In the case of a command, you get a formal description of how to run it and what command line options it takes. For program functions, a list of the parameters the function takes and which header file contains its definition. For experienced users, this may be all the documentation they need.

    </definition>
    </definition_list_item>
    <definition_list_item><term>DESCRIPTION</term><definition>

    A textual description of the functioning of the command or function.

    </definition>
    </definition_list_item>
    <definition_list_item><term>EXAMPLES</term><definition>

    Some examples of common usage.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SEE ALSO</term><definition>

    A list of related commands or functions.

    </definition>
    </definition_list_item>
    </definition_list>

    Other sections may be present, but these are not well standardized across man pages. Common examples include: OPTIONS, EXIT STATUS, ENVIRONMENT, KNOWN BUGS, FILES, AUTHOR, REPORTING BUGS, HISTORY and COPYRIGHT.

    +

    These days virtually every Unix command line application comes with its man page, and many Unix users perceive a lack of man pages as a sign of low quality; indeed, some projects, such as Debian, go out of their way to write man pages for programs lacking one. Few alternatives to man have enjoyed much popularity, with the possible exception of the GNU project's "info" system, an early and simple hypertext system.

    +

    However, the format of a single page for each application, the lack of classification within the sections and the relatively unsophisticated formatting facilities have motivated the development of alternative documentation systems, such as the previously mentioned "info" system.

    +

    Most Unix GUI applications (particularly those built using the GNOME and KDE development environments) now provide end-user documentation in HTML and include embedded HTML viewers such as yelp for reading the help within the application.

    +

    Usually the man pages are written in English. Translations into other languages can be also available on the system.

    +

    The default format of the man pages is troff, with either the macro package man (appearance oriented) or on some systems mdoc (semantic oriented). This makes it possible to typeset a man page to PostScript, PDF and various other formats for viewing or printing.

    +
    +
    +
    +

    +4.3. info

    +

    info is a software utility which forms a hypertextual, multipage documentation and help viewer working on a command line interface, useful when there is no GUI available.

    +

    The syntax is

    +
    $ info <command_name>
    +

    info processes info files, which are Texinfo formatted files, and presents the documentation as a tree, with simple commands to traverse the tree and to follow cross references. For instance

    +<block_quote>
      +
    • n goes to the next page.

    • +
    • p goes to the previous page.

    • +
    • u goes to the upper page.

    • +
    • l goes to the last(visited) node

    • +
    • To follow a cross reference, the cursor can be moved over a link (a word preceded by a <title_reference>*</title_reference>) and enter pressed.

    • +
    </block_quote>

    info was initially written for use with GNU/Linux and then ported to other Unix-like operating systems.

    +
    +
    +

    +4.4. --help

    +

    Most GNU commands support the --help, which gives a short explanation about how to use the command and a list of available options. Below is the output of this option with the cat command:

    +
    $ userprompt@host: cat --help
    +Usage: cat [OPTION] [FILE]...
    +Concatenate FILE(s), or standard input, to standard output.
    +
    +  -A, --show-all           equivalent to -vET
    +  -b, --number-nonblank    number nonempty output lines
    +  -e                       equivalent to -vE
    +  -E, --show-ends          display $ at end of each line
    +  -n, --number             number all output lines
    +  -s, --squeeze-blank      suppress repeated empty output lines
    +  -t                       equivalent to -vT
    +  -T, --show-tabs          display TAB characters as ^I
    +  -u                       (ignored)
    +  -v, --show-nonprinting   use ^ and M- notation, except for LFD and              TAB
    +  --help     display this help and exit
    +  --version  output version information and exit
    +
    +With no FILE, or when FILE is -, read standard input.
    +
    +Examples:
    +  cat f - g  Output f's contents, then standard input, then g's           contents.
    +  cat        Copy standard input to standard output.
    +
    +Report bugs to <bug-coreutils@gnu.org>.
    +
    +
    +
    +

    +5. Basic file handling

    +
    +

    +5.1. cp

    +

    cp is the command entered in a Unix shell to copy a file from one place to another, possibly on a different filesystem. The original file remains unchanged, and the new file may have the same or a different name.

    +
    +

    +5.1.1. Usage

    +

    To copy a file to another file:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ][ -- ] SourceFile TargetFile
    +

    To copy a file to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -r | -R ] [ -- ] SourceFile ...              TargetDirectory
    +

    To copy a directory to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -- ] { -r | -R }
    +SourceDirectory ... TargetDirectory
    +
    +
    +

    +5.1.2. Flags

    +

    -f (force) – specifies removal of the target file if it cannot be opened for write operations. The removal precedes any copying performed by the cp command.

    +

    -P – makes the cp command copy symbolic links. The default is to follow symbolic links, that is, to copy files to which symbolic links point.

    +

    -i (interactive) – prompts you with the name of a file to be overwritten. This occurs if the TargetDirectory or TargetFile parameter contains a file with the same name as a file specified in the SourceFile or SourceDirectory parameter. If you enter y or the locale's equivalent of y, the cp command continues. Any other answer prevents the cp command from overwriting the file.

    +

    -p (preserve) – duplicates the following characteristics of each SourceFile/SourceDirectory in the corresponding TargetFile and/or TargetDirectory:

    +<block_quote>
      +
    • The time of the last data modification and the time of the last access.

    • +
    • The user ID and group ID (only if it has permissions to do this)

    • +
    • The file permission bits and the SUID and SGID bits.

    • +
    </block_quote>

    -R (recursive) – copy directories (recursively copying all the contents)

    +
    +
    +

    +5.1.3. Examples

    +

    To make a copy of a file in the current directory, enter:

    +
    $ cp prog.c prog.bak
    +

    This copies prog.c to prog.bak. If the prog.bak file does not already exist, the cp command creates it. If it does exist, the cp command replaces it with a copy of the prog.c file.

    +

    To copy a file in your current directory into another directory, enter:

    +
    $ cp zaphod /home/books/hhgg
    +

    This copies the jones file to /home/books/hhgg/zaphod.

    +

    To copy a file to a new file and preserve the modification date, time, and access control list associated with the source file, enter:

    +
    $ cp -p martin_luther_king martin_luther_king.jr
    +

    This copies the martin_luther_king file to the martin_luther_king.jr file. Instead of creating the file with the current date and time stamp, the system gives the martin_luther_king.jr file the same date and time as the martin_luther_king file. The martin_luther_king.jr file also inherits the martin_luther_king file's access control protection.

    +

    To copy all the files in a directory to a new directory, enter:

    +
    $ cp /home/galactica/clients/* /home/hhgg/customers
    +

    This copies only the files in the clients directory to the customers directory.

    +

    To copy a directory, including all its files and subdirectories, to another directory, enter:

    +<block_quote>

    $ cp -R /home/hhgg/clients /home/hhgg/customers

    </block_quote>

    This copies the clients directory, including all its files, subdirectories, and the files in those subdirectories, to the customers/clients directory.

    +

    To copy a specific set of files of any extension to another directory, enter:

    +
    $ cp zaphod arthur ford /home/hhgg/clients
    +

    This copies the zaphod, arthur, and ford files in your current working directory to the /home/hhgg/clients directory.

    +

    To use pattern-matching characters to copy files, enter:

    +
    $ cp programs/*.py .
    +

    This copies the files in the programs directory that end with .py to the current directory, signified by the single "." (dot). You must type a space between the py and the final dot.

    +
    +
    +
    +

    +5.2. mv

    +

    mv (short for move) is a Unix command that moves one or more files or directories from one place to another. The original file is deleted, and the new file may have the same or a different name. If possible (i.e. when the original and new files are on the same file system), mv will rename the file instead. Write permission is required on all directories being modified.

    +
    +

    +5.2.1. Conflicting existing file

    +

    In all cases, when a file is moved to have the name of an existing file (in the same directory), the existing file is deleted. If the existing file is not writable but is in a directory that is writable, then the mv command asks for confirmation if possible (i.e. if run from a terminal) before proceeding, unless the -f (force) option is used.

    +
    +
    +

    +5.2.2. Differences with copy and delete

    +

    Note that, usually, when moving files within the same volume, moving (and/or renaming) is not the same as simply copying and then deleting the original. When moving a file, the link is simply removed from the old parent directory and added to the new parent directory. However, the file itself is untouched (i.e. it has the same inodes and resides at the same place on the disk). For example, you cannot copy a file you cannot read, but you can move (and/or rename) it (provided you have write permission to its old and new parent directories). Also, suppose there is a non-empty directory you do not have write permission to. You cannot delete this directory (since you cannot delete its contents); but you can move (and/or rename) it. Also, since moving between filenames on a single volume does not involve copying, it is faster and does not place strain of lots of reads and writes on the disk. Moving files across different volumes, however, does necessitate copying and deleting.

    +
    +
    +

    +5.2.3. Examples

    +
    $ mv myfile mynewfilename    renames a file
    +$ mv myfile otherfilename    renames a file and deletes the existing            file "myfile"
    +$ mv myfile /myfile          moves 'myfile' from the current            directory to the root directory
    +$ mv myfile dir/myfile       moves 'myfile' to 'dir/myfile' relative            to the current directory
    +$ mv myfile dir              same as the previous command (the          filename is implied to be the same)
    +$ mv myfile dir/myfile2      moves 'myfile' to dir and renames it to            'myfile2'
    +$ mv foo bar baz dir         moves multiple files to directory dir
    +$ mv --help                  shows a very concise help about the                syntax of the command
    +$ man mv                     prints an extensive user manual for                'mv' in the terminal
    +

    In all cases, the file or files being moved or renamed can be a directory.

    +

    Note that when the command is called with two arguments (as mv name1 name2 or mv name1 /dir/name2), it can have three different effects, depending on whether name2 does not exist, is an existing file, or is an existing directory. If the user intends to refer to an existing directory, /. (or in some Unix versions / is sufficient) may be appended to the name to force the system to check this. To move a file to a new directory, the directory must be created first.

    +
    +
    +
    +

    +5.3. rm

    +

    rm (short for "remove") is one of several basic Unix command lines that operates on files. It is used to delete files from a filesystem. The data is not actually destroyed. Only the index listing where the file is stored is destroyed, and the storage is made available for reuse. There are undelete utilities that will attempt to reconstruct the index and can bring the file back if the parts were not reused.

    +

    Here's example to remove a file named "foo" from a directory, here shown with the -i option:

    +
    $ rm -i foo
    +remove foo? y
    +
    +

    +5.3.1. Options

    +

    Common options that rm accepts include:

    +<block_quote>
      +
    • -r, which removes directories, removing the contents recursively beforehand (so as not to leave files without a directory to reside in) ("recursive")

    • +
    • -i, which asks for every deletion to be confirmed ("interactive")

    • +
    • -f, which ignores non-existent files and overrides any confirmation prompts ("force")

    • +
    • -v, which shows what is being removed as it happens ("verbose")

    • +
    </block_quote>

    rm is often aliased to "rm -i" so as to avoid accidental deletion of files. If a user still wishes to delete a large number of files without confirmation, they can manually cancel out the -i argument by adding the -f option (as the option specified later on the expanded command line "rm -i -f" takes precedence).

    +

    rm -rf (variously, rm -rf /, rm -rf <title_reference>*</title_reference>, and others) is frequently used in jokes and anecdotes about Unix disasters. The rm -rf variant of the command, if run by a superuser on the root directory, would cause the contents of every writable mounted filesystem on the computer to be deleted.

    +

    rm is often used in conjunction with xargs to supply a list of files to delete:

    +
    xargs rm < filelist
    +

    When rm is used on a symbolic link, it deletes the link, but does not affect the target of the link.

    +
    +
    +

    +5.3.2. Permissions

    +

    Usually, on most filesystems, deleting a file requires write permission on the parent directory (and execute permission, in order to enter the directory in the first place). (Note that, confusingly for beginners, permissions on the file itself are irrelevant. However, GNU rm asks for confirmation if a write-protected file is to be deleted, unless the -f option is used.)

    +

    To delete a directory (with rm -r), one must delete all of its contents recursively. This requires that one must have read and write and execute permission to that directory (if it's not empty) and all non-empty subdirectories recursively (if there are any). The read permissions are needed to list the contents of the directory in order to delete them. This sometimes leads to an odd situation where a non-empty directory cannot be deleted because one doesn't have write permission to it and so cannot delete its contents; but if the same directory were empty, one would be able to delete it.

    +

    If a file resides in a directory with the sticky bit set, then deleting the file requires one to be the owner of the file.

    +
    +
    +
    +
    +

    +6. Command Line Arguments

    +

    In computer command line interfaces, a command line argument is an argument sent to a program being called. In general, a program can take any number of command line arguments, which may be necessary for the program to run, or may even be ignored, depending on the function of that program.

    +

    For example, in Unix and Unix-like environments, an example of a command-line argument is:

    +
    rm file.s
    +

    "file.s" is a command line argument which tells the program rm to remove the file "file.s".

    +

    Programming languages such as C, C++ and Java allow a program to interpret the command line arguments by handling them as string parameters in the main function.

    +

    A command line option or simply option (also known as a command line parameter, flag, or a switch) is an indication by a user that a computer program should change its default output.

    +

    Long options are introduced via "--", and are typically whole words. For example, ls --long --classify --all. Arguments to long options are provided with "=", as ls --block-size=1024. Some Unix programs use long options with single dashes, for example MPlayer as in mplayer -nosound.

    +

    Linux also uses "--" to terminate option lists. For example, an attempt to delete a file called -file1 by using rm -file1 may produce an error, since rm may interpret -file1 as a command line switch. Using rm -- -file1 removes ambiguity.

    +
    +
    +

    +7. Basic Text Processing

    +
    +

    +7.1. head

    +

    head is a program on Unix and Unix-like systems used to display the first few lines of a text file or piped data. The command syntax is:

    +
    $ head [options] <file_name>
    +

    By default, head will print the first 10 lines of its input to the standard output. The number of lines printed may be changed with a command line option. The following example shows the first 20 lines of filename:

    +
    $ head -n 20 filename
    +

    This displays the first 5 lines of all files starting with foo:

    +
    $ head -n 5 foo*
    +

    Some versions omit the n and just let you say -5.

    +
    +

    +7.1.1. Flags

    +
    -c <x number of bytes> Copy first x number of bytes.
    +

    Other options: sed

    +

    Many early versions of Unix did not have this command, and so documentation and books had sed do this job:

    +
    sed 5q foo
    +

    This says to print every line (implicit), and quit after the fifth.

    +
    +
    +
    +

    +7.2. tail

    +

    tail is a program on Unix and Unix-like systems used to display the last few lines of a text file or piped data.

    +

    The command-syntax is:

    +
    $ tail [options] <file_name>
    +

    By default, tail will print the last 10 lines of its input to the standard output. With command line options the number of lines printed and the printing units (lines, blocks or bytes) may be changed. The following example shows the last 20 lines of filename:

    +
    $ tail -n 20 filename
    +

    This example shows the last 15 bytes of all files starting with foo:

    +
    $ tail -c 15 foo*
    +

    This example shows all lines of filename from the second line onwards:

    +
    $ tail -n +2 filename
    +

    Using an older syntax (still used in Sun Solaris as the -n option is not supported), the last 20 lines and the last 50 bytes of filename can be shown with the following command:

    +
    $ tail -20 filename
    +$ tail -50c filename
    +

    However this syntax is now obsolete and does not conform with the POSIX 1003.1-2001 standard. Even if still supported in current versions, when used with other options (like -f, see below), these switches could not work at all.

    +
    +

    +7.2.1. File monitoring

    +

    tail has a special command line option -f (follow) that allows a file to be monitored. Instead of displaying the last few lines and exiting, tail displays the lines and then monitors the file. As new lines are added to the file by another process, tail updates the display. This is particularly useful for monitoring log files. The following command will display the last 10 lines of messages and append new lines to the display as new lines are added to messages:

    +
    $ tail -f /var/adm/messages
    +

    To interrupt tail while it is monitoring, break-in with Ctrl+C. This command can be run "in the background" with &, see job control.

    +

    If you have a command's result to monitor, you can use the watch command.

    +
    +
    +
    +

    +7.3. cut

    +

    In computing, cut is a Unix command line utility which is used to extract sections from each line of input — usually from a file.

    +

    Extraction of line segments can typically be done by bytes (-b), characters (-c), or fields (-f) separated by a delimiter (-d — the tab character by default). A range must be provided in each case which consists of one of N, N-M, N- (N to the end of the line), or -M (beginning of the line to M), where N and M are counted from 1 (there is no zeroth value). Since version 6, an error is thrown if you include a zeroth value. Prior to this the value was ignored and assumed to be 1.

    +

    Assuming a file named file containing the lines:

    +
    foo:bar:baz:qux:quux
    +one:two:three:four:five:six:seven
    +alpha:beta:gamma:delta:epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +

    To output the fourth through tenth characters of each line:

    +
    $ cut -c 4-10 file
    +

    This gives the output:

    +
    :bar:ba
    +:two:th
    +ha:beta
    +

    To output the fifth field through the end of the line of each line using the colon character as the field delimiter:

    +
    $ cut -d : -f 5- file
    +

    This gives the output:

    +
    quux
    +five:six:seven
    +epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +
    +
    +

    +7.4. paste

    +

    paste is a Unix command line utility which is used to join files horizontally (parallel merging) by outputting lines consisting of the sequentially corresponding lines of each file specified, separated by tabs, to the standard output. It is effectively the horizontal equivalent to the utility cat command which operates on the vertical plane of two or more files.

    +

    To paste several columns of data together into the file www from files who, where, and when:

    +
    $ paste who where when > www
    +

    If the files contain:

    +++++
    +

    This creates the file named www containing:

    +
    Batman            GothamCity       January 3
    +Trillian          Andromeda        February 4
    +Jeeves            London           March 19
    +
    +
    +
    +

    +8. Shell Meta Characters

    +

    Unix recognizes certain special characters, called "meta characters," as command directives. The shell meta characters are recognized anywhere they appear in the command line, even if they are not surrounded by blank space. For that reason, it is safest to only use the characters A-Z, a-z, 0-9, and the period, dash, and underscore characters when naming files and directories on Unix. If your file or directory has a shell meta character in the name, you will find it difficult to use the name in a shell command.

    +

    The shell meta characters include:

    +

    / < > ! $ % ^ & * | { } [ ] " ' ` ~ ;

    +

    Different shells may differ in the meta characters recognized.

    +

    As an example,

    +
    $ ls file.*
    +

    run on a directory containing the files file, file.c, file.lst, and myfile would list the files file.c and file.lst. However,:

    +
    $ ls file.?
    +

    run on the same directory would only list file.c because the ? only matches one character, no more, no less. This can save you a great deal of typing time. For example, if there is a file called california_cornish_hens_with_wild_rice and no other files whose names begin with 'c', you could view the file without typing the whole name by typing this:

    +
    $ more c*
    +

    because the c* matches that long file name.

    +

    Filenames containing metacharacters can pose many problems and should never be intentionally created. If you do find that you've created a file with metacharacters, and you would like to remove it, you have three options. You may use wildcards to match metacharacter, use the to directly enter the filename, or put the command in double quotes (except in the case of double quotes within the file name, these must be captured with one of the first two methods). For example, deleting a file named <title_reference>"``*`|more</title_reference>"` can be accomplished with:

    +
    $ rm ??more
    +

    or:

    +
    $ rm $\backslash$*$\backslash$|more
    +

    or:

    +
    $ rm ''*|more''
    +
    +
    +

    +9. Looking At Files

    +
    +

    +9.1. cat

    +

    The cat command is a standard Unix program used to concatenate and display files. The name is from "catenate", a synonym of concatenate.

    +

    The Single Unix Specification specifies the behavior that the contents of each of the files given in sequence as arguments will be written to the standard output in the same sequence, and mandates one option, -u, where each byte is printed as it is read.

    +

    If the filename is specified as -, then cat will read from standard input at that point in the sequence. If no files are specified, cat will read from standard input entered.

    +
    +

    +9.1.1. Jargon File Definition

    +

    The Jargon File version 4.4.7 lists this as the definition of cat:

    +
    1. To spew an entire file to the screen or some other output sink without
    +     pause (syn. blast).
    +
    +2. By extension, to dump large amounts of data at an unprepared target or
    +     with no intention of browsing it carefully. Usage: considered silly.
    +     Rare outside Unix sites. See also dd, BLT.
    +
    +     Among Unix fans, *cat(1)* is considered an excellent example of
    +     user-interface design, because it delivers the file contents without
    +     such verbosity as spacing or headers between the files, and because
    +     it does not require the files to consist of lines of text, but works
    +     with any sort of data.
    +
    +     Among Unix critics, *cat(1)* is considered the canonical example of
    +     bad user-interface design, because of its woefully unobvious name.
    +     It is far more often used to blast a single file to standard output
    +     than to concatenate two or more files. The name cat for the former
    +     operation is just as unintuitive as, say, LISP's cdr.
    +
    +     Of such oppositions are holy wars made...
    +
    +
    +

    +9.1.2. Useless Use of 'cat'

    +

    UUOC (from comp.unix.shell on Usenet) stands for “Useless Use of cat”. As it is observed on comp.unix.shell, “The purpose of cat is to concatenate (or 'catenate') files. If it's only one file, concatenating it with nothing at all is a waste of time, and costs you a process.”

    +

    Nevertheless one sees people doing:

    +
    $ cat file | some_command and its args ...
    +

    instead of the equivalent and cheaper:

    +
    <file some_command and its args ...
    +

    or (equivalently and more classically):

    +
    some_command and its args ... <file
    +

    Since 1995, occasional awards for UUOC have been given out. The activity of fixing instances of UUOC is sometimes called 'demoggification'.

    +

    Amongst many, it is still considered safer to use cat for such cases given that the < and > keys are next to each other in many popular keyboard mappings. While the risk might be low, the impact of using > instead of < can be high and prohibitive.

    +
    +
    +

    +9.1.3. zcat

    +

    zcat is a Unix program similar to cat, that decompresses individual files and concatenates them to standard output. Traditionally zcat operated on files compressed by compress but today it is usually able to operate on gzip or even bzip2 archives. On such systems, it is equivalent to gunzip -c

    +
    +
    +
    +

    +9.2. more

    +

    In computing, more is a command to view (but not modify) the contents of a text file one screen at a time (terminal pager). It is available on Unix and Unix-like systems, DOS, OS/2 and Microsoft Windows. Programs of this sort are called pagers.

    +
    +

    +9.2.1. Usage

    +

    The command-syntax is:

    +
    $ more [options] [file_name]
    +

    If no file name is provided, more looks for input from stdin.

    +

    Once more has obtained input, it displays as much as can fit on the current screen and waits for user input to advance, with the exception that a form feed (^L) will also cause more to wait at that line, regardless of the amount of text on the screen. In the lower-left corner of the screen is displayed the text "--More--" and a percentage, representing the percent of the file that more has paged through. (This percentage includes the text displayed on the current screen.) When more reaches the end of a file (100%) it exits. The most common methods of navigating through a file are Enter, which advances the output by one line, and Space, which advances the output by one screen.

    +

    There are also other commands that can be used while navigating through the document; consult more's man page for more details.

    +

    Options are typically entered before the file name, but can also be entered in the environment variable $MORE. Options entered in the actual command line will override those entered in the $MORE environment variable. Available options may vary between Unix systems.

    +
    +
    +
    +

    +9.3. less

    +

    less is a terminal pager program on Unix, Windows and Unix-like systems used to view (but not change) the contents of a text file one screen at a time. It is similar to more, but has the extended capability of allowing both forward and backward navigation through the file. Unlike most Unix text editors/viewers, less does not need to read the entire file before starting, resulting in faster load times with large files.

    +
    +

    +9.3.1. Usage

    +

    less can be invoked with options to change its behaviour, for example, the number of lines to display on the screen. A few options vary depending on the operating system. While less is displaying the file, various commands can be used to navigate through the file. These commands are based on those used by both more and vi. It is also possible to search for character patterns in the file.

    +

    By default, less displays the contents of the file to the standard output (one screen at a time). If the file name argument is omitted, it displays the contents from standard input (usually the output of another command through a pipe). If the output is redirected to anything other than a terminal, for example a pipe to another command, less behaves like cat.

    +

    The command-syntax is:

    +
    $ less [options] file_name
    +
    +
    +

    +9.3.2. Frequently Used Options

    +<block_quote>
      +
    • -g: Highlights just the current match of any searched string.

    • +
    • -I: Case-insensitive searches.

    • +
    • -M: Shows more detailed prompt, including file position.

    • +
    • -N: Shows line numbers (useful for source code viewing).

    • +
    • -S: Disables line wrap ("chop long lines"). Long lines can be seen by side scrolling.

    • +
    • -?: Shows help.

    • +
    </block_quote>
    +
    +
    +

    +9.3.3. Frequently Used Commands

    +<block_quote>
      +
    • [Arrows]/[Page Up]/[Page Down]/[Home]/[End]: Navigation.

    • +
    • [Space bar]: Next page.

    • +
    • b: Previous page.

    • +
    • ng: Jump to line number n. Default is the start of the file.

    • +
    • nG: Jump to line number n. Default is the end of the file.

    • +
    • /pattern: Search for pattern. Regular expressions can be used.

    • +
    • '^ or g: Go to start of file.

    • +
    • '$ or G: Go to end of file.

    • +
    • s: Save current content (got from another program like grep) in a file.

    • +
    • =: File information.

    • +
    • h: Help.

    • +
    • q: Quit.

    • +
    </block_quote>
    +
    +
    +

    +9.3.4. Examples

    +
    $ less -M readme.txt                     #Read "readme.txt."
    +$ less +F /var/log/mail.log              #Follow mode for log
    +$ file * | less                          #Easier file analysis.
    +$ grep -i void *.c | less -I -p void     #Case insensitive search                                                         for "void" in all .c files
    +
    +
    +
    +
    +

    +10. Directory Structure

    +

    In the File Hierarchy Standard (FHS) all files and directories appear under the root directory "/", even if they are stored on different physical devices. Note however that some of these directories may or may not be present on a Unix system depending on whether certain subsystems, such as the X Window System, are installed.

    +

    The majority of these directories exist in all UNIX operating systems and are generally used in much the same way; however, the descriptions here are those used specifically for the FHS, and are not considered authoritative for platforms other than Linux.

    ++++
    +
    +

    +10.1. man hier

    +

    This is the manual page on the UNIX filesystem. The syntax for this is:

    +
    $ man hier
    +
    +
    +

    +10.2. ls -l

    +

    Shows you huge amounts of information (permissions, owners, size, and when last modified) for folders and files. The syntax is

    +
    $ ls -l
    +

    This can be done after entering the required directory.

    +
    +
    +
    +

    +11. Permissions and Ownership

    +
    +

    +11.1. chmod

    +

    The chmod command (abbreviated from 'change mode') is a shell command and C language function in Unix and Unix-like environments. When executed, it can change file system modes of files and directories. The modes include permissions and special modes.A chmod command first appeared in AT&T Unix version 1, and is still used today on Unix-like machines.

    +
    +

    +11.1.1. Usage

    +

    The chmod command options are specified like this:

    +
    $ chmod [options] mode[,mode] file1 [file2 ...]
    +

    To view what the permissions currently are, type:

    +
    $ ls -l file
    +
    +
    +

    +11.1.2. Command line options

    +

    The chmod command has a number of command line options that affect its behavior. The most common options are:

    +<block_quote>
      +
    • -R: Changes the modes of directories and files recursively

    • +
    • -v: Verbose mode; lists all files as they are being processed

    • +
    </block_quote>
    +
    +11.1.2.1. Symbolic modes
    +

    To the chmod utility, all permissions and special modes are represented by its mode parameter. One way to adjust the mode of files or directories is to specify a symbolic mode. The symbolic mode is composed of three components, which are combined to form a single string of text:

    +
    $ chmod [references][operator][modes] file1 ...
    +

    The references (or classes) are used to distinguish the users to whom the permissions apply. If no references are specified it defaults to “all” but modifies only the permissions allowed by the umask. The references are represented by one or more of the following letters:

    +++++
    +

    The chmod program uses an operator to specify how the modes of a file should be adjusted. The following operators are accepted:

    ++++
    +

    The modes indicate which permissions are to be granted or taken away from the specified classes. There are three basic modes which correspond to the basic permissions:

    +++++
    +

    The combination of these three components produces a string that is understood by the chmod command. Multiple changes can be specified by separating multiple symbolic modes with commas.

    +
    +
    +
    +11.1.2.2. Symbolic examples
    +

    Add the 'read' and 'write' permissions to the 'user' and 'group' classes of a directory:

    +
    $ chmod ug+rw mydir
    +$ ls -ld mydir
    +drw-rw----   2 starwars  yoda  96 Dec 8 12:53 mydir
    +

    For a file, remove write permissions for all classes:

    +
    $ chmod a-w myfile
    +$ ls -l myfile
    +-r-xr-xr-x   2 starwars  yoda 96 Dec 8 12:53 myfile
    +

    Set the permissions for the u*ser and the *g*roup to read and execute only (no write permission) on *mydir.

    +
    $ chmod ug=rx mydir
    +$ ls -ld mydir
    +dr-xr-x---   2 starwars  yoda 96 Dec 8 12:53 mydir
    +
    +
    +
    +11.1.2.3. Octal numbers
    +

    The chmod command also accepts three and four-digit octal numbers representing modes. Using a three-digit octal number to set the modes of a file named myfile :

    +
    $ chmod 664 myfile
    +$ ls -l myfile
    +-rw-rw-r--  1   57 Jul  3 10:13  myfile
    +

    Since the setuid, setgid and sticky bits are not set, this is equivalent to:

    +
    $ chmod 0664 myfile
    +
    +
    +
    +11.1.2.4. Special modes
    +

    The chmod command is also capable of changing the additional permissions or special modes of a file or directory. The symbolic modes use s to represent the setuid and setgid modes, and t to represent the sticky mode. The modes are only applied to the appropriate classes, regardless of whether or not other classes are specified.

    +

    Most operating systems support the specification of special modes using octal modes, but some do not. On these systems, only the symbolic modes can be used.

    +
    +
    +
    +
    +
    +

    +12. Redirection and Piping

    +

    In computing, redirection is a function common to most command-line interpreters, including the various Unix shells that can redirect standard streams to user-specified locations.

    +

    Programs do redirection with the dup2(2) system call, or its less-flexible but higher-level stdio analogues, freopen(3) and popen(3).

    +
    +

    +12.1. Redirecting standard input and standard output

    +

    Redirection is usually implemented by placing certain characters between commands. Typically, the syntax of these characters is as follows:

    +
    $ command1 > file1
    +

    executes command1, placing the output in file1. Note that this will truncate any existing data in file1. To append output to the end of the file, use the >> operator.:

    +
    $ command1 < file1
    +

    executes command1, using file1 as the source of input (as opposed to the keyboard).:

    +
    $ command1 < infile > outfile
    +

    combines the two capabilities: command1 reads from infile and writes to outfile

    +
    +
    +

    +12.2. Piping

    +

    Programs can be run together such that one program reads the output from another with no need for an explicit intermediate file: +A pipeline of three programs run on a text terminal:

    +
    $ command1 | command2
    +

    executes command1, using its output as the input for command2 (commonly called piping, since the "|" character is known as a "pipe").

    +

    This is equivalent to using two redirects and a temporary file:

    +
    $ command1 > tempfile
    +$ command2 < tempfile
    +$ rm tempfile
    +

    A good example for command piping is combining echo with another command to achieve something interactive in a non-interactive shell, e.g.:

    +
    $ echo -e "user\npass" | ftp localhost
    +

    This runs the ftp client with input user, press return, then pass.

    +
    +
    +

    +12.3. Redirecting to and from the standard file handles

    +

    In Unix shells derived from the original Bourne shell, the first two actions can be further modified by placing a number (the file descriptor) immediately before the character; this will affect which stream is used for the redirection. The Unix standard I/O streams are:

    +++++
    +

    For example:

    +
    $ command1 2> file1
    +

    executes command1, directing the standard error stream to file1.

    +

    In shells derived from csh (the C shell), the syntax instead appends the & character to the redirect characters, thus achieving a similar result.

    +

    Another useful capability is to redirect one standard file handle to another. The most popular variation is to merge standard error into standard output so error messages can be processed together with (or alternately to) the usual output. Example:

    +
    $ find / -name .profile > results 2>&1
    +

    will try to find all files named .profile. Executed without redirection, it will output hits to stdout and errors (e.g. for lack of privilege to traverse protected directories) to stderr. If standard output is directed to file results, error messages appear on the console. To see both hits and error messages in file results, merge stderr (handle 2) into stdout (handle 1) using 2>&1 .

    +

    It's possible use 2>&1 before ">" but it doesn't work. In fact, when the interpreter reads 2>&1, it doesn't know yet where standard output is redirected and then standard error isn't merged.

    +

    If the merged output is to be piped into another program, the file merge sequence 2>&1 must precede the pipe symbol, thus:

    +
    $ find / -name .profile 2>&1 | less
    +

    A simplified form of the command:

    +
    $ command > file 2>&1
    +

    is:

    +
    $ command &>file
    +

    or:

    +
    $command >&file
    +
    +
    +

    +12.4. Chained pipelines

    +

    The redirection and piping tokens can be chained together to create complex commands. For example:

    +
    $ ls | grep '\.sh' | sort > shlist
    +

    lists the contents of the current directory, where this output is filtered to only contain lines which contain .sh, sort this resultant output lexicographically, and place the final output in shlist. This type of construction is used very commonly in shell scripts and batch files.

    +
    +
    +

    +12.5. Redirect to multiple outputs

    +

    The standard command tee can redirect output from a command to several destinations.

    +
    $ ls -lrt | tee xyz
    +

    This directs the file list output to both standard output as well as to the file xyz.

    +
    +
    +
    +

    +13. More Text Processing

    +
    +

    +13.1. grep

    +

    grep is a command line text search utility originally written for Unix. The name is taken from the first letters in global / regular expression / print, a series of instructions for the ed text editor. The grep command searches files or standard input globally for lines matching a given regular expression, and prints them to the program's standard output.

    +
    +

    +13.1.1. Usage

    +

    This is an example of a common grep usage:

    +
    $ grep apple fruitlist.txt
    +

    In this case, grep prints all lines containing 'apple' from the file fruitlist.txt, regardless of word boundaries; therefore lines containing 'pineapple' or 'apples' are also printed. The grep command is case sensitive by default, so this example's output does not include lines containing 'Apple' (with a capital A) unless they also contain 'apple'.

    +

    Like most Unix commands, grep accepts command line arguments to change this and many other behaviors. For example:

    +
    $ grep -i apple fruitlist.txt
    +

    This prints all lines containing 'apple' regardless of capitalization. The '-i' argument tells grep to be case insensitive, or to ignore case.

    +

    To print all lines containing 'apple' as a word ('pineapple' and 'apples' will not match):

    +
    $ grep -w apple fruitlist.txt
    +

    Regular expressions can be used to match more complicated queries.

    +
    +
    +13.1.1.1. Variations
    +

    There are countless implementations and derivatives of grep available for many operating systems. Early variants of grep included egrep and fgrep. The former applies an extended regular expression syntax that was added to Unix after Ken Thompson's original regular expression implementation. The latter searches for any of a list of 'fixed' strings using the Aho-Corasick algorithm. These variants are embodied in most modern grep implementations as command-line switches (and standardized as -E and -F in POSIX). In such combined implementations, grep may also behave differently depending on the name by which it is invoked, allowing fgrep, egrep, and grep to be links to the same program.

    +

    pcregrep is an implementation of grep that uses Perl regular expression syntax.

    +

    Other commands contain the word 'grep' to indicate that they search (usually for regular expression matches). The pgrep utility, for instance, displays the processes whose names match a given regular expression.

    +
    +
    +
    +
    +

    +13.2. tr

    +

    tr (abbreviated from translate or transliterate) is a command in Unix-like operating systems.

    +

    When executed, the program reads from the standard input and writes to the standard output. It takes as parameters two sets of characters, and replaces occurrences of the characters in the first set with the corresponding elements from the other set. For example,

    +
    $ tr 'abcd' 'jkmn'
    +

    maps 'a' to 'j', 'b' to 'k', 'c' to 'm', and 'd' to 'n'.

    +

    Sets of characters may be abbreviated by using character ranges. The previous example could be written:

    +
    $ tr 'a-d' 'jkmn'
    +

    In POSIX compliant versions of tr the set represented by a character range depends on the locale's collating order, so it is safer to avoid character ranges in scripts that might be executed in a locale different from that in which they were written. Ranges can often be replaced with POSIX character sets such as [:alpha:].

    +

    The -c flag complements the first set of characters.

    +
    $ tr -cd '[:alnum:]'
    +

    therefore removes all non-alphanumeric characters.

    +

    The -s flag causes tr to compress sequences of identical adjacent characters in its output to a single token. For example,

    +
    $ tr -s '\n' '\n'
    +

    replaces sequences of one or more newline characters with a single newline.

    +

    The -d flag causes tr to delete all tokens of the specified set of characters from its input. In this case, only a single character set argument is used. The following command removes carriage return characters, thereby converting a file in DOS/Windows format to one in Unix format.

    +
    $ tr -d '\r'
    +

    Most versions of tr, including GNU tr and classic Unix tr, operate on single byte characters and are not Unicode compliant. An exception is the Heirloom Toolchest implementation, which provides basic Unicode support.

    +

    Ruby and Perl also have an internal tr operator, which operates analogously. Tcl's string map command is more general in that it maps strings to strings while tr maps characters to characters.

    +
    +
    +
    +

    +14. Elementary Regex

    +

    In computing, regular expressions provide a concise and flexible means for identifying strings of text of interest, such as particular characters, words, or patterns of characters. A regular expression (often shortened to regex or regexp) is written in a formal language that can be interpreted by a regular expression processor, a program that either serves as a parser generator or examines text and identifies parts that match the provided specification.

    +

    Regular expressions are used by many text editors, utilities, and programming languages to search and manipulate text based on patterns. For example, Perl, Ruby and Tcl have a powerful regular expression engine built directly into their syntax. Several utilities provided by Unix distributions—including the editor ed and the filter grep — were the first to popularize the concept of regular expressions.

    +

    Traditional Unix regular expression syntax followed common conventions but often differed from tool to tool. The IEEE POSIX Basic Regular Expressions (BRE) standard (released alongside an alternative flavor called Extended Regular Expressions or ERE) was designed mostly for backward compatibility with the traditional (Simple Regular Expression) syntax but provided a common standard which has since been adopted as the default syntax of many Unix regular expression tools, though there is often some variation or additional features. Many such tools also provide support for ERE syntax with command line arguments.

    +

    In the BRE syntax, most characters are treated as literals — they match only themselves (i.e., a matches "a"). The exceptions, listed below, are called metacharacters or metasequences.

    ++++
    +
    +

    +14.1. Lazy quantification

    +

    The standard quantifiers in regular expressions are greedy, meaning they match as much as they can, only giving back as necessary to match the remainder of the regex. For example, someone new to regexes wishing to find the first instance of an item between < and > symbols in this example:

    +
    Another whale explosion occurred on <January 26>, <2004>.
    +

    ...would likely come up with the pattern <.*>, or similar. However, this pattern will actually return "<January 26>, <2004>" instead of the "<January 26>" which might be expected, because the <title_reference>*</title_reference> quantifier is greedy — it will consume as many characters as possible from the input, and "January 26>, <2004" has more characters than "January 26".

    +

    Though this problem can be avoided in a number of ways (e.g., by specifying the text that is not to be matched: <[^>]*>), modern regular expression tools allow a quantifier to be specified as lazy (also known as non-greedy, reluctant, minimal, or ungreedy) by putting a question mark after the quantifier (e.g., <.*?>), or by using a modifier which reverses the greediness of quantifiers (though changing the meaning of the standard quantifiers can be confusing). By using a lazy quantifier, the expression tries the minimal match first. Though in the previous example lazy matching is used to select one of many matching results, in some cases it can also be used to improve performance when greedy matching would require more backtracking.

    +
    +
    +
    +

    +15. One Liners

    +

    A one-liner is textual input to the command-line of an operating system shell that performs some function in just one line of input.

    +

    The one liner can be

    +<block_quote>
      +
    1. An expression written in the language of the shell.

    2. +
    3. The invocation of an interpreter together with program source for the interpreter to run.

    4. +
    5. The invocation of a compiler together with source to compile and +instructions for executing the compiled program.

    6. +
    </block_quote>

    Certain dynamic scripting languages such as AWK, sed, and perl have traditionally been adept at expressing one-liners. Specialist shell interpreters such as these Unix shells or the Windows PowerShell, allow for the construction of powerful one-liners.

    +

    The use of the phrase one-liner has been widened to also include program-source for any language that does something useful in one line.

    +

    The word One-liner has two references in the index of the book The AWK Programming Language (the book is often referred to by the abbreviation TAPL). It explains the programming language AWK, which is part of the Unix operating system. The authors explain the birth of the One-liner paradigm with their daily work on early Unix machines:

    +
    “The 1977 version had only a few built-in variables and predefined functions. It was designed for writing short programs [...] Our model was that an invocation would be one or two lines long, typed in and used immediately. Defaults were chosen to match this style [...] We, being the authors, knew how the language was supposed to be used, and so we only wrote one-liners.”
    +

    Notice that this original definition of a One-liner implies immediate execution of the program without any compilation. So, in a strict sense, only source code for interpreted languages qualifies as a One-liner. But this strict understanding of a One-liner was broadened in 1985 when the IOCCC introduced the category of Best One Liner for C, which is a compiled language.

    +

    The TAPL book contains 20 examples of One-liners (A Handful of Useful awk One-Liners) at the end of the book's first chapter.

    +

    Here are the first few of them:

    +<block_quote>
      +
    1. +

      Print the total number of input lines:

      +

      END { print NR }

      +
    2. +
    3. +

      Print the tenth input line:

      +

      NR == 10

      +
    4. +
    5. +

      Print the last field of every input line:

      +

      { print $NF }

      +
    6. +
    </block_quote>

    One-liners are also used to show off the differential expressive power of programming languages. Frequently, one-liners are used to demonstrate programming ability. Contests are often held to see who can create the most exceptional one-liner.

    +

    The following example is a C program (a winning entry in the "Best one-liner" category of the IOCCC, here split to two lines for presentation).:

    +
    main(int c,char**v){return!m(v[1],v[2]);}m(char*s,char*t){return
    +*t-42?*s?63==*t|*s==*t&&m(s+1,t+1):!*t:m(s,t+1)||*s&&m(s+1,t);}
    +

    This one-liner program is a glob pattern matcher. It understands the glob characters '*' meaning 'zero or more characters' and '?' meaning exactly one character, just like most Unix shells.

    +

    Run it with two args, the string and the glob pattern. The exit status is 0 (shell true) when the pattern matches, 1 otherwise. The glob pattern must match the whole string, so you may want to use * at the beginning and end of the pattern if you are looking for something in the middle. Examples:

    +
    $ prog foo 'f??'; echo $?
    +
    +$ prog 'best short program' '??st*o**p?*'; echo $?
    +

    Here is a one line shell script to show directories:

    +
    $ ls -R | grep ":$" | sed -e 's/:$//' -e 's/[^-][^\/]*\//--/g' -e 's/^/   /' -e 's/-/|/'
    +
    +
    +
    + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/ch2intro.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/ch2intro.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1138 @@ + + + +Intro + + + + + + +
    +
    + +
    +
    +

    +
    +
    + +
    +

    +1. Introducing Linux

    +

    (Attribution : A significant chunk of the content under this section is based on data from Wikipedia and the Linux Documentation Project)

    +

    Linux (usually pronounced ˈlɪnəks') is a generic term referring to Unix-like computer operating systems based on the Linux kernel, where a kernel is the intermediate layer between the hardware and the applications. The kernel is, on an abstract level, the core of (most) operating systems, that manages the various system resources. The development of the Linux OS is considered the basis for Free and Open Source Software (FOSS) collaboration since typically the underlying source code can be used, modified freely, and redistributed by anyone under the terms of the GNU (a recursive acronym for "GNU's Not Unix!") Global Public License (GPL) and other free software licences. This freedom to access and reuse various components of a system, is one of the primary reasons for the popularity of Linux.

    +

    Linux is installed on a variety of computer hardware, that include mobile phones, embedded devices and supercomputers, but is infamous for its use in servers.

    +

    The name "Linux" comes from the Linux kernel, originally written in 1991 by Linus Torvalds. The rest of the system usually comprises components such as the Apache HTTP Server, the X Window System, the GNOME and KDE desktop environments, and utilities and libraries from the GNU Project (announced in 1983 by Richard Stallman). Commonly-used applications with desktop Linux systems include the Mozilla Firefox web-browser and the OpenOffice.org office application suite. The GNU contribution is the basis for the Free Software Foundation's preferred name GNU/Linux. The kernel's mascot is a penguin named "Tux". Mozilla Firefox and OpenOffice.org are open-source projects which can be run on most Operating Systems, including proprietary ones.

    +
    +

    +1.1. Historical Background

    +
    +

    +1.1.1. Events leading to the creation

    +
      +
    • The Unix operating system was developed in the 1960s and released for public use in 1970. Its accessibility and portability caused it to be widely adopted, copied and modified by academic institutions and businesses. Its design became influential to authors of other systems. Other free operating systems include the Berkeley Software Distribution (BSD), developed at the University of California at Berkeley, and MINIX which was released by Andrew S. Tanenbaum. The development and adoption of BSD and MINIX were limited due to various reasons, and this lack of a widely-adopted and free kernel triggered Linus Torvalds into starting his project.

    • +
    • In 1983, Richard Stallman started the GNU project with the goal of creating a free UNIX-like operating system. As part of this work, he wrote the GNU General Public License (GPL). By the early 1990s there was almost enough available software to create a full operating system. However, the GNU kernel, called Hurd, failed to attract enough attention from developers leaving GNU incomplete.

    • +
    +
    +
    +

    +1.1.2. The Creation of Linux

    +

    In 1991, Linus Torvalds began a project at the University of Helsinki that later became the Linux kernel. It was initially a terminal (command-line) emulator, which Torvalds used to access the large UNIX servers of the university. He wrote the program targeting just the hardware he was using and independent of an operating system because he wanted to use the functions of his computer with an 80386 processor. Development was done on Minix using the GNU C compiler. This application is still the main choice for compiling Linux today (although the code can be built with other compilers, such as the Intel C Compiler).

    +

    Torvalds continues to direct the development of the kernel. Stallman heads the Free Software Foundation, which in turn supports the GNU components. Finally, individuals and corporations develop third-party non-GNU components, which constitute a vast body of work and including kernel modules, and user applications and libraries. Linux vendors and communities combine and distribute the kernel, GNU components, and non-GNU components, with additional package management software in the form of Linux distributions.

    +
    +
    +
    +

    +1.2. Design and Implications

    +

    A Linux-based system is a modular Unix-like operating system, deriving much of its basic design from principles established in Unix earlier. Such a system uses a monolithic kernel, called the Linux kernel, which handles process control, networking, and peripheral and file system access. Device drivers are integrated directly with the kernel. Separate projects that interface with the kernel provide much of the system's higher-level functionality. The GNU userland is an important part of most Linux-based systems, providing the most common implementation of the C library, a popular shell, and many of the common Unix tools which carry out many basic operating system tasks. The graphical user interface (or GUI) used by most Linux systems is based on the "X Window System".

    +
    +

    +1.2.1. User Interface

    +

    Users can control a Linux-based system through a command line interface (or CLI), a graphical user interface (or GUI), or through controls attached to the associated hardware (this is common for embedded systems). For desktop systems, the default mode is usually the GUI. On desktop machines, "KDE", "GNOME" and "Xfce" are the most popular user interfaces,though a variety of additional user interfaces exist. Most popular user interfaces run on top of the "X Window System" (or X), which enables a graphical application running on one machine to be displayed and controlled from another in a network.

    +

    A Linux system also provides a CLI of some sort through a shell, which is the traditional way of interacting with a Unix system. A Linux distribution specialized for servers may use the CLI as its only interface. A “headless system” (system run without even a monitor) can be controlled by the command line via a remote-control protocol such as SSH or telnet. The CLI is particularly suited for automation of repetitive or delayed tasks, and provides very simple inter-process communication. A graphical terminal emulator program is often used to access the CLI from a Linux desktop.

    +
    +
    +

    +1.2.2. Development

    +

    The primary difference between Linux and many other popular contemporary operating systems is that the Linux kernel and other components are free and open source software. Linux is not the only such operating system, although it is by far the most widely used. Some free and open source software licenses are based on the principle of "copyleft", a kind of reciprocity: any work derived from a copyleft piece of software must also be copyleft itself. The most common free software license, the GNU GPL, is a form of copyleft, and is used for the Linux kernel and many of the components from the GNU project.

    +

    Linux based distributions are intended by developers for interoperability with other operating systems and established computing standards. Linux systems adhere to POSIX, SUS, ISO and ANSI standards where possible, although to date only one Linux distribution has been POSIX.1 certified, Linux-FT.Free software projects, although developed in a collaborative fashion, are often produced independently of each other. The fact that the software licenses explicitly permit redistribution, however, provides a basis for larger scale projects that collect the software produced by stand-alone projects and make it available all at once in the form of a Linux distribution.

    +

    A Linux distribution, commonly called a "distro", is a project that manages a remote collection of system software and application software packages available for download and installation through a network connection. This allows the user to adapt the operating system to his/her specific needs. Distributions are maintained by individuals, loose-knit teams, volunteer organizations, and commercial entities. A distribution can be installed using a CD that contains distribution-specific software for initial system installation and configuration. A package manager such as Synaptic or YAST allows later package upgrades and installations. A distribution is responsible for the default configuration of the installed Linux kernel, general system security, and more generally integration of the different software packages into a coherent whole.

    +
    +
    +

    +1.2.3. Community

    +

    A distribution is largely driven by its developer and user communities. Some vendors develop and fund their distributions on a volunteer basis. Examples include Debian and the Debian-based, Ubuntu. Others maintain a community version of their commercial distributions, as Red Hat does with Fedora.

    +

    In many cities and regions, local associations known as Linux Users Groups (LUGs) seek to promote their preferred distribution and by extension free software. They hold meetings and provide free demonstrations, training, technical support, and operating system installation to new users. Many Internet communities also provide support to Linux users and developers. Most distributions and free software / open source projects have IRC (Internet Relay Chat) chatrooms or newsgroups. Online forums are another means for support. Linux distributions host mailing lists; commonly there will be a specific topic such as usage or development for a given list. All these can be found simply by running an appropriate search on Google.

    +

    Although Linux distributions are generally available without charge, several large corporations sell, support, and contribute to the development of the components of the system and of free software. These include Dell, IBM, HP, Oracle, Sun Microsystems, Novell, Nokia. A number of corporations, notably Red Hat, have built their entire business around Linux distributions.

    +
    +
    +

    +1.2.4. Can I make a profit out of running a business involving Linux?

    +

    The answer is, "Yes!". The free software licenses, on which the various software packages of a distribution built on the Linux kernel are based, explicitly accommodate and encourage commercialization; the relationship between a Linux distribution as a whole and individual vendors may be seen as symbiotic. One common business model of commercial suppliers is charging for support, especially for business users. A number of companies also offer a specialized business version of their distribution, which adds proprietary support packages and tools to administer higher numbers of installations or to simplify administrative tasks. Another business model is to give away the software in order to sell hardware. Examples of corporations that are extensively (and sometimes exclusively) open-source and Linux-powered , with successful revenue generation models involving these, are Google, SUN, Mozilla, etc.

    +
    +
    +

    +1.2.5. Programming on Linux

    +

    Most Linux distributions support dozens of programming languages. The most common collection of utilities for building both Linux applications and operating system programs is found within the GNU toolchain, which includes the GNU Compiler Collection (GCC) and the GNU build system. Amongst others, GCC provides compilers for Ada, C, C++, Java, and Fortran. The Linux kernel itself is written to be compiled with GCC. Proprietary compilers for Linux include the Intel C++ Compiler, Sun Studio, and IBM XL C/C++ Compiler.

    +

    Most distributions also include support for PHP, Perl, Ruby, Python and other dynamic languages. Examples of languages that are less common, but still supported, are C# via the Mono project, sponsored by Novell, and Scheme. A number of Java Virtual Machines and development kits run on Linux, including the original Sun Microsystems JVM (HotSpot), and IBM's J2SE RE, as well as many open-source projects like Kaffe.

    +

    The two main frameworks for developing graphical applications are those of GNOME and KDE. These projects are based on the GTK+ and Qt widget toolkits, respectively, which can also be used independently of the larger framework. Both support a wide variety of languages. There are a number of Integrated Development Environments (IDEs) available including Anjuta, Code::Blocks, Eclipse, KDevelop, Lazarus, MonoDevelop, NetBeans, and Omnis Studio while the long-established editors Vim and Emacs remain popular.

    +
    +
    +
    +

    +1.3. Reasons for Using Linux

    +
    • Linux is free:

    +

    As in "free beer". Linux can be downloaded in its entirety from the Internet completely for free. No registration fees, no costs per user, free updates, and freely available source code in case you want to change the behavior of your system. +Most of all, Linux is free as in "free speech": +The license commonly used is the GNU Public License (GPL). The license says that anybody who may want to do so, has the right to change Linux and eventually to redistribute a changed version, on the one condition that the code is still available after redistribution. In practice, you are free to grab a kernel image and sell the new code, as long as your customers can still have a copy of that code.

    +
    • Linux is portable to any hardware platform:

    +

    A vendor, who wants to sell a new type of computer and who does not know what kind of OS his/her new machine will run, can take a Linux kernel and make it work on his/her hardware, because documentation related to this activity is freely available.

    +
    • Linux was made to keep on running:

    +

    As with UNIX, a Linux system expects to run without rebooting all the time. That is why a lot of tasks are being executed at night or scheduled automatically for other times, resulting in higher availability during busier periods and a more balanced use of the hardware. This property allows for Linux to be applicable to environments where people do not have the time or the possibility to control their systems constantly.

    +
    • Linux is secure and versatile:

    +

    The security model used in Linux is based on the UNIX idea of security, which is known to be robust and of proven quality. But Linux is not only safe from attacks from the Internet: it will adapt equally to other situations, utilizing the same high standards for security.

    +
    • Linux is scalable:

    +

    From a Palmtop with 2 MB of memory to a petabyte storage cluster with hundreds of nodes: add or remove the appropriate packages and Linux fits all. One does not need a supercomputer anymore,because you can use Linux to do big things using the building blocks provided with the system. If one wants to do little things, such as making an operating system for an embedded processor or just recycling your old 486, Linux will do that as well.

    +
    • The Linux OS and Linux applications have very short debug−times:

    +

    Because Linux has been developed and tested by thousands of people, both errors and people to fix them are found very quickly. It often happens that there are only a couple of hours between discovery and fixing of a bug.

    +
    +
    +
    +

    +2. Getting Started

    +
    +

    +2.1. Logging in, activating the user interface and logging out

    +

    In order to work on a Linux system directly, one needs to provide a user name and password. You always need to authenticate to the system. Most PC−based Linux systems have two basic modes for a system to run in: either quick and clean in text console mode,which includes with mouse, multitasking and multi−user features, or in graphical console mode, which looks better but eats more system resources.

    +
    +

    +2.1.1. Graphical Mode

    +

    This is the default nowadays on most desktop computers. You know you will be connecting to the system using graphical mode when you are first asked for your user name, and then to type your password.

    +

    To log in, make sure the mouse pointer is in the login window, provide your user name and password to the system and click OK or press Enter. +It is generally considered a bad idea to connect (graphically) using the root user name, the system adminstrator's account, since the use of graphics includes running a lot of extra programs, in root's case with a lot of extra permissions. To keep all risks as low as possible, use a normal user account to connect graphically. But there are enough risks to keep this in mind as a general advice, for all use of the root account: only log in as root when extra privileges are required.

    +

    After entering your user name/password combination, it can take a little while before the graphical environment is started, depending on the CPU speed of your computer, on the software you use and on your personal settings.

    +

    To continue, you will need to open a terminal window or xterm for short (X being the name for the underlying software supporting the graphical environment). This program can be found in the Applications−>Utilities->System Tools or Internet menu, depending on what window manager you are using. There might be icons that you can use as a shortcut to get an xterm window as well, and clicking the right mouse button on the desktop background will usually present you with a menu containing a terminal window application.

    +

    While browsing the menus, you will notice that a lot of things can be done without entering commands via the keyboard. For most users, the good old point−n−click method of dealing with the computer will do. But for those who want to enter the "heart" of the system, a tool stronger than a mouse will be required to handle the various tasks. This tool is the shell, and when in graphical mode, we activate our shell by opening a terminal window.

    +

    A terminal window should always show a command prompt when you open one. This terminal shows a standard prompt, which displays the user's login name, and the current working directory, represented by the twiddle (~)

    +

    Another common form for a prompt is this one: +[

    +
    +

    user@host
    +

    dir]

    +

    In the above example, user will be your login name, hosts the name of the machine you are working on, and dir an indication of your current location in the file system. Prompts can display all kinds of information, but they are not part of the commands you are giving to your system. To disconnect from the system in graphical mode, you need to close all terminal windows and other applications. After that, hit the logout icon or find Log Out in the menu. Closing everything is not really necessary, and the system can do this for you, but session management might put all currently open applications back on your screen when you connect again, which takes longer and is not always the desired effect. However, this behavior is configurable.

    +

    When you see the login screen again, asking to enter user name and password, logout was successful.

    +
    +
    +

    +2.1.2. Text Mode

    +

    One is in text mode when the whole screen is black, showing (in most cases white) characters. A text mode login screen typically shows some information about the machine you are working on, the name of the machine and a prompt waiting for you to log in.

    +

    The login is different from a graphical login, in that you have to hit the Enter key after providing your user name, because there are no buttons on the screen that you can click with the mouse. Then you should type your password, followed by another Enter. You will not see any indication that you are entering something, not even an asterisk, and you won't see the cursor move. But this is normal on Linux and is done for security +reasons.

    +

    When the system has accepted you as a valid user, you may get some more information, called the message of the day, which can be anything. Additionally, it is popular on UNIX systems to display a fortune cookie, which contains some general wise or unwise (this is up to you) thoughts. After that, you will be given a shell, indicated with the same prompt that you would get in graphical mode.

    +

    Also in text mode: log in as root only to do setup and configuration that absolutely requires administrator privileges, such as adding users, installing software packages, and performing network and other system configuration. Once you are finished, immediately leave the special account and resume your work as a non−privileged user.

    +

    Logging out is done by entering the logout command, followed by Enter. You are successfully disconnected from the system when you see the login screen again.Don't power−off the computer after logging out. It is not meant to be shut off without application of the proper procedures for halting the system. Powering it off without going through the halting process might cause severe damage!

    +
    +
    +
    +
    +

    +3. Basic Commands

    +
    +

    +3.1. ls

    +

    When invoked without any arguments, ls lists the files in the current working directory. A directory that is not the current working directory can be specified and ls will list the files there. The user also may specify any list of files and directories. In this case, all files and all contents of specified directories will be listed. The name ls is derived from list segments which was used in earlier systems.

    +

    Files whose names start with "." are not listed, unless the -a flag is specified or the files are specified explicitly.

    +

    Without options, ls displays files in a bare format. This bare format however makes it difficult to establish the type, permissions, and size of the files. The most common options to reveal this information or change the list of files are:

    +<block_quote>
      +
    • -l long format, displaying Unix file types, permissions, number of hard links, owner, group, size, date, and filename

    • +
    • -F appends a character revealing the nature of a file, for example, * for an executable, or / for a directory. Regular files have no suffix.

    • +
    • -a lists all files in the given directory, including those whose names start with "." (which are hidden files in Unix). By default, these files are excluded from the list.

    • +
    • -R recursively lists subdirectories. The command ls -R / would therefore list all files.

    • +
    • -d shows information about a symbolic link or directory, rather than about the link's target or listing the contents of a directory.

    • +
    • -t sort the list of files by modification time.

    • +
    • -h print sizes in human readable format. (e.g., 1K, 234M, 2G, etc.)

    • +
    </block_quote>

    In some environments, providing the option --color (for GNU ls) or -G (FreeBSD ls) causes ls to highlight different types of files with different colors, instead of with characters as -F would. To determine what color to use for a file, GNU ls checks the Unix file type, the file permissions, and the file extension, while FreeBSD ls checks only the Unix file type and file permissions.:

    +
    $ ls
    +jeeves.rst psmith.html blandings.html
    +$ ls -l
    +drwxr--r--   1 plum  editors   4096  jeeves
    +-rw-r--r--   1 plum  editors  30405  psmith
    +-r-xr-xr-x   1 plum  plum      8460  blandings
    +

    Here "$" actually is the beginning of the prompt. This is typical in most Unix-based systems.

    +
    +
    +

    +3.2. date

    +

    The Unix date command displays the time and date. The super-user can use it to set the system clock.

    +

    With no options, the date command displays the current date and time, including the abbreviated day name, abbreviated month name, day of the month, the time separated by colons, the timezone name, and the year. For example:

    +
    $date
    +Tue Sep  8 12:01:45 IST 2009
    +

    On some systems to set the current date and time to September 8, 2004 01:22 you type:

    +
    $date --set="20040908 01:22"
    +

    In order to view the various options for the date command, type:

    +
    $man date
    +

    This will take you to the "Manual" page comprising of all the details on the date command. You can return to the terminal from the "man" page by pressing the Esc key in the keyboard and typing ":q" in that order.

    +
    +
    +

    +3.3. cd

    +

    This stands for "change directory". When one wants to go up to the parent directory, bypassing the tree of directories one has entered, “ cd ..” can be used.

    +

    One dot '.' represents the current directory while two dots '..' represent the parent directory.

    +

    “ cd -” will return you to the previous directory (a bit like an “undo”).

    +

    You can also use cd absolute path or cd relative path (see below):

    +

    Absolute paths:

    +<block_quote>

    An “ absolute path” is easily recognised from the leading forward slash, /. The / means that you start at the top level directory and continue down.

    </block_quote>

    For example to get to /boot/grub you would type:

    +
    $cd /boot/grub
    +

    This is an absolute path because you start at the top of the hierarchy and go downwards from there (it doesn't matter where in the filesystem you were when you typed the command).

    +

    Relative paths:

    +<block_quote>

    A “ relative path” doesn't have a preceding slash. Use a relative path when you start from a directory below the top level directory structure. This is dependent on where you are in the filesystem.

    +

    For example if you are in root's home directory and want to get to /root/music, you type:

    +
    $ cd music
    </block_quote>

    Please note that there is no / using the above cd command. Using a / would cause this to be an absolute path, working from the top of the hierarchy downward.

    +
    +
    +

    +3.4. who

    +

    The standard Unix command who displays a list of users who are currently logged into a computer.

    +

    The who command is related to the command w, which provides the same information but also displays additional data and statistics.:

    +
    $who
    +beeblebrox tty7         2009-09-08 10:50 (:0)
    +beeblebrox pts/0        2009-09-08 11:25 (:0.0)
    +dumbledore pts/1        2009-09-08 18:11 (potter.xyz.in)
    +beeblebrox pts/2        2009-09-08 18:53 (:0.0)
    +

    The command can be invoked with the arguments am i or am I (so it is invoked as who am i or * who am I*), showing information about the current terminal only (see the -m option below, of which this invocation is equivalent).

    +

    In order to find out the various options that can be appended to the who command, check the man page by typing out the following in the terminal:

    +
    $man who
    +

    This will take you to the "Manual" page containing details about the who command

    +
    +
    +

    +3.5. mkdir

    +

    This command is used to make a new directory. Normal usage is as straightforward as follows:

    +
    $mkdir name_of_directory
    +

    Where name_of_directory is the name of the directory one wants to create. When typed as above (ie. normal usage), the new directory would be created within the current directory. On Unix, multiple directories can be specified, and mkdir will try to create all of them.

    +
    +

    +3.5.1. Options

    +

    On Unix-like operating systems, mkdir takes options. Three of the most common options are:

    +<block_quote>
      +
    • -p: will also create all directories leading up to the given directory that do not exist already. If the given directory already exists, ignore the error.

    • +
    • -v: display each directory that mkdir creates. Most often used with -p.

    • +
    • -m: specify the octal permissions of directories created by mkdir.

    • +
    </block_quote>

    -p is most often used when using mkdir to build up complex directory hierarchies, in case a necessary directory is missing or already there. -m is commonly used to lock down temporary directories used by shell scripts.

    +
    +
    +

    +3.5.2. Examples

    +

    An example of -p in action is:

    +
    $mkdir -p /tmp/a/b/c
    +

    If /tmp/a exists but /tmp/a/b does not, mkdir will create /tmp/a/b before creating /tmp/a/b/c.

    +

    And an even more powerful command, creating a full tree at once (this however is a Shell extension, nothing mkdir does itself):

    +
    $mkdir -p tmpdir/{trunk/sources/{includes,docs},branches,tags}
    +

    This will create:

    +<definition_list><definition_list_item><term>tmpdir - branches</term><definition>
      +
    • tag

    • +
    • <definition_list><definition_list_item><term>trunk - sources - includes</term><definition>
      • docs

      </definition>
      </definition_list_item>
      </definition_list>
    • +
    </definition>
    </definition_list_item>
    </definition_list>
    +
    +
    +
    +
    +

    +4. Getting Help

    +
    +

    +4.1. apropos and whatis

    +

    This is a command to search the manual pages files in Unix and Unix-like operating systems.

    +
    $ apropos grep
    +egrep       egrep (1)       Search a file for a pattern using full regular expressions
    +fgrep       fgrep (1)       Search a file for a fixed-character string
    +fmlgrep     fmlgrep (1)     Search a file for a pattern
    +grep        grep (1)        Search a file for a pattern
    +gzgrep      gzgrep (1)      Search a possibly compressed file for a regular expression
    +nisgrep     nismatch (1)    Utilities for searching NIS+ tables
    +pgrep       pgrep (1)       Find or signal a process by name or other attribute
    +zgrep       zgrep (1)       Search a possibly compressed file for a regular expression
    +...
    +

    In this example, the user uses apropos to search for the string "grep", and apropos returns the indicated man pages that include the term "grep".

    +

    A short index of explanations for commands is available using the whatis command, like in the examples below:

    +
    $whatis ls
    +ls (1)           - list directory contents
    +

    This displays short information about a command, and the first section in the collection of man pages that contains an appropriate page.

    +

    If you don't know where to get started and which man page to read, apropos gives more information. Say that you do not know how to start a browser, then you could enter the following command:

    +
    $apropos browser
    +gmusicbrowser (1)    - Jukebox for large collections of audio files
    +infobrowser (1)      - read Info documents
    +libsmbclient (7)     - An extension library for browsers and that               can be used...
    +opera (1)            - a standards-compliant graphical Web browser
    +sensible-browser (1) - sensible editing, paging, and web browsing
    +smbtree (1)          - A text based smb network browser
    +tvtk_doc (1)         - A GUI based TVTK documentation search browser.
    +viewres (1)          - graphical class browser for Xt
    +w3m (1)              - a text based Web browser and pager
    +www-browser (1)      - a text based Web browser and pager
    +...
    +
    +
    +

    +4.2. man

    +

    Man pages (short for "manual pages") are the extensive documentation that comes preinstalled with almost all substantial Unix and Unix-like operating systems. The Unix command used to display them is man. Each page is a self-contained document.

    +

    To read a manual page for a Unix command, one can use:

    +
    $ man <command_name>
    +

    at a shell prompt; for example, "man ftp". In order to simplify navigation through the output, man generally uses the less terminal pager.

    +

    Pages are traditionally referred to using the notation "name(section)"; for example, ftp(1). The same page name may appear in more than one section of the manual, this can occur when the names of system calls, user commands, or macro packages coincide. Two examples are man(1) and man(7), or exit(2) and exit(3). The syntax for accessing the non-default manual section varies between different man implementations. On Linux and BSD, for example, the syntax for reading *printf(3) is:

    +
    $man 3 printf
    +

    Another example:

    +
    $man man
    +

    The previous example will take you to the "Manual" page entry about manual pages!

    +
    +

    +4.2.1. Layout

    +

    All man pages follow a common layout that is optimized for presentation on a simple ASCII text display, possibly without any form of highlighting or font control. Sections present may include:

    +<definition_list><definition_list_item><term>NAME</term><definition>

    The name of the command or function, followed by a one-line description of what it does.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SYNOPSIS</term><definition>

    In the case of a command, you get a formal description of how to run it and what command line options it takes. For program functions, a list of the parameters the function takes and which header file contains its definition. For experienced users, this may be all the documentation they need.

    </definition>
    </definition_list_item>
    <definition_list_item><term>DESCRIPTION</term><definition>

    A textual description of the functioning of the command or function.

    </definition>
    </definition_list_item>
    <definition_list_item><term>EXAMPLES</term><definition>

    Some examples of common usage.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SEE ALSO</term><definition>

    A list of related commands or functions.

    </definition>
    </definition_list_item>
    </definition_list>

    Other sections may be present, but these are not well standardized across man pages. Common examples include: OPTIONS, EXIT STATUS, ENVIRONMENT, KNOWN BUGS, FILES, AUTHOR, REPORTING BUGS, HISTORY and COPYRIGHT.

    +

    These days virtually every Unix command line application comes with its man page, and many Unix users perceive a lack of man pages as a sign of low quality; indeed, some projects, such as Debian, go out of their way to write man pages for programs lacking one. Few alternatives to man have enjoyed much popularity, with the possible exception of the GNU project's "info" system, an early and simple hypertext system.

    +

    However, the format of a single page for each application, the lack of classification within the sections and the relatively unsophisticated formatting facilities have motivated the development of alternative documentation systems, such as the previously mentioned "info" system.

    +

    Most Unix GUI applications (particularly those built using the GNOME and KDE development environments) now provide end-user documentation in HTML and include embedded HTML viewers such as yelp for reading the help within the application.

    +

    Usually the man pages are written in English. Translations into other languages can be also available on the system.

    +

    The default format of the man pages is troff, with either the macro package man (appearance oriented) or on some systems mdoc (semantic oriented). This makes it possible to typeset a man page to PostScript, PDF and various other formats for viewing or printing.

    +
    +
    +
    +

    +4.3. info

    +

    info is a software utility which forms a hypertextual, multipage documentation and help viewer working on a command line interface, useful when there is no GUI available.

    +

    The syntax is

    +
    $ info <command_name>
    +

    info processes info files, which are Texinfo formatted files, and presents the documentation as a tree, with simple commands to traverse the tree and to follow cross references. For instance

    +<block_quote>
      +
    • n goes to the next page.

    • +
    • p goes to the previous page.

    • +
    • u goes to the upper page.

    • +
    • l goes to the last(visited) node

    • +
    • To follow a cross reference, the cursor can be moved over a link (a word preceded by a <title_reference>*</title_reference>) and enter pressed.

    • +
    </block_quote>

    info was initially written for use with GNU/Linux and then ported to other Unix-like operating systems.

    +
    +
    +

    +4.4. --help

    +

    Most GNU commands support the --help, which gives a short explanation about how to use the command and a list of available options. Below is the output of this option with the cat command:

    +
    $ userprompt@host: cat --help
    +Usage: cat [OPTION] [FILE]...
    +Concatenate FILE(s), or standard input, to standard output.
    +
    +  -A, --show-all           equivalent to -vET
    +  -b, --number-nonblank    number nonempty output lines
    +  -e                       equivalent to -vE
    +  -E, --show-ends          display $ at end of each line
    +  -n, --number             number all output lines
    +  -s, --squeeze-blank      suppress repeated empty output lines
    +  -t                       equivalent to -vT
    +  -T, --show-tabs          display TAB characters as ^I
    +  -u                       (ignored)
    +  -v, --show-nonprinting   use ^ and M- notation, except for LFD and              TAB
    +  --help     display this help and exit
    +  --version  output version information and exit
    +
    +With no FILE, or when FILE is -, read standard input.
    +
    +Examples:
    +  cat f - g  Output f's contents, then standard input, then g's           contents.
    +  cat        Copy standard input to standard output.
    +
    +Report bugs to <bug-coreutils@gnu.org>.
    +
    +
    +
    +

    +5. Basic file handling

    +
    +

    +5.1. cp

    +

    cp is the command entered in a Unix shell to copy a file from one place to another, possibly on a different filesystem. The original file remains unchanged, and the new file may have the same or a different name.

    +
    +

    +5.1.1. Usage

    +

    To copy a file to another file:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ][ -- ] SourceFile TargetFile
    +

    To copy a file to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -r | -R ] [ -- ] SourceFile ...              TargetDirectory
    +

    To copy a directory to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -- ] { -r | -R }
    +SourceDirectory ... TargetDirectory
    +
    +
    +

    +5.1.2. Flags

    +

    -f (force) – specifies removal of the target file if it cannot be opened for write operations. The removal precedes any copying performed by the cp command.

    +

    -P – makes the cp command copy symbolic links. The default is to follow symbolic links, that is, to copy files to which symbolic links point.

    +

    -i (interactive) – prompts you with the name of a file to be overwritten. This occurs if the TargetDirectory or TargetFile parameter contains a file with the same name as a file specified in the SourceFile or SourceDirectory parameter. If you enter y or the locale's equivalent of y, the cp command continues. Any other answer prevents the cp command from overwriting the file.

    +

    -p (preserve) – duplicates the following characteristics of each SourceFile/SourceDirectory in the corresponding TargetFile and/or TargetDirectory:

    +<block_quote>
      +
    • The time of the last data modification and the time of the last access.

    • +
    • The user ID and group ID (only if it has permissions to do this)

    • +
    • The file permission bits and the SUID and SGID bits.

    • +
    </block_quote>

    -R (recursive) – copy directories (recursively copying all the contents)

    +
    +
    +

    +5.1.3. Examples

    +

    To make a copy of a file in the current directory, enter:

    +
    $ cp prog.c prog.bak
    +

    This copies prog.c to prog.bak. If the prog.bak file does not already exist, the cp command creates it. If it does exist, the cp command replaces it with a copy of the prog.c file.

    +

    To copy a file in your current directory into another directory, enter:

    +
    $ cp zaphod /home/books/hhgg
    +

    This copies the jones file to /home/books/hhgg/zaphod.

    +

    To copy a file to a new file and preserve the modification date, time, and access control list associated with the source file, enter:

    +
    $ cp -p martin_luther_king martin_luther_king.jr
    +

    This copies the martin_luther_king file to the martin_luther_king.jr file. Instead of creating the file with the current date and time stamp, the system gives the martin_luther_king.jr file the same date and time as the martin_luther_king file. The martin_luther_king.jr file also inherits the martin_luther_king file's access control protection.

    +

    To copy all the files in a directory to a new directory, enter:

    +
    $ cp /home/galactica/clients/* /home/hhgg/customers
    +

    This copies only the files in the clients directory to the customers directory.

    +

    To copy a directory, including all its files and subdirectories, to another directory, enter:

    +<block_quote>

    $ cp -R /home/hhgg/clients /home/hhgg/customers

    </block_quote>

    This copies the clients directory, including all its files, subdirectories, and the files in those subdirectories, to the customers/clients directory.

    +

    To copy a specific set of files of any extension to another directory, enter:

    +
    $ cp zaphod arthur ford /home/hhgg/clients
    +

    This copies the zaphod, arthur, and ford files in your current working directory to the /home/hhgg/clients directory.

    +

    To use pattern-matching characters to copy files, enter:

    +
    $ cp programs/*.py .
    +

    This copies the files in the programs directory that end with .py to the current directory, signified by the single "." (dot). You must type a space between the py and the final dot.

    +
    +
    +
    +

    +5.2. mv

    +

    mv (short for move) is a Unix command that moves one or more files or directories from one place to another. The original file is deleted, and the new file may have the same or a different name. If possible (i.e. when the original and new files are on the same file system), mv will rename the file instead. Write permission is required on all directories being modified.

    +
    +

    +5.2.1. Conflicting existing file

    +

    In all cases, when a file is moved to have the name of an existing file (in the same directory), the existing file is deleted. If the existing file is not writable but is in a directory that is writable, then the mv command asks for confirmation if possible (i.e. if run from a terminal) before proceeding, unless the -f (force) option is used.

    +
    +
    +

    +5.2.2. Differences with copy and delete

    +

    Note that, usually, when moving files within the same volume, moving (and/or renaming) is not the same as simply copying and then deleting the original. When moving a file, the link is simply removed from the old parent directory and added to the new parent directory. However, the file itself is untouched (i.e. it has the same inodes and resides at the same place on the disk). For example, you cannot copy a file you cannot read, but you can move (and/or rename) it (provided you have write permission to its old and new parent directories). Also, suppose there is a non-empty directory you do not have write permission to. You cannot delete this directory (since you cannot delete its contents); but you can move (and/or rename) it. Also, since moving between filenames on a single volume does not involve copying, it is faster and does not place strain of lots of reads and writes on the disk. Moving files across different volumes, however, does necessitate copying and deleting.

    +
    +
    +

    +5.2.3. Examples

    +
    $ mv myfile mynewfilename    renames a file
    +$ mv myfile otherfilename    renames a file and deletes the existing            file "myfile"
    +$ mv myfile /myfile          moves 'myfile' from the current            directory to the root directory
    +$ mv myfile dir/myfile       moves 'myfile' to 'dir/myfile' relative            to the current directory
    +$ mv myfile dir              same as the previous command (the          filename is implied to be the same)
    +$ mv myfile dir/myfile2      moves 'myfile' to dir and renames it to            'myfile2'
    +$ mv foo bar baz dir         moves multiple files to directory dir
    +$ mv --help                  shows a very concise help about the                syntax of the command
    +$ man mv                     prints an extensive user manual for                'mv' in the terminal
    +

    In all cases, the file or files being moved or renamed can be a directory.

    +

    Note that when the command is called with two arguments (as mv name1 name2 or mv name1 /dir/name2), it can have three different effects, depending on whether name2 does not exist, is an existing file, or is an existing directory. If the user intends to refer to an existing directory, /. (or in some Unix versions / is sufficient) may be appended to the name to force the system to check this. To move a file to a new directory, the directory must be created first.

    +
    +
    +
    +

    +5.3. rm

    +

    rm (short for "remove") is one of several basic Unix command lines that operates on files. It is used to delete files from a filesystem. The data is not actually destroyed. Only the index listing where the file is stored is destroyed, and the storage is made available for reuse. There are undelete utilities that will attempt to reconstruct the index and can bring the file back if the parts were not reused.

    +

    Here's example to remove a file named "foo" from a directory, here shown with the -i option:

    +
    $ rm -i foo
    +remove foo? y
    +
    +

    +5.3.1. Options

    +

    Common options that rm accepts include:

    +<block_quote>
      +
    • -r, which removes directories, removing the contents recursively beforehand (so as not to leave files without a directory to reside in) ("recursive")

    • +
    • -i, which asks for every deletion to be confirmed ("interactive")

    • +
    • -f, which ignores non-existent files and overrides any confirmation prompts ("force")

    • +
    • -v, which shows what is being removed as it happens ("verbose")

    • +
    </block_quote>

    rm is often aliased to "rm -i" so as to avoid accidental deletion of files. If a user still wishes to delete a large number of files without confirmation, they can manually cancel out the -i argument by adding the -f option (as the option specified later on the expanded command line "rm -i -f" takes precedence).

    +

    rm -rf (variously, rm -rf /, rm -rf <title_reference>*</title_reference>, and others) is frequently used in jokes and anecdotes about Unix disasters. The rm -rf variant of the command, if run by a superuser on the root directory, would cause the contents of every writable mounted filesystem on the computer to be deleted.

    +

    rm is often used in conjunction with xargs to supply a list of files to delete:

    +
    xargs rm < filelist
    +

    When rm is used on a symbolic link, it deletes the link, but does not affect the target of the link.

    +
    +
    +

    +5.3.2. Permissions

    +

    Usually, on most filesystems, deleting a file requires write permission on the parent directory (and execute permission, in order to enter the directory in the first place). (Note that, confusingly for beginners, permissions on the file itself are irrelevant. However, GNU rm asks for confirmation if a write-protected file is to be deleted, unless the -f option is used.)

    +

    To delete a directory (with rm -r), one must delete all of its contents recursively. This requires that one must have read and write and execute permission to that directory (if it's not empty) and all non-empty subdirectories recursively (if there are any). The read permissions are needed to list the contents of the directory in order to delete them. This sometimes leads to an odd situation where a non-empty directory cannot be deleted because one doesn't have write permission to it and so cannot delete its contents; but if the same directory were empty, one would be able to delete it.

    +

    If a file resides in a directory with the sticky bit set, then deleting the file requires one to be the owner of the file.

    +
    +
    +
    +
    +

    +6. Command Line Arguments

    +

    In computer command line interfaces, a command line argument is an argument sent to a program being called. In general, a program can take any number of command line arguments, which may be necessary for the program to run, or may even be ignored, depending on the function of that program.

    +

    For example, in Unix and Unix-like environments, an example of a command-line argument is:

    +
    rm file.s
    +

    "file.s" is a command line argument which tells the program rm to remove the file "file.s".

    +

    Programming languages such as C, C++ and Java allow a program to interpret the command line arguments by handling them as string parameters in the main function.

    +

    A command line option or simply option (also known as a command line parameter, flag, or a switch) is an indication by a user that a computer program should change its default output.

    +

    Long options are introduced via "--", and are typically whole words. For example, ls --long --classify --all. Arguments to long options are provided with "=", as ls --block-size=1024. Some Unix programs use long options with single dashes, for example MPlayer as in mplayer -nosound.

    +

    Linux also uses "--" to terminate option lists. For example, an attempt to delete a file called -file1 by using rm -file1 may produce an error, since rm may interpret -file1 as a command line switch. Using rm -- -file1 removes ambiguity.

    +
    +
    +

    +7. Basic Text Processing

    +
    +

    +7.1. head

    +

    head is a program on Unix and Unix-like systems used to display the first few lines of a text file or piped data. The command syntax is:

    +
    $ head [options] <file_name>
    +

    By default, head will print the first 10 lines of its input to the standard output. The number of lines printed may be changed with a command line option. The following example shows the first 20 lines of filename:

    +
    $ head -n 20 filename
    +

    This displays the first 5 lines of all files starting with foo:

    +
    $ head -n 5 foo*
    +

    Some versions omit the n and just let you say -5.

    +
    +

    +7.1.1. Flags

    +
    -c <x number of bytes> Copy first x number of bytes.
    +

    Other options: sed

    +

    Many early versions of Unix did not have this command, and so documentation and books had sed do this job:

    +
    sed 5q foo
    +

    This says to print every line (implicit), and quit after the fifth.

    +
    +
    +
    +

    +7.2. tail

    +

    tail is a program on Unix and Unix-like systems used to display the last few lines of a text file or piped data.

    +

    The command-syntax is:

    +
    $ tail [options] <file_name>
    +

    By default, tail will print the last 10 lines of its input to the standard output. With command line options the number of lines printed and the printing units (lines, blocks or bytes) may be changed. The following example shows the last 20 lines of filename:

    +
    $ tail -n 20 filename
    +

    This example shows the last 15 bytes of all files starting with foo:

    +
    $ tail -c 15 foo*
    +

    This example shows all lines of filename from the second line onwards:

    +
    $ tail -n +2 filename
    +

    Using an older syntax (still used in Sun Solaris as the -n option is not supported), the last 20 lines and the last 50 bytes of filename can be shown with the following command:

    +
    $ tail -20 filename
    +$ tail -50c filename
    +

    However this syntax is now obsolete and does not conform with the POSIX 1003.1-2001 standard. Even if still supported in current versions, when used with other options (like -f, see below), these switches could not work at all.

    +
    +

    +7.2.1. File monitoring

    +

    tail has a special command line option -f (follow) that allows a file to be monitored. Instead of displaying the last few lines and exiting, tail displays the lines and then monitors the file. As new lines are added to the file by another process, tail updates the display. This is particularly useful for monitoring log files. The following command will display the last 10 lines of messages and append new lines to the display as new lines are added to messages:

    +
    $ tail -f /var/adm/messages
    +

    To interrupt tail while it is monitoring, break-in with Ctrl+C. This command can be run "in the background" with &, see job control.

    +

    If you have a command's result to monitor, you can use the watch command.

    +
    +
    +
    +

    +7.3. cut

    +

    In computing, cut is a Unix command line utility which is used to extract sections from each line of input — usually from a file.

    +

    Extraction of line segments can typically be done by bytes (-b), characters (-c), or fields (-f) separated by a delimiter (-d — the tab character by default). A range must be provided in each case which consists of one of N, N-M, N- (N to the end of the line), or -M (beginning of the line to M), where N and M are counted from 1 (there is no zeroth value). Since version 6, an error is thrown if you include a zeroth value. Prior to this the value was ignored and assumed to be 1.

    +

    Assuming a file named file containing the lines:

    +
    foo:bar:baz:qux:quux
    +one:two:three:four:five:six:seven
    +alpha:beta:gamma:delta:epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +

    To output the fourth through tenth characters of each line:

    +
    $ cut -c 4-10 file
    +

    This gives the output:

    +
    :bar:ba
    +:two:th
    +ha:beta
    +

    To output the fifth field through the end of the line of each line using the colon character as the field delimiter:

    +
    $ cut -d : -f 5- file
    +

    This gives the output:

    +
    quux
    +five:six:seven
    +epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +
    +
    +

    +7.4. paste

    +

    paste is a Unix command line utility which is used to join files horizontally (parallel merging) by outputting lines consisting of the sequentially corresponding lines of each file specified, separated by tabs, to the standard output. It is effectively the horizontal equivalent to the utility cat command which operates on the vertical plane of two or more files.

    +

    To paste several columns of data together into the file www from files who, where, and when:

    +
    $ paste who where when > www
    +

    If the files contain:

    +++++
    +

    This creates the file named www containing:

    +
    Batman            GothamCity       January 3
    +Trillian          Andromeda        February 4
    +Jeeves            London           March 19
    +
    +
    +
    +

    +8. Shell Meta Characters

    +

    Unix recognizes certain special characters, called "meta characters," as command directives. The shell meta characters are recognized anywhere they appear in the command line, even if they are not surrounded by blank space. For that reason, it is safest to only use the characters A-Z, a-z, 0-9, and the period, dash, and underscore characters when naming files and directories on Unix. If your file or directory has a shell meta character in the name, you will find it difficult to use the name in a shell command.

    +

    The shell meta characters include:

    +

    / < > ! $ % ^ & * | { } [ ] " ' ` ~ ;

    +

    Different shells may differ in the meta characters recognized.

    +

    As an example,

    +
    $ ls file.*
    +

    run on a directory containing the files file, file.c, file.lst, and myfile would list the files file.c and file.lst. However,:

    +
    $ ls file.?
    +

    run on the same directory would only list file.c because the ? only matches one character, no more, no less. This can save you a great deal of typing time. For example, if there is a file called california_cornish_hens_with_wild_rice and no other files whose names begin with 'c', you could view the file without typing the whole name by typing this:

    +
    $ more c*
    +

    because the c* matches that long file name.

    +

    Filenames containing metacharacters can pose many problems and should never be intentionally created. If you do find that you've created a file with metacharacters, and you would like to remove it, you have three options. You may use wildcards to match metacharacter, use the to directly enter the filename, or put the command in double quotes (except in the case of double quotes within the file name, these must be captured with one of the first two methods). For example, deleting a file named <title_reference>"``*`|more</title_reference>"` can be accomplished with:

    +
    $ rm ??more
    +

    or:

    +
    $ rm $\backslash$*$\backslash$|more
    +

    or:

    +
    $ rm ''*|more''
    +
    +
    +

    +9. Looking At Files

    +
    +

    +9.1. cat

    +

    The cat command is a standard Unix program used to concatenate and display files. The name is from "catenate", a synonym of concatenate.

    +

    The Single Unix Specification specifies the behavior that the contents of each of the files given in sequence as arguments will be written to the standard output in the same sequence, and mandates one option, -u, where each byte is printed as it is read.

    +

    If the filename is specified as -, then cat will read from standard input at that point in the sequence. If no files are specified, cat will read from standard input entered.

    +
    +

    +9.1.1. Jargon File Definition

    +

    The Jargon File version 4.4.7 lists this as the definition of cat:

    +
    1. To spew an entire file to the screen or some other output sink without
    +     pause (syn. blast).
    +
    +2. By extension, to dump large amounts of data at an unprepared target or
    +     with no intention of browsing it carefully. Usage: considered silly.
    +     Rare outside Unix sites. See also dd, BLT.
    +
    +     Among Unix fans, *cat(1)* is considered an excellent example of
    +     user-interface design, because it delivers the file contents without
    +     such verbosity as spacing or headers between the files, and because
    +     it does not require the files to consist of lines of text, but works
    +     with any sort of data.
    +
    +     Among Unix critics, *cat(1)* is considered the canonical example of
    +     bad user-interface design, because of its woefully unobvious name.
    +     It is far more often used to blast a single file to standard output
    +     than to concatenate two or more files. The name cat for the former
    +     operation is just as unintuitive as, say, LISP's cdr.
    +
    +     Of such oppositions are holy wars made...
    +
    +
    +

    +9.1.2. Useless Use of 'cat'

    +

    UUOC (from comp.unix.shell on Usenet) stands for “Useless Use of cat”. As it is observed on comp.unix.shell, “The purpose of cat is to concatenate (or 'catenate') files. If it's only one file, concatenating it with nothing at all is a waste of time, and costs you a process.”

    +

    Nevertheless one sees people doing:

    +
    $ cat file | some_command and its args ...
    +

    instead of the equivalent and cheaper:

    +
    <file some_command and its args ...
    +

    or (equivalently and more classically):

    +
    some_command and its args ... <file
    +

    Since 1995, occasional awards for UUOC have been given out. The activity of fixing instances of UUOC is sometimes called 'demoggification'.

    +

    Amongst many, it is still considered safer to use cat for such cases given that the < and > keys are next to each other in many popular keyboard mappings. While the risk might be low, the impact of using > instead of < can be high and prohibitive.

    +
    +
    +

    +9.1.3. zcat

    +

    zcat is a Unix program similar to cat, that decompresses individual files and concatenates them to standard output. Traditionally zcat operated on files compressed by compress but today it is usually able to operate on gzip or even bzip2 archives. On such systems, it is equivalent to gunzip -c

    +
    +
    +
    +

    +9.2. more

    +

    In computing, more is a command to view (but not modify) the contents of a text file one screen at a time (terminal pager). It is available on Unix and Unix-like systems, DOS, OS/2 and Microsoft Windows. Programs of this sort are called pagers.

    +
    +

    +9.2.1. Usage

    +

    The command-syntax is:

    +
    $ more [options] [file_name]
    +

    If no file name is provided, more looks for input from stdin.

    +

    Once more has obtained input, it displays as much as can fit on the current screen and waits for user input to advance, with the exception that a form feed (^L) will also cause more to wait at that line, regardless of the amount of text on the screen. In the lower-left corner of the screen is displayed the text "--More--" and a percentage, representing the percent of the file that more has paged through. (This percentage includes the text displayed on the current screen.) When more reaches the end of a file (100%) it exits. The most common methods of navigating through a file are Enter, which advances the output by one line, and Space, which advances the output by one screen.

    +

    There are also other commands that can be used while navigating through the document; consult more's man page for more details.

    +

    Options are typically entered before the file name, but can also be entered in the environment variable $MORE. Options entered in the actual command line will override those entered in the $MORE environment variable. Available options may vary between Unix systems.

    +
    +
    +
    +

    +9.3. less

    +

    less is a terminal pager program on Unix, Windows and Unix-like systems used to view (but not change) the contents of a text file one screen at a time. It is similar to more, but has the extended capability of allowing both forward and backward navigation through the file. Unlike most Unix text editors/viewers, less does not need to read the entire file before starting, resulting in faster load times with large files.

    +
    +

    +9.3.1. Usage

    +

    less can be invoked with options to change its behaviour, for example, the number of lines to display on the screen. A few options vary depending on the operating system. While less is displaying the file, various commands can be used to navigate through the file. These commands are based on those used by both more and vi. It is also possible to search for character patterns in the file.

    +

    By default, less displays the contents of the file to the standard output (one screen at a time). If the file name argument is omitted, it displays the contents from standard input (usually the output of another command through a pipe). If the output is redirected to anything other than a terminal, for example a pipe to another command, less behaves like cat.

    +

    The command-syntax is:

    +
    $ less [options] file_name
    +
    +
    +

    +9.3.2. Frequently Used Options

    +<block_quote>
      +
    • -g: Highlights just the current match of any searched string.

    • +
    • -I: Case-insensitive searches.

    • +
    • -M: Shows more detailed prompt, including file position.

    • +
    • -N: Shows line numbers (useful for source code viewing).

    • +
    • -S: Disables line wrap ("chop long lines"). Long lines can be seen by side scrolling.

    • +
    • -?: Shows help.

    • +
    </block_quote>
    +
    +
    +

    +9.3.3. Frequently Used Commands

    +<block_quote>
      +
    • [Arrows]/[Page Up]/[Page Down]/[Home]/[End]: Navigation.

    • +
    • [Space bar]: Next page.

    • +
    • b: Previous page.

    • +
    • ng: Jump to line number n. Default is the start of the file.

    • +
    • nG: Jump to line number n. Default is the end of the file.

    • +
    • /pattern: Search for pattern. Regular expressions can be used.

    • +
    • '^ or g: Go to start of file.

    • +
    • '$ or G: Go to end of file.

    • +
    • s: Save current content (got from another program like grep) in a file.

    • +
    • =: File information.

    • +
    • h: Help.

    • +
    • q: Quit.

    • +
    </block_quote>
    +
    +
    +

    +9.3.4. Examples

    +
    $ less -M readme.txt                     #Read "readme.txt."
    +$ less +F /var/log/mail.log              #Follow mode for log
    +$ file * | less                          #Easier file analysis.
    +$ grep -i void *.c | less -I -p void     #Case insensitive search                                                         for "void" in all .c files
    +
    +
    +
    +
    +

    +10. Directory Structure

    +

    In the File Hierarchy Standard (FHS) all files and directories appear under the root directory "/", even if they are stored on different physical devices. Note however that some of these directories may or may not be present on a Unix system depending on whether certain subsystems, such as the X Window System, are installed.

    +

    The majority of these directories exist in all UNIX operating systems and are generally used in much the same way; however, the descriptions here are those used specifically for the FHS, and are not considered authoritative for platforms other than Linux.

    ++++
    +
    +

    +10.1. man hier

    +

    This is the manual page on the UNIX filesystem. The syntax for this is:

    +
    $ man hier
    +
    +
    +

    +10.2. ls -l

    +

    Shows you huge amounts of information (permissions, owners, size, and when last modified) for folders and files. The syntax is

    +
    $ ls -l
    +

    This can be done after entering the required directory.

    +
    +
    +
    +

    +11. Permissions and Ownership

    +
    +

    +11.1. chmod

    +

    The chmod command (abbreviated from 'change mode') is a shell command and C language function in Unix and Unix-like environments. When executed, it can change file system modes of files and directories. The modes include permissions and special modes.A chmod command first appeared in AT&T Unix version 1, and is still used today on Unix-like machines.

    +
    +

    +11.1.1. Usage

    +

    The chmod command options are specified like this:

    +
    $ chmod [options] mode[,mode] file1 [file2 ...]
    +

    To view what the permissions currently are, type:

    +
    $ ls -l file
    +
    +
    +

    +11.1.2. Command line options

    +

    The chmod command has a number of command line options that affect its behavior. The most common options are:

    +<block_quote>
      +
    • -R: Changes the modes of directories and files recursively

    • +
    • -v: Verbose mode; lists all files as they are being processed

    • +
    </block_quote>
    +
    +11.1.2.1. Symbolic modes
    +

    To the chmod utility, all permissions and special modes are represented by its mode parameter. One way to adjust the mode of files or directories is to specify a symbolic mode. The symbolic mode is composed of three components, which are combined to form a single string of text:

    +
    $ chmod [references][operator][modes] file1 ...
    +

    The references (or classes) are used to distinguish the users to whom the permissions apply. If no references are specified it defaults to “all” but modifies only the permissions allowed by the umask. The references are represented by one or more of the following letters:

    +++++
    +

    The chmod program uses an operator to specify how the modes of a file should be adjusted. The following operators are accepted:

    ++++
    +

    The modes indicate which permissions are to be granted or taken away from the specified classes. There are three basic modes which correspond to the basic permissions:

    +++++
    +

    The combination of these three components produces a string that is understood by the chmod command. Multiple changes can be specified by separating multiple symbolic modes with commas.

    +
    +
    +
    +11.1.2.2. Symbolic examples
    +

    Add the 'read' and 'write' permissions to the 'user' and 'group' classes of a directory:

    +
    $ chmod ug+rw mydir
    +$ ls -ld mydir
    +drw-rw----   2 starwars  yoda  96 Dec 8 12:53 mydir
    +

    For a file, remove write permissions for all classes:

    +
    $ chmod a-w myfile
    +$ ls -l myfile
    +-r-xr-xr-x   2 starwars  yoda 96 Dec 8 12:53 myfile
    +

    Set the permissions for the u*ser and the *g*roup to read and execute only (no write permission) on *mydir.

    +
    $ chmod ug=rx mydir
    +$ ls -ld mydir
    +dr-xr-x---   2 starwars  yoda 96 Dec 8 12:53 mydir
    +
    +
    +
    +11.1.2.3. Octal numbers
    +

    The chmod command also accepts three and four-digit octal numbers representing modes. Using a three-digit octal number to set the modes of a file named myfile :

    +
    $ chmod 664 myfile
    +$ ls -l myfile
    +-rw-rw-r--  1   57 Jul  3 10:13  myfile
    +

    Since the setuid, setgid and sticky bits are not set, this is equivalent to:

    +
    $ chmod 0664 myfile
    +
    +
    +
    +11.1.2.4. Special modes
    +

    The chmod command is also capable of changing the additional permissions or special modes of a file or directory. The symbolic modes use s to represent the setuid and setgid modes, and t to represent the sticky mode. The modes are only applied to the appropriate classes, regardless of whether or not other classes are specified.

    +

    Most operating systems support the specification of special modes using octal modes, but some do not. On these systems, only the symbolic modes can be used.

    +
    +
    +
    +
    +
    +

    +12. Redirection and Piping

    +

    In computing, redirection is a function common to most command-line interpreters, including the various Unix shells that can redirect standard streams to user-specified locations.

    +

    Programs do redirection with the dup2(2) system call, or its less-flexible but higher-level stdio analogues, freopen(3) and popen(3).

    +
    +

    +12.1. Redirecting standard input and standard output

    +

    Redirection is usually implemented by placing certain characters between commands. Typically, the syntax of these characters is as follows:

    +
    $ command1 > file1
    +

    executes command1, placing the output in file1. Note that this will truncate any existing data in file1. To append output to the end of the file, use the >> operator.:

    +
    $ command1 < file1
    +

    executes command1, using file1 as the source of input (as opposed to the keyboard).:

    +
    $ command1 < infile > outfile
    +

    combines the two capabilities: command1 reads from infile and writes to outfile

    +
    +
    +

    +12.2. Piping

    +

    Programs can be run together such that one program reads the output from another with no need for an explicit intermediate file: +A pipeline of three programs run on a text terminal:

    +
    $ command1 | command2
    +

    executes command1, using its output as the input for command2 (commonly called piping, since the "|" character is known as a "pipe").

    +

    This is equivalent to using two redirects and a temporary file:

    +
    $ command1 > tempfile
    +$ command2 < tempfile
    +$ rm tempfile
    +

    A good example for command piping is combining echo with another command to achieve something interactive in a non-interactive shell, e.g.:

    +
    $ echo -e "user\npass" | ftp localhost
    +

    This runs the ftp client with input user, press return, then pass.

    +
    +
    +

    +12.3. Redirecting to and from the standard file handles

    +

    In Unix shells derived from the original Bourne shell, the first two actions can be further modified by placing a number (the file descriptor) immediately before the character; this will affect which stream is used for the redirection. The Unix standard I/O streams are:

    +++++
    +

    For example:

    +
    $ command1 2> file1
    +

    executes command1, directing the standard error stream to file1.

    +

    In shells derived from csh (the C shell), the syntax instead appends the & character to the redirect characters, thus achieving a similar result.

    +

    Another useful capability is to redirect one standard file handle to another. The most popular variation is to merge standard error into standard output so error messages can be processed together with (or alternately to) the usual output. Example:

    +
    $ find / -name .profile > results 2>&1
    +

    will try to find all files named .profile. Executed without redirection, it will output hits to stdout and errors (e.g. for lack of privilege to traverse protected directories) to stderr. If standard output is directed to file results, error messages appear on the console. To see both hits and error messages in file results, merge stderr (handle 2) into stdout (handle 1) using 2>&1 .

    +

    It's possible use 2>&1 before ">" but it doesn't work. In fact, when the interpreter reads 2>&1, it doesn't know yet where standard output is redirected and then standard error isn't merged.

    +

    If the merged output is to be piped into another program, the file merge sequence 2>&1 must precede the pipe symbol, thus:

    +
    $ find / -name .profile 2>&1 | less
    +

    A simplified form of the command:

    +
    $ command > file 2>&1
    +

    is:

    +
    $ command &>file
    +

    or:

    +
    $command >&file
    +
    +
    +

    +12.4. Chained pipelines

    +

    The redirection and piping tokens can be chained together to create complex commands. For example:

    +
    $ ls | grep '\.sh' | sort > shlist
    +

    lists the contents of the current directory, where this output is filtered to only contain lines which contain .sh, sort this resultant output lexicographically, and place the final output in shlist. This type of construction is used very commonly in shell scripts and batch files.

    +
    +
    +

    +12.5. Redirect to multiple outputs

    +

    The standard command tee can redirect output from a command to several destinations.

    +
    $ ls -lrt | tee xyz
    +

    This directs the file list output to both standard output as well as to the file xyz.

    +
    +
    +
    +

    +13. More Text Processing

    +
    +

    +13.1. grep

    +

    grep is a command line text search utility originally written for Unix. The name is taken from the first letters in global / regular expression / print, a series of instructions for the ed text editor. The grep command searches files or standard input globally for lines matching a given regular expression, and prints them to the program's standard output.

    +
    +

    +13.1.1. Usage

    +

    This is an example of a common grep usage:

    +
    $ grep apple fruitlist.txt
    +

    In this case, grep prints all lines containing 'apple' from the file fruitlist.txt, regardless of word boundaries; therefore lines containing 'pineapple' or 'apples' are also printed. The grep command is case sensitive by default, so this example's output does not include lines containing 'Apple' (with a capital A) unless they also contain 'apple'.

    +

    Like most Unix commands, grep accepts command line arguments to change this and many other behaviors. For example:

    +
    $ grep -i apple fruitlist.txt
    +

    This prints all lines containing 'apple' regardless of capitalization. The '-i' argument tells grep to be case insensitive, or to ignore case.

    +

    To print all lines containing 'apple' as a word ('pineapple' and 'apples' will not match):

    +
    $ grep -w apple fruitlist.txt
    +

    Regular expressions can be used to match more complicated queries.

    +
    +
    +13.1.1.1. Variations
    +

    There are countless implementations and derivatives of grep available for many operating systems. Early variants of grep included egrep and fgrep. The former applies an extended regular expression syntax that was added to Unix after Ken Thompson's original regular expression implementation. The latter searches for any of a list of 'fixed' strings using the Aho-Corasick algorithm. These variants are embodied in most modern grep implementations as command-line switches (and standardized as -E and -F in POSIX). In such combined implementations, grep may also behave differently depending on the name by which it is invoked, allowing fgrep, egrep, and grep to be links to the same program.

    +

    pcregrep is an implementation of grep that uses Perl regular expression syntax.

    +

    Other commands contain the word 'grep' to indicate that they search (usually for regular expression matches). The pgrep utility, for instance, displays the processes whose names match a given regular expression.

    +
    +
    +
    +
    +

    +13.2. tr

    +

    tr (abbreviated from translate or transliterate) is a command in Unix-like operating systems.

    +

    When executed, the program reads from the standard input and writes to the standard output. It takes as parameters two sets of characters, and replaces occurrences of the characters in the first set with the corresponding elements from the other set. For example,

    +
    $ tr 'abcd' 'jkmn'
    +

    maps 'a' to 'j', 'b' to 'k', 'c' to 'm', and 'd' to 'n'.

    +

    Sets of characters may be abbreviated by using character ranges. The previous example could be written:

    +
    $ tr 'a-d' 'jkmn'
    +

    In POSIX compliant versions of tr the set represented by a character range depends on the locale's collating order, so it is safer to avoid character ranges in scripts that might be executed in a locale different from that in which they were written. Ranges can often be replaced with POSIX character sets such as [:alpha:].

    +

    The -c flag complements the first set of characters.

    +
    $ tr -cd '[:alnum:]'
    +

    therefore removes all non-alphanumeric characters.

    +

    The -s flag causes tr to compress sequences of identical adjacent characters in its output to a single token. For example,

    +
    $ tr -s '\n' '\n'
    +

    replaces sequences of one or more newline characters with a single newline.

    +

    The -d flag causes tr to delete all tokens of the specified set of characters from its input. In this case, only a single character set argument is used. The following command removes carriage return characters, thereby converting a file in DOS/Windows format to one in Unix format.

    +
    $ tr -d '\r'
    +

    Most versions of tr, including GNU tr and classic Unix tr, operate on single byte characters and are not Unicode compliant. An exception is the Heirloom Toolchest implementation, which provides basic Unicode support.

    +

    Ruby and Perl also have an internal tr operator, which operates analogously. Tcl's string map command is more general in that it maps strings to strings while tr maps characters to characters.

    +
    +
    +
    +

    +14. Elementary Regex

    +

    In computing, regular expressions provide a concise and flexible means for identifying strings of text of interest, such as particular characters, words, or patterns of characters. A regular expression (often shortened to regex or regexp) is written in a formal language that can be interpreted by a regular expression processor, a program that either serves as a parser generator or examines text and identifies parts that match the provided specification.

    +

    Regular expressions are used by many text editors, utilities, and programming languages to search and manipulate text based on patterns. For example, Perl, Ruby and Tcl have a powerful regular expression engine built directly into their syntax. Several utilities provided by Unix distributions—including the editor ed and the filter grep — were the first to popularize the concept of regular expressions.

    +

    Traditional Unix regular expression syntax followed common conventions but often differed from tool to tool. The IEEE POSIX Basic Regular Expressions (BRE) standard (released alongside an alternative flavor called Extended Regular Expressions or ERE) was designed mostly for backward compatibility with the traditional (Simple Regular Expression) syntax but provided a common standard which has since been adopted as the default syntax of many Unix regular expression tools, though there is often some variation or additional features. Many such tools also provide support for ERE syntax with command line arguments.

    +

    In the BRE syntax, most characters are treated as literals — they match only themselves (i.e., a matches "a"). The exceptions, listed below, are called metacharacters or metasequences.

    ++++
    +
    +

    +14.1. Lazy quantification

    +

    The standard quantifiers in regular expressions are greedy, meaning they match as much as they can, only giving back as necessary to match the remainder of the regex. For example, someone new to regexes wishing to find the first instance of an item between < and > symbols in this example:

    +
    Another whale explosion occurred on <January 26>, <2004>.
    +

    ...would likely come up with the pattern <.*>, or similar. However, this pattern will actually return "<January 26>, <2004>" instead of the "<January 26>" which might be expected, because the <title_reference>*</title_reference> quantifier is greedy — it will consume as many characters as possible from the input, and "January 26>, <2004" has more characters than "January 26".

    +

    Though this problem can be avoided in a number of ways (e.g., by specifying the text that is not to be matched: <[^>]*>), modern regular expression tools allow a quantifier to be specified as lazy (also known as non-greedy, reluctant, minimal, or ungreedy) by putting a question mark after the quantifier (e.g., <.*?>), or by using a modifier which reverses the greediness of quantifiers (though changing the meaning of the standard quantifiers can be confusing). By using a lazy quantifier, the expression tries the minimal match first. Though in the previous example lazy matching is used to select one of many matching results, in some cases it can also be used to improve performance when greedy matching would require more backtracking.

    +
    +
    +
    +

    +15. One Liners

    +

    A one-liner is textual input to the command-line of an operating system shell that performs some function in just one line of input.

    +

    The one liner can be

    +<block_quote>
      +
    1. An expression written in the language of the shell.

    2. +
    3. The invocation of an interpreter together with program source for the interpreter to run.

    4. +
    5. The invocation of a compiler together with source to compile and +instructions for executing the compiled program.

    6. +
    </block_quote>

    Certain dynamic scripting languages such as AWK, sed, and perl have traditionally been adept at expressing one-liners. Specialist shell interpreters such as these Unix shells or the Windows PowerShell, allow for the construction of powerful one-liners.

    +

    The use of the phrase one-liner has been widened to also include program-source for any language that does something useful in one line.

    +

    The word One-liner has two references in the index of the book The AWK Programming Language (the book is often referred to by the abbreviation TAPL). It explains the programming language AWK, which is part of the Unix operating system. The authors explain the birth of the One-liner paradigm with their daily work on early Unix machines:

    +
    “The 1977 version had only a few built-in variables and predefined functions. It was designed for writing short programs [...] Our model was that an invocation would be one or two lines long, typed in and used immediately. Defaults were chosen to match this style [...] We, being the authors, knew how the language was supposed to be used, and so we only wrote one-liners.”
    +

    Notice that this original definition of a One-liner implies immediate execution of the program without any compilation. So, in a strict sense, only source code for interpreted languages qualifies as a One-liner. But this strict understanding of a One-liner was broadened in 1985 when the IOCCC introduced the category of Best One Liner for C, which is a compiled language.

    +

    The TAPL book contains 20 examples of One-liners (A Handful of Useful awk One-Liners) at the end of the book's first chapter.

    +

    Here are the first few of them:

    +<block_quote>
      +
    1. +

      Print the total number of input lines:

      +

      END { print NR }

      +
    2. +
    3. +

      Print the tenth input line:

      +

      NR == 10

      +
    4. +
    5. +

      Print the last field of every input line:

      +

      { print $NF }

      +
    6. +
    </block_quote>

    One-liners are also used to show off the differential expressive power of programming languages. Frequently, one-liners are used to demonstrate programming ability. Contests are often held to see who can create the most exceptional one-liner.

    +

    The following example is a C program (a winning entry in the "Best one-liner" category of the IOCCC, here split to two lines for presentation).:

    +
    main(int c,char**v){return!m(v[1],v[2]);}m(char*s,char*t){return
    +*t-42?*s?63==*t|*s==*t&&m(s+1,t+1):!*t:m(s,t+1)||*s&&m(s+1,t);}
    +

    This one-liner program is a glob pattern matcher. It understands the glob characters '*' meaning 'zero or more characters' and '?' meaning exactly one character, just like most Unix shells.

    +

    Run it with two args, the string and the glob pattern. The exit status is 0 (shell true) when the pattern matches, 1 otherwise. The glob pattern must match the whole string, so you may want to use * at the beginning and end of the pattern if you are looking for something in the middle. Examples:

    +
    $ prog foo 'f??'; echo $?
    +
    +$ prog 'best short program' '??st*o**p?*'; echo $?
    +

    Here is a one line shell script to show directories:

    +
    $ ls -R | grep ":$" | sed -e 's/:$//' -e 's/[^-][^\/]*\//--/g' -e 's/^/   /' -e 's/-/|/'
    +
    +
    +
    + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/ch3list_tuples.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/ch3list_tuples.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1138 @@ + + + +List and Tuples  + + + + + + +
    +
    + +
    +
    +

    +
    +
    + +
    +

    +1. Introducing Linux

    +

    (Attribution : A significant chunk of the content under this section is based on data from Wikipedia and the Linux Documentation Project)

    +

    Linux (usually pronounced ˈlɪnəks') is a generic term referring to Unix-like computer operating systems based on the Linux kernel, where a kernel is the intermediate layer between the hardware and the applications. The kernel is, on an abstract level, the core of (most) operating systems, that manages the various system resources. The development of the Linux OS is considered the basis for Free and Open Source Software (FOSS) collaboration since typically the underlying source code can be used, modified freely, and redistributed by anyone under the terms of the GNU (a recursive acronym for "GNU's Not Unix!") Global Public License (GPL) and other free software licences. This freedom to access and reuse various components of a system, is one of the primary reasons for the popularity of Linux.

    +

    Linux is installed on a variety of computer hardware, that include mobile phones, embedded devices and supercomputers, but is infamous for its use in servers.

    +

    The name "Linux" comes from the Linux kernel, originally written in 1991 by Linus Torvalds. The rest of the system usually comprises components such as the Apache HTTP Server, the X Window System, the GNOME and KDE desktop environments, and utilities and libraries from the GNU Project (announced in 1983 by Richard Stallman). Commonly-used applications with desktop Linux systems include the Mozilla Firefox web-browser and the OpenOffice.org office application suite. The GNU contribution is the basis for the Free Software Foundation's preferred name GNU/Linux. The kernel's mascot is a penguin named "Tux". Mozilla Firefox and OpenOffice.org are open-source projects which can be run on most Operating Systems, including proprietary ones.

    +
    +

    +1.1. Historical Background

    +
    +

    +1.1.1. Events leading to the creation

    +
      +
    • The Unix operating system was developed in the 1960s and released for public use in 1970. Its accessibility and portability caused it to be widely adopted, copied and modified by academic institutions and businesses. Its design became influential to authors of other systems. Other free operating systems include the Berkeley Software Distribution (BSD), developed at the University of California at Berkeley, and MINIX which was released by Andrew S. Tanenbaum. The development and adoption of BSD and MINIX were limited due to various reasons, and this lack of a widely-adopted and free kernel triggered Linus Torvalds into starting his project.

    • +
    • In 1983, Richard Stallman started the GNU project with the goal of creating a free UNIX-like operating system. As part of this work, he wrote the GNU General Public License (GPL). By the early 1990s there was almost enough available software to create a full operating system. However, the GNU kernel, called Hurd, failed to attract enough attention from developers leaving GNU incomplete.

    • +
    +
    +
    +

    +1.1.2. The Creation of Linux

    +

    In 1991, Linus Torvalds began a project at the University of Helsinki that later became the Linux kernel. It was initially a terminal (command-line) emulator, which Torvalds used to access the large UNIX servers of the university. He wrote the program targeting just the hardware he was using and independent of an operating system because he wanted to use the functions of his computer with an 80386 processor. Development was done on Minix using the GNU C compiler. This application is still the main choice for compiling Linux today (although the code can be built with other compilers, such as the Intel C Compiler).

    +

    Torvalds continues to direct the development of the kernel. Stallman heads the Free Software Foundation, which in turn supports the GNU components. Finally, individuals and corporations develop third-party non-GNU components, which constitute a vast body of work and including kernel modules, and user applications and libraries. Linux vendors and communities combine and distribute the kernel, GNU components, and non-GNU components, with additional package management software in the form of Linux distributions.

    +
    +
    +
    +

    +1.2. Design and Implications

    +

    A Linux-based system is a modular Unix-like operating system, deriving much of its basic design from principles established in Unix earlier. Such a system uses a monolithic kernel, called the Linux kernel, which handles process control, networking, and peripheral and file system access. Device drivers are integrated directly with the kernel. Separate projects that interface with the kernel provide much of the system's higher-level functionality. The GNU userland is an important part of most Linux-based systems, providing the most common implementation of the C library, a popular shell, and many of the common Unix tools which carry out many basic operating system tasks. The graphical user interface (or GUI) used by most Linux systems is based on the "X Window System".

    +
    +

    +1.2.1. User Interface

    +

    Users can control a Linux-based system through a command line interface (or CLI), a graphical user interface (or GUI), or through controls attached to the associated hardware (this is common for embedded systems). For desktop systems, the default mode is usually the GUI. On desktop machines, "KDE", "GNOME" and "Xfce" are the most popular user interfaces,though a variety of additional user interfaces exist. Most popular user interfaces run on top of the "X Window System" (or X), which enables a graphical application running on one machine to be displayed and controlled from another in a network.

    +

    A Linux system also provides a CLI of some sort through a shell, which is the traditional way of interacting with a Unix system. A Linux distribution specialized for servers may use the CLI as its only interface. A “headless system” (system run without even a monitor) can be controlled by the command line via a remote-control protocol such as SSH or telnet. The CLI is particularly suited for automation of repetitive or delayed tasks, and provides very simple inter-process communication. A graphical terminal emulator program is often used to access the CLI from a Linux desktop.

    +
    +
    +

    +1.2.2. Development

    +

    The primary difference between Linux and many other popular contemporary operating systems is that the Linux kernel and other components are free and open source software. Linux is not the only such operating system, although it is by far the most widely used. Some free and open source software licenses are based on the principle of "copyleft", a kind of reciprocity: any work derived from a copyleft piece of software must also be copyleft itself. The most common free software license, the GNU GPL, is a form of copyleft, and is used for the Linux kernel and many of the components from the GNU project.

    +

    Linux based distributions are intended by developers for interoperability with other operating systems and established computing standards. Linux systems adhere to POSIX, SUS, ISO and ANSI standards where possible, although to date only one Linux distribution has been POSIX.1 certified, Linux-FT.Free software projects, although developed in a collaborative fashion, are often produced independently of each other. The fact that the software licenses explicitly permit redistribution, however, provides a basis for larger scale projects that collect the software produced by stand-alone projects and make it available all at once in the form of a Linux distribution.

    +

    A Linux distribution, commonly called a "distro", is a project that manages a remote collection of system software and application software packages available for download and installation through a network connection. This allows the user to adapt the operating system to his/her specific needs. Distributions are maintained by individuals, loose-knit teams, volunteer organizations, and commercial entities. A distribution can be installed using a CD that contains distribution-specific software for initial system installation and configuration. A package manager such as Synaptic or YAST allows later package upgrades and installations. A distribution is responsible for the default configuration of the installed Linux kernel, general system security, and more generally integration of the different software packages into a coherent whole.

    +
    +
    +

    +1.2.3. Community

    +

    A distribution is largely driven by its developer and user communities. Some vendors develop and fund their distributions on a volunteer basis. Examples include Debian and the Debian-based, Ubuntu. Others maintain a community version of their commercial distributions, as Red Hat does with Fedora.

    +

    In many cities and regions, local associations known as Linux Users Groups (LUGs) seek to promote their preferred distribution and by extension free software. They hold meetings and provide free demonstrations, training, technical support, and operating system installation to new users. Many Internet communities also provide support to Linux users and developers. Most distributions and free software / open source projects have IRC (Internet Relay Chat) chatrooms or newsgroups. Online forums are another means for support. Linux distributions host mailing lists; commonly there will be a specific topic such as usage or development for a given list. All these can be found simply by running an appropriate search on Google.

    +

    Although Linux distributions are generally available without charge, several large corporations sell, support, and contribute to the development of the components of the system and of free software. These include Dell, IBM, HP, Oracle, Sun Microsystems, Novell, Nokia. A number of corporations, notably Red Hat, have built their entire business around Linux distributions.

    +
    +
    +

    +1.2.4. Can I make a profit out of running a business involving Linux?

    +

    The answer is, "Yes!". The free software licenses, on which the various software packages of a distribution built on the Linux kernel are based, explicitly accommodate and encourage commercialization; the relationship between a Linux distribution as a whole and individual vendors may be seen as symbiotic. One common business model of commercial suppliers is charging for support, especially for business users. A number of companies also offer a specialized business version of their distribution, which adds proprietary support packages and tools to administer higher numbers of installations or to simplify administrative tasks. Another business model is to give away the software in order to sell hardware. Examples of corporations that are extensively (and sometimes exclusively) open-source and Linux-powered , with successful revenue generation models involving these, are Google, SUN, Mozilla, etc.

    +
    +
    +

    +1.2.5. Programming on Linux

    +

    Most Linux distributions support dozens of programming languages. The most common collection of utilities for building both Linux applications and operating system programs is found within the GNU toolchain, which includes the GNU Compiler Collection (GCC) and the GNU build system. Amongst others, GCC provides compilers for Ada, C, C++, Java, and Fortran. The Linux kernel itself is written to be compiled with GCC. Proprietary compilers for Linux include the Intel C++ Compiler, Sun Studio, and IBM XL C/C++ Compiler.

    +

    Most distributions also include support for PHP, Perl, Ruby, Python and other dynamic languages. Examples of languages that are less common, but still supported, are C# via the Mono project, sponsored by Novell, and Scheme. A number of Java Virtual Machines and development kits run on Linux, including the original Sun Microsystems JVM (HotSpot), and IBM's J2SE RE, as well as many open-source projects like Kaffe.

    +

    The two main frameworks for developing graphical applications are those of GNOME and KDE. These projects are based on the GTK+ and Qt widget toolkits, respectively, which can also be used independently of the larger framework. Both support a wide variety of languages. There are a number of Integrated Development Environments (IDEs) available including Anjuta, Code::Blocks, Eclipse, KDevelop, Lazarus, MonoDevelop, NetBeans, and Omnis Studio while the long-established editors Vim and Emacs remain popular.

    +
    +
    +
    +

    +1.3. Reasons for Using Linux

    +
    • Linux is free:

    +

    As in "free beer". Linux can be downloaded in its entirety from the Internet completely for free. No registration fees, no costs per user, free updates, and freely available source code in case you want to change the behavior of your system. +Most of all, Linux is free as in "free speech": +The license commonly used is the GNU Public License (GPL). The license says that anybody who may want to do so, has the right to change Linux and eventually to redistribute a changed version, on the one condition that the code is still available after redistribution. In practice, you are free to grab a kernel image and sell the new code, as long as your customers can still have a copy of that code.

    +
    • Linux is portable to any hardware platform:

    +

    A vendor, who wants to sell a new type of computer and who does not know what kind of OS his/her new machine will run, can take a Linux kernel and make it work on his/her hardware, because documentation related to this activity is freely available.

    +
    • Linux was made to keep on running:

    +

    As with UNIX, a Linux system expects to run without rebooting all the time. That is why a lot of tasks are being executed at night or scheduled automatically for other times, resulting in higher availability during busier periods and a more balanced use of the hardware. This property allows for Linux to be applicable to environments where people do not have the time or the possibility to control their systems constantly.

    +
    • Linux is secure and versatile:

    +

    The security model used in Linux is based on the UNIX idea of security, which is known to be robust and of proven quality. But Linux is not only safe from attacks from the Internet: it will adapt equally to other situations, utilizing the same high standards for security.

    +
    • Linux is scalable:

    +

    From a Palmtop with 2 MB of memory to a petabyte storage cluster with hundreds of nodes: add or remove the appropriate packages and Linux fits all. One does not need a supercomputer anymore,because you can use Linux to do big things using the building blocks provided with the system. If one wants to do little things, such as making an operating system for an embedded processor or just recycling your old 486, Linux will do that as well.

    +
    • The Linux OS and Linux applications have very short debug−times:

    +

    Because Linux has been developed and tested by thousands of people, both errors and people to fix them are found very quickly. It often happens that there are only a couple of hours between discovery and fixing of a bug.

    +
    +
    +
    +

    +2. Getting Started

    +
    +

    +2.1. Logging in, activating the user interface and logging out

    +

    In order to work on a Linux system directly, one needs to provide a user name and password. You always need to authenticate to the system. Most PC−based Linux systems have two basic modes for a system to run in: either quick and clean in text console mode,which includes with mouse, multitasking and multi−user features, or in graphical console mode, which looks better but eats more system resources.

    +
    +

    +2.1.1. Graphical Mode

    +

    This is the default nowadays on most desktop computers. You know you will be connecting to the system using graphical mode when you are first asked for your user name, and then to type your password.

    +

    To log in, make sure the mouse pointer is in the login window, provide your user name and password to the system and click OK or press Enter. +It is generally considered a bad idea to connect (graphically) using the root user name, the system adminstrator's account, since the use of graphics includes running a lot of extra programs, in root's case with a lot of extra permissions. To keep all risks as low as possible, use a normal user account to connect graphically. But there are enough risks to keep this in mind as a general advice, for all use of the root account: only log in as root when extra privileges are required.

    +

    After entering your user name/password combination, it can take a little while before the graphical environment is started, depending on the CPU speed of your computer, on the software you use and on your personal settings.

    +

    To continue, you will need to open a terminal window or xterm for short (X being the name for the underlying software supporting the graphical environment). This program can be found in the Applications−>Utilities->System Tools or Internet menu, depending on what window manager you are using. There might be icons that you can use as a shortcut to get an xterm window as well, and clicking the right mouse button on the desktop background will usually present you with a menu containing a terminal window application.

    +

    While browsing the menus, you will notice that a lot of things can be done without entering commands via the keyboard. For most users, the good old point−n−click method of dealing with the computer will do. But for those who want to enter the "heart" of the system, a tool stronger than a mouse will be required to handle the various tasks. This tool is the shell, and when in graphical mode, we activate our shell by opening a terminal window.

    +

    A terminal window should always show a command prompt when you open one. This terminal shows a standard prompt, which displays the user's login name, and the current working directory, represented by the twiddle (~)

    +

    Another common form for a prompt is this one: +[

    +
    +

    user@host
    +

    dir]

    +

    In the above example, user will be your login name, hosts the name of the machine you are working on, and dir an indication of your current location in the file system. Prompts can display all kinds of information, but they are not part of the commands you are giving to your system. To disconnect from the system in graphical mode, you need to close all terminal windows and other applications. After that, hit the logout icon or find Log Out in the menu. Closing everything is not really necessary, and the system can do this for you, but session management might put all currently open applications back on your screen when you connect again, which takes longer and is not always the desired effect. However, this behavior is configurable.

    +

    When you see the login screen again, asking to enter user name and password, logout was successful.

    +
    +
    +

    +2.1.2. Text Mode

    +

    One is in text mode when the whole screen is black, showing (in most cases white) characters. A text mode login screen typically shows some information about the machine you are working on, the name of the machine and a prompt waiting for you to log in.

    +

    The login is different from a graphical login, in that you have to hit the Enter key after providing your user name, because there are no buttons on the screen that you can click with the mouse. Then you should type your password, followed by another Enter. You will not see any indication that you are entering something, not even an asterisk, and you won't see the cursor move. But this is normal on Linux and is done for security +reasons.

    +

    When the system has accepted you as a valid user, you may get some more information, called the message of the day, which can be anything. Additionally, it is popular on UNIX systems to display a fortune cookie, which contains some general wise or unwise (this is up to you) thoughts. After that, you will be given a shell, indicated with the same prompt that you would get in graphical mode.

    +

    Also in text mode: log in as root only to do setup and configuration that absolutely requires administrator privileges, such as adding users, installing software packages, and performing network and other system configuration. Once you are finished, immediately leave the special account and resume your work as a non−privileged user.

    +

    Logging out is done by entering the logout command, followed by Enter. You are successfully disconnected from the system when you see the login screen again.Don't power−off the computer after logging out. It is not meant to be shut off without application of the proper procedures for halting the system. Powering it off without going through the halting process might cause severe damage!

    +
    +
    +
    +
    +

    +3. Basic Commands

    +
    +

    +3.1. ls

    +

    When invoked without any arguments, ls lists the files in the current working directory. A directory that is not the current working directory can be specified and ls will list the files there. The user also may specify any list of files and directories. In this case, all files and all contents of specified directories will be listed. The name ls is derived from list segments which was used in earlier systems.

    +

    Files whose names start with "." are not listed, unless the -a flag is specified or the files are specified explicitly.

    +

    Without options, ls displays files in a bare format. This bare format however makes it difficult to establish the type, permissions, and size of the files. The most common options to reveal this information or change the list of files are:

    +<block_quote>
      +
    • -l long format, displaying Unix file types, permissions, number of hard links, owner, group, size, date, and filename

    • +
    • -F appends a character revealing the nature of a file, for example, * for an executable, or / for a directory. Regular files have no suffix.

    • +
    • -a lists all files in the given directory, including those whose names start with "." (which are hidden files in Unix). By default, these files are excluded from the list.

    • +
    • -R recursively lists subdirectories. The command ls -R / would therefore list all files.

    • +
    • -d shows information about a symbolic link or directory, rather than about the link's target or listing the contents of a directory.

    • +
    • -t sort the list of files by modification time.

    • +
    • -h print sizes in human readable format. (e.g., 1K, 234M, 2G, etc.)

    • +
    </block_quote>

    In some environments, providing the option --color (for GNU ls) or -G (FreeBSD ls) causes ls to highlight different types of files with different colors, instead of with characters as -F would. To determine what color to use for a file, GNU ls checks the Unix file type, the file permissions, and the file extension, while FreeBSD ls checks only the Unix file type and file permissions.:

    +
    $ ls
    +jeeves.rst psmith.html blandings.html
    +$ ls -l
    +drwxr--r--   1 plum  editors   4096  jeeves
    +-rw-r--r--   1 plum  editors  30405  psmith
    +-r-xr-xr-x   1 plum  plum      8460  blandings
    +

    Here "$" actually is the beginning of the prompt. This is typical in most Unix-based systems.

    +
    +
    +

    +3.2. date

    +

    The Unix date command displays the time and date. The super-user can use it to set the system clock.

    +

    With no options, the date command displays the current date and time, including the abbreviated day name, abbreviated month name, day of the month, the time separated by colons, the timezone name, and the year. For example:

    +
    $date
    +Tue Sep  8 12:01:45 IST 2009
    +

    On some systems to set the current date and time to September 8, 2004 01:22 you type:

    +
    $date --set="20040908 01:22"
    +

    In order to view the various options for the date command, type:

    +
    $man date
    +

    This will take you to the "Manual" page comprising of all the details on the date command. You can return to the terminal from the "man" page by pressing the Esc key in the keyboard and typing ":q" in that order.

    +
    +
    +

    +3.3. cd

    +

    This stands for "change directory". When one wants to go up to the parent directory, bypassing the tree of directories one has entered, “ cd ..” can be used.

    +

    One dot '.' represents the current directory while two dots '..' represent the parent directory.

    +

    “ cd -” will return you to the previous directory (a bit like an “undo”).

    +

    You can also use cd absolute path or cd relative path (see below):

    +

    Absolute paths:

    +<block_quote>

    An “ absolute path” is easily recognised from the leading forward slash, /. The / means that you start at the top level directory and continue down.

    </block_quote>

    For example to get to /boot/grub you would type:

    +
    $cd /boot/grub
    +

    This is an absolute path because you start at the top of the hierarchy and go downwards from there (it doesn't matter where in the filesystem you were when you typed the command).

    +

    Relative paths:

    +<block_quote>

    A “ relative path” doesn't have a preceding slash. Use a relative path when you start from a directory below the top level directory structure. This is dependent on where you are in the filesystem.

    +

    For example if you are in root's home directory and want to get to /root/music, you type:

    +
    $ cd music
    </block_quote>

    Please note that there is no / using the above cd command. Using a / would cause this to be an absolute path, working from the top of the hierarchy downward.

    +
    +
    +

    +3.4. who

    +

    The standard Unix command who displays a list of users who are currently logged into a computer.

    +

    The who command is related to the command w, which provides the same information but also displays additional data and statistics.:

    +
    $who
    +beeblebrox tty7         2009-09-08 10:50 (:0)
    +beeblebrox pts/0        2009-09-08 11:25 (:0.0)
    +dumbledore pts/1        2009-09-08 18:11 (potter.xyz.in)
    +beeblebrox pts/2        2009-09-08 18:53 (:0.0)
    +

    The command can be invoked with the arguments am i or am I (so it is invoked as who am i or * who am I*), showing information about the current terminal only (see the -m option below, of which this invocation is equivalent).

    +

    In order to find out the various options that can be appended to the who command, check the man page by typing out the following in the terminal:

    +
    $man who
    +

    This will take you to the "Manual" page containing details about the who command

    +
    +
    +

    +3.5. mkdir

    +

    This command is used to make a new directory. Normal usage is as straightforward as follows:

    +
    $mkdir name_of_directory
    +

    Where name_of_directory is the name of the directory one wants to create. When typed as above (ie. normal usage), the new directory would be created within the current directory. On Unix, multiple directories can be specified, and mkdir will try to create all of them.

    +
    +

    +3.5.1. Options

    +

    On Unix-like operating systems, mkdir takes options. Three of the most common options are:

    +<block_quote>
      +
    • -p: will also create all directories leading up to the given directory that do not exist already. If the given directory already exists, ignore the error.

    • +
    • -v: display each directory that mkdir creates. Most often used with -p.

    • +
    • -m: specify the octal permissions of directories created by mkdir.

    • +
    </block_quote>

    -p is most often used when using mkdir to build up complex directory hierarchies, in case a necessary directory is missing or already there. -m is commonly used to lock down temporary directories used by shell scripts.

    +
    +
    +

    +3.5.2. Examples

    +

    An example of -p in action is:

    +
    $mkdir -p /tmp/a/b/c
    +

    If /tmp/a exists but /tmp/a/b does not, mkdir will create /tmp/a/b before creating /tmp/a/b/c.

    +

    And an even more powerful command, creating a full tree at once (this however is a Shell extension, nothing mkdir does itself):

    +
    $mkdir -p tmpdir/{trunk/sources/{includes,docs},branches,tags}
    +

    This will create:

    +<definition_list><definition_list_item><term>tmpdir - branches</term><definition>
      +
    • tag

    • +
    • <definition_list><definition_list_item><term>trunk - sources - includes</term><definition>
      • docs

      </definition>
      </definition_list_item>
      </definition_list>
    • +
    </definition>
    </definition_list_item>
    </definition_list>
    +
    +
    +
    +
    +

    +4. Getting Help

    +
    +

    +4.1. apropos and whatis

    +

    This is a command to search the manual pages files in Unix and Unix-like operating systems.

    +
    $ apropos grep
    +egrep       egrep (1)       Search a file for a pattern using full regular expressions
    +fgrep       fgrep (1)       Search a file for a fixed-character string
    +fmlgrep     fmlgrep (1)     Search a file for a pattern
    +grep        grep (1)        Search a file for a pattern
    +gzgrep      gzgrep (1)      Search a possibly compressed file for a regular expression
    +nisgrep     nismatch (1)    Utilities for searching NIS+ tables
    +pgrep       pgrep (1)       Find or signal a process by name or other attribute
    +zgrep       zgrep (1)       Search a possibly compressed file for a regular expression
    +...
    +

    In this example, the user uses apropos to search for the string "grep", and apropos returns the indicated man pages that include the term "grep".

    +

    A short index of explanations for commands is available using the whatis command, like in the examples below:

    +
    $whatis ls
    +ls (1)           - list directory contents
    +

    This displays short information about a command, and the first section in the collection of man pages that contains an appropriate page.

    +

    If you don't know where to get started and which man page to read, apropos gives more information. Say that you do not know how to start a browser, then you could enter the following command:

    +
    $apropos browser
    +gmusicbrowser (1)    - Jukebox for large collections of audio files
    +infobrowser (1)      - read Info documents
    +libsmbclient (7)     - An extension library for browsers and that               can be used...
    +opera (1)            - a standards-compliant graphical Web browser
    +sensible-browser (1) - sensible editing, paging, and web browsing
    +smbtree (1)          - A text based smb network browser
    +tvtk_doc (1)         - A GUI based TVTK documentation search browser.
    +viewres (1)          - graphical class browser for Xt
    +w3m (1)              - a text based Web browser and pager
    +www-browser (1)      - a text based Web browser and pager
    +...
    +
    +
    +

    +4.2. man

    +

    Man pages (short for "manual pages") are the extensive documentation that comes preinstalled with almost all substantial Unix and Unix-like operating systems. The Unix command used to display them is man. Each page is a self-contained document.

    +

    To read a manual page for a Unix command, one can use:

    +
    $ man <command_name>
    +

    at a shell prompt; for example, "man ftp". In order to simplify navigation through the output, man generally uses the less terminal pager.

    +

    Pages are traditionally referred to using the notation "name(section)"; for example, ftp(1). The same page name may appear in more than one section of the manual, this can occur when the names of system calls, user commands, or macro packages coincide. Two examples are man(1) and man(7), or exit(2) and exit(3). The syntax for accessing the non-default manual section varies between different man implementations. On Linux and BSD, for example, the syntax for reading *printf(3) is:

    +
    $man 3 printf
    +

    Another example:

    +
    $man man
    +

    The previous example will take you to the "Manual" page entry about manual pages!

    +
    +

    +4.2.1. Layout

    +

    All man pages follow a common layout that is optimized for presentation on a simple ASCII text display, possibly without any form of highlighting or font control. Sections present may include:

    +<definition_list><definition_list_item><term>NAME</term><definition>

    The name of the command or function, followed by a one-line description of what it does.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SYNOPSIS</term><definition>

    In the case of a command, you get a formal description of how to run it and what command line options it takes. For program functions, a list of the parameters the function takes and which header file contains its definition. For experienced users, this may be all the documentation they need.

    </definition>
    </definition_list_item>
    <definition_list_item><term>DESCRIPTION</term><definition>

    A textual description of the functioning of the command or function.

    </definition>
    </definition_list_item>
    <definition_list_item><term>EXAMPLES</term><definition>

    Some examples of common usage.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SEE ALSO</term><definition>

    A list of related commands or functions.

    </definition>
    </definition_list_item>
    </definition_list>

    Other sections may be present, but these are not well standardized across man pages. Common examples include: OPTIONS, EXIT STATUS, ENVIRONMENT, KNOWN BUGS, FILES, AUTHOR, REPORTING BUGS, HISTORY and COPYRIGHT.

    +

    These days virtually every Unix command line application comes with its man page, and many Unix users perceive a lack of man pages as a sign of low quality; indeed, some projects, such as Debian, go out of their way to write man pages for programs lacking one. Few alternatives to man have enjoyed much popularity, with the possible exception of the GNU project's "info" system, an early and simple hypertext system.

    +

    However, the format of a single page for each application, the lack of classification within the sections and the relatively unsophisticated formatting facilities have motivated the development of alternative documentation systems, such as the previously mentioned "info" system.

    +

    Most Unix GUI applications (particularly those built using the GNOME and KDE development environments) now provide end-user documentation in HTML and include embedded HTML viewers such as yelp for reading the help within the application.

    +

    Usually the man pages are written in English. Translations into other languages can be also available on the system.

    +

    The default format of the man pages is troff, with either the macro package man (appearance oriented) or on some systems mdoc (semantic oriented). This makes it possible to typeset a man page to PostScript, PDF and various other formats for viewing or printing.

    +
    +
    +
    +

    +4.3. info

    +

    info is a software utility which forms a hypertextual, multipage documentation and help viewer working on a command line interface, useful when there is no GUI available.

    +

    The syntax is

    +
    $ info <command_name>
    +

    info processes info files, which are Texinfo formatted files, and presents the documentation as a tree, with simple commands to traverse the tree and to follow cross references. For instance

    +<block_quote>
      +
    • n goes to the next page.

    • +
    • p goes to the previous page.

    • +
    • u goes to the upper page.

    • +
    • l goes to the last(visited) node

    • +
    • To follow a cross reference, the cursor can be moved over a link (a word preceded by a <title_reference>*</title_reference>) and enter pressed.

    • +
    </block_quote>

    info was initially written for use with GNU/Linux and then ported to other Unix-like operating systems.

    +
    +
    +

    +4.4. --help

    +

    Most GNU commands support the --help, which gives a short explanation about how to use the command and a list of available options. Below is the output of this option with the cat command:

    +
    $ userprompt@host: cat --help
    +Usage: cat [OPTION] [FILE]...
    +Concatenate FILE(s), or standard input, to standard output.
    +
    +  -A, --show-all           equivalent to -vET
    +  -b, --number-nonblank    number nonempty output lines
    +  -e                       equivalent to -vE
    +  -E, --show-ends          display $ at end of each line
    +  -n, --number             number all output lines
    +  -s, --squeeze-blank      suppress repeated empty output lines
    +  -t                       equivalent to -vT
    +  -T, --show-tabs          display TAB characters as ^I
    +  -u                       (ignored)
    +  -v, --show-nonprinting   use ^ and M- notation, except for LFD and              TAB
    +  --help     display this help and exit
    +  --version  output version information and exit
    +
    +With no FILE, or when FILE is -, read standard input.
    +
    +Examples:
    +  cat f - g  Output f's contents, then standard input, then g's           contents.
    +  cat        Copy standard input to standard output.
    +
    +Report bugs to <bug-coreutils@gnu.org>.
    +
    +
    +
    +

    +5. Basic file handling

    +
    +

    +5.1. cp

    +

    cp is the command entered in a Unix shell to copy a file from one place to another, possibly on a different filesystem. The original file remains unchanged, and the new file may have the same or a different name.

    +
    +

    +5.1.1. Usage

    +

    To copy a file to another file:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ][ -- ] SourceFile TargetFile
    +

    To copy a file to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -r | -R ] [ -- ] SourceFile ...              TargetDirectory
    +

    To copy a directory to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -- ] { -r | -R }
    +SourceDirectory ... TargetDirectory
    +
    +
    +

    +5.1.2. Flags

    +

    -f (force) – specifies removal of the target file if it cannot be opened for write operations. The removal precedes any copying performed by the cp command.

    +

    -P – makes the cp command copy symbolic links. The default is to follow symbolic links, that is, to copy files to which symbolic links point.

    +

    -i (interactive) – prompts you with the name of a file to be overwritten. This occurs if the TargetDirectory or TargetFile parameter contains a file with the same name as a file specified in the SourceFile or SourceDirectory parameter. If you enter y or the locale's equivalent of y, the cp command continues. Any other answer prevents the cp command from overwriting the file.

    +

    -p (preserve) – duplicates the following characteristics of each SourceFile/SourceDirectory in the corresponding TargetFile and/or TargetDirectory:

    +<block_quote>
      +
    • The time of the last data modification and the time of the last access.

    • +
    • The user ID and group ID (only if it has permissions to do this)

    • +
    • The file permission bits and the SUID and SGID bits.

    • +
    </block_quote>

    -R (recursive) – copy directories (recursively copying all the contents)

    +
    +
    +

    +5.1.3. Examples

    +

    To make a copy of a file in the current directory, enter:

    +
    $ cp prog.c prog.bak
    +

    This copies prog.c to prog.bak. If the prog.bak file does not already exist, the cp command creates it. If it does exist, the cp command replaces it with a copy of the prog.c file.

    +

    To copy a file in your current directory into another directory, enter:

    +
    $ cp zaphod /home/books/hhgg
    +

    This copies the jones file to /home/books/hhgg/zaphod.

    +

    To copy a file to a new file and preserve the modification date, time, and access control list associated with the source file, enter:

    +
    $ cp -p martin_luther_king martin_luther_king.jr
    +

    This copies the martin_luther_king file to the martin_luther_king.jr file. Instead of creating the file with the current date and time stamp, the system gives the martin_luther_king.jr file the same date and time as the martin_luther_king file. The martin_luther_king.jr file also inherits the martin_luther_king file's access control protection.

    +

    To copy all the files in a directory to a new directory, enter:

    +
    $ cp /home/galactica/clients/* /home/hhgg/customers
    +

    This copies only the files in the clients directory to the customers directory.

    +

    To copy a directory, including all its files and subdirectories, to another directory, enter:

    +<block_quote>

    $ cp -R /home/hhgg/clients /home/hhgg/customers

    </block_quote>

    This copies the clients directory, including all its files, subdirectories, and the files in those subdirectories, to the customers/clients directory.

    +

    To copy a specific set of files of any extension to another directory, enter:

    +
    $ cp zaphod arthur ford /home/hhgg/clients
    +

    This copies the zaphod, arthur, and ford files in your current working directory to the /home/hhgg/clients directory.

    +

    To use pattern-matching characters to copy files, enter:

    +
    $ cp programs/*.py .
    +

    This copies the files in the programs directory that end with .py to the current directory, signified by the single "." (dot). You must type a space between the py and the final dot.

    +
    +
    +
    +

    +5.2. mv

    +

    mv (short for move) is a Unix command that moves one or more files or directories from one place to another. The original file is deleted, and the new file may have the same or a different name. If possible (i.e. when the original and new files are on the same file system), mv will rename the file instead. Write permission is required on all directories being modified.

    +
    +

    +5.2.1. Conflicting existing file

    +

    In all cases, when a file is moved to have the name of an existing file (in the same directory), the existing file is deleted. If the existing file is not writable but is in a directory that is writable, then the mv command asks for confirmation if possible (i.e. if run from a terminal) before proceeding, unless the -f (force) option is used.

    +
    +
    +

    +5.2.2. Differences with copy and delete

    +

    Note that, usually, when moving files within the same volume, moving (and/or renaming) is not the same as simply copying and then deleting the original. When moving a file, the link is simply removed from the old parent directory and added to the new parent directory. However, the file itself is untouched (i.e. it has the same inodes and resides at the same place on the disk). For example, you cannot copy a file you cannot read, but you can move (and/or rename) it (provided you have write permission to its old and new parent directories). Also, suppose there is a non-empty directory you do not have write permission to. You cannot delete this directory (since you cannot delete its contents); but you can move (and/or rename) it. Also, since moving between filenames on a single volume does not involve copying, it is faster and does not place strain of lots of reads and writes on the disk. Moving files across different volumes, however, does necessitate copying and deleting.

    +
    +
    +

    +5.2.3. Examples

    +
    $ mv myfile mynewfilename    renames a file
    +$ mv myfile otherfilename    renames a file and deletes the existing            file "myfile"
    +$ mv myfile /myfile          moves 'myfile' from the current            directory to the root directory
    +$ mv myfile dir/myfile       moves 'myfile' to 'dir/myfile' relative            to the current directory
    +$ mv myfile dir              same as the previous command (the          filename is implied to be the same)
    +$ mv myfile dir/myfile2      moves 'myfile' to dir and renames it to            'myfile2'
    +$ mv foo bar baz dir         moves multiple files to directory dir
    +$ mv --help                  shows a very concise help about the                syntax of the command
    +$ man mv                     prints an extensive user manual for                'mv' in the terminal
    +

    In all cases, the file or files being moved or renamed can be a directory.

    +

    Note that when the command is called with two arguments (as mv name1 name2 or mv name1 /dir/name2), it can have three different effects, depending on whether name2 does not exist, is an existing file, or is an existing directory. If the user intends to refer to an existing directory, /. (or in some Unix versions / is sufficient) may be appended to the name to force the system to check this. To move a file to a new directory, the directory must be created first.

    +
    +
    +
    +

    +5.3. rm

    +

    rm (short for "remove") is one of several basic Unix command lines that operates on files. It is used to delete files from a filesystem. The data is not actually destroyed. Only the index listing where the file is stored is destroyed, and the storage is made available for reuse. There are undelete utilities that will attempt to reconstruct the index and can bring the file back if the parts were not reused.

    +

    Here's example to remove a file named "foo" from a directory, here shown with the -i option:

    +
    $ rm -i foo
    +remove foo? y
    +
    +

    +5.3.1. Options

    +

    Common options that rm accepts include:

    +<block_quote>
      +
    • -r, which removes directories, removing the contents recursively beforehand (so as not to leave files without a directory to reside in) ("recursive")

    • +
    • -i, which asks for every deletion to be confirmed ("interactive")

    • +
    • -f, which ignores non-existent files and overrides any confirmation prompts ("force")

    • +
    • -v, which shows what is being removed as it happens ("verbose")

    • +
    </block_quote>

    rm is often aliased to "rm -i" so as to avoid accidental deletion of files. If a user still wishes to delete a large number of files without confirmation, they can manually cancel out the -i argument by adding the -f option (as the option specified later on the expanded command line "rm -i -f" takes precedence).

    +

    rm -rf (variously, rm -rf /, rm -rf <title_reference>*</title_reference>, and others) is frequently used in jokes and anecdotes about Unix disasters. The rm -rf variant of the command, if run by a superuser on the root directory, would cause the contents of every writable mounted filesystem on the computer to be deleted.

    +

    rm is often used in conjunction with xargs to supply a list of files to delete:

    +
    xargs rm < filelist
    +

    When rm is used on a symbolic link, it deletes the link, but does not affect the target of the link.

    +
    +
    +

    +5.3.2. Permissions

    +

    Usually, on most filesystems, deleting a file requires write permission on the parent directory (and execute permission, in order to enter the directory in the first place). (Note that, confusingly for beginners, permissions on the file itself are irrelevant. However, GNU rm asks for confirmation if a write-protected file is to be deleted, unless the -f option is used.)

    +

    To delete a directory (with rm -r), one must delete all of its contents recursively. This requires that one must have read and write and execute permission to that directory (if it's not empty) and all non-empty subdirectories recursively (if there are any). The read permissions are needed to list the contents of the directory in order to delete them. This sometimes leads to an odd situation where a non-empty directory cannot be deleted because one doesn't have write permission to it and so cannot delete its contents; but if the same directory were empty, one would be able to delete it.

    +

    If a file resides in a directory with the sticky bit set, then deleting the file requires one to be the owner of the file.

    +
    +
    +
    +
    +

    +6. Command Line Arguments

    +

    In computer command line interfaces, a command line argument is an argument sent to a program being called. In general, a program can take any number of command line arguments, which may be necessary for the program to run, or may even be ignored, depending on the function of that program.

    +

    For example, in Unix and Unix-like environments, an example of a command-line argument is:

    +
    rm file.s
    +

    "file.s" is a command line argument which tells the program rm to remove the file "file.s".

    +

    Programming languages such as C, C++ and Java allow a program to interpret the command line arguments by handling them as string parameters in the main function.

    +

    A command line option or simply option (also known as a command line parameter, flag, or a switch) is an indication by a user that a computer program should change its default output.

    +

    Long options are introduced via "--", and are typically whole words. For example, ls --long --classify --all. Arguments to long options are provided with "=", as ls --block-size=1024. Some Unix programs use long options with single dashes, for example MPlayer as in mplayer -nosound.

    +

    Linux also uses "--" to terminate option lists. For example, an attempt to delete a file called -file1 by using rm -file1 may produce an error, since rm may interpret -file1 as a command line switch. Using rm -- -file1 removes ambiguity.

    +
    +
    +

    +7. Basic Text Processing

    +
    +

    +7.1. head

    +

    head is a program on Unix and Unix-like systems used to display the first few lines of a text file or piped data. The command syntax is:

    +
    $ head [options] <file_name>
    +

    By default, head will print the first 10 lines of its input to the standard output. The number of lines printed may be changed with a command line option. The following example shows the first 20 lines of filename:

    +
    $ head -n 20 filename
    +

    This displays the first 5 lines of all files starting with foo:

    +
    $ head -n 5 foo*
    +

    Some versions omit the n and just let you say -5.

    +
    +

    +7.1.1. Flags

    +
    -c <x number of bytes> Copy first x number of bytes.
    +

    Other options: sed

    +

    Many early versions of Unix did not have this command, and so documentation and books had sed do this job:

    +
    sed 5q foo
    +

    This says to print every line (implicit), and quit after the fifth.

    +
    +
    +
    +

    +7.2. tail

    +

    tail is a program on Unix and Unix-like systems used to display the last few lines of a text file or piped data.

    +

    The command-syntax is:

    +
    $ tail [options] <file_name>
    +

    By default, tail will print the last 10 lines of its input to the standard output. With command line options the number of lines printed and the printing units (lines, blocks or bytes) may be changed. The following example shows the last 20 lines of filename:

    +
    $ tail -n 20 filename
    +

    This example shows the last 15 bytes of all files starting with foo:

    +
    $ tail -c 15 foo*
    +

    This example shows all lines of filename from the second line onwards:

    +
    $ tail -n +2 filename
    +

    Using an older syntax (still used in Sun Solaris as the -n option is not supported), the last 20 lines and the last 50 bytes of filename can be shown with the following command:

    +
    $ tail -20 filename
    +$ tail -50c filename
    +

    However this syntax is now obsolete and does not conform with the POSIX 1003.1-2001 standard. Even if still supported in current versions, when used with other options (like -f, see below), these switches could not work at all.

    +
    +

    +7.2.1. File monitoring

    +

    tail has a special command line option -f (follow) that allows a file to be monitored. Instead of displaying the last few lines and exiting, tail displays the lines and then monitors the file. As new lines are added to the file by another process, tail updates the display. This is particularly useful for monitoring log files. The following command will display the last 10 lines of messages and append new lines to the display as new lines are added to messages:

    +
    $ tail -f /var/adm/messages
    +

    To interrupt tail while it is monitoring, break-in with Ctrl+C. This command can be run "in the background" with &, see job control.

    +

    If you have a command's result to monitor, you can use the watch command.

    +
    +
    +
    +

    +7.3. cut

    +

    In computing, cut is a Unix command line utility which is used to extract sections from each line of input — usually from a file.

    +

    Extraction of line segments can typically be done by bytes (-b), characters (-c), or fields (-f) separated by a delimiter (-d — the tab character by default). A range must be provided in each case which consists of one of N, N-M, N- (N to the end of the line), or -M (beginning of the line to M), where N and M are counted from 1 (there is no zeroth value). Since version 6, an error is thrown if you include a zeroth value. Prior to this the value was ignored and assumed to be 1.

    +

    Assuming a file named file containing the lines:

    +
    foo:bar:baz:qux:quux
    +one:two:three:four:five:six:seven
    +alpha:beta:gamma:delta:epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +

    To output the fourth through tenth characters of each line:

    +
    $ cut -c 4-10 file
    +

    This gives the output:

    +
    :bar:ba
    +:two:th
    +ha:beta
    +

    To output the fifth field through the end of the line of each line using the colon character as the field delimiter:

    +
    $ cut -d : -f 5- file
    +

    This gives the output:

    +
    quux
    +five:six:seven
    +epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +
    +
    +

    +7.4. paste

    +

    paste is a Unix command line utility which is used to join files horizontally (parallel merging) by outputting lines consisting of the sequentially corresponding lines of each file specified, separated by tabs, to the standard output. It is effectively the horizontal equivalent to the utility cat command which operates on the vertical plane of two or more files.

    +

    To paste several columns of data together into the file www from files who, where, and when:

    +
    $ paste who where when > www
    +

    If the files contain:

    +++++
    +

    This creates the file named www containing:

    +
    Batman            GothamCity       January 3
    +Trillian          Andromeda        February 4
    +Jeeves            London           March 19
    +
    +
    +
    +

    +8. Shell Meta Characters

    +

    Unix recognizes certain special characters, called "meta characters," as command directives. The shell meta characters are recognized anywhere they appear in the command line, even if they are not surrounded by blank space. For that reason, it is safest to only use the characters A-Z, a-z, 0-9, and the period, dash, and underscore characters when naming files and directories on Unix. If your file or directory has a shell meta character in the name, you will find it difficult to use the name in a shell command.

    +

    The shell meta characters include:

    +

    / < > ! $ % ^ & * | { } [ ] " ' ` ~ ;

    +

    Different shells may differ in the meta characters recognized.

    +

    As an example,

    +
    $ ls file.*
    +

    run on a directory containing the files file, file.c, file.lst, and myfile would list the files file.c and file.lst. However,:

    +
    $ ls file.?
    +

    run on the same directory would only list file.c because the ? only matches one character, no more, no less. This can save you a great deal of typing time. For example, if there is a file called california_cornish_hens_with_wild_rice and no other files whose names begin with 'c', you could view the file without typing the whole name by typing this:

    +
    $ more c*
    +

    because the c* matches that long file name.

    +

    Filenames containing metacharacters can pose many problems and should never be intentionally created. If you do find that you've created a file with metacharacters, and you would like to remove it, you have three options. You may use wildcards to match metacharacter, use the to directly enter the filename, or put the command in double quotes (except in the case of double quotes within the file name, these must be captured with one of the first two methods). For example, deleting a file named <title_reference>"``*`|more</title_reference>"` can be accomplished with:

    +
    $ rm ??more
    +

    or:

    +
    $ rm $\backslash$*$\backslash$|more
    +

    or:

    +
    $ rm ''*|more''
    +
    +
    +

    +9. Looking At Files

    +
    +

    +9.1. cat

    +

    The cat command is a standard Unix program used to concatenate and display files. The name is from "catenate", a synonym of concatenate.

    +

    The Single Unix Specification specifies the behavior that the contents of each of the files given in sequence as arguments will be written to the standard output in the same sequence, and mandates one option, -u, where each byte is printed as it is read.

    +

    If the filename is specified as -, then cat will read from standard input at that point in the sequence. If no files are specified, cat will read from standard input entered.

    +
    +

    +9.1.1. Jargon File Definition

    +

    The Jargon File version 4.4.7 lists this as the definition of cat:

    +
    1. To spew an entire file to the screen or some other output sink without
    +     pause (syn. blast).
    +
    +2. By extension, to dump large amounts of data at an unprepared target or
    +     with no intention of browsing it carefully. Usage: considered silly.
    +     Rare outside Unix sites. See also dd, BLT.
    +
    +     Among Unix fans, *cat(1)* is considered an excellent example of
    +     user-interface design, because it delivers the file contents without
    +     such verbosity as spacing or headers between the files, and because
    +     it does not require the files to consist of lines of text, but works
    +     with any sort of data.
    +
    +     Among Unix critics, *cat(1)* is considered the canonical example of
    +     bad user-interface design, because of its woefully unobvious name.
    +     It is far more often used to blast a single file to standard output
    +     than to concatenate two or more files. The name cat for the former
    +     operation is just as unintuitive as, say, LISP's cdr.
    +
    +     Of such oppositions are holy wars made...
    +
    +
    +

    +9.1.2. Useless Use of 'cat'

    +

    UUOC (from comp.unix.shell on Usenet) stands for “Useless Use of cat”. As it is observed on comp.unix.shell, “The purpose of cat is to concatenate (or 'catenate') files. If it's only one file, concatenating it with nothing at all is a waste of time, and costs you a process.”

    +

    Nevertheless one sees people doing:

    +
    $ cat file | some_command and its args ...
    +

    instead of the equivalent and cheaper:

    +
    <file some_command and its args ...
    +

    or (equivalently and more classically):

    +
    some_command and its args ... <file
    +

    Since 1995, occasional awards for UUOC have been given out. The activity of fixing instances of UUOC is sometimes called 'demoggification'.

    +

    Amongst many, it is still considered safer to use cat for such cases given that the < and > keys are next to each other in many popular keyboard mappings. While the risk might be low, the impact of using > instead of < can be high and prohibitive.

    +
    +
    +

    +9.1.3. zcat

    +

    zcat is a Unix program similar to cat, that decompresses individual files and concatenates them to standard output. Traditionally zcat operated on files compressed by compress but today it is usually able to operate on gzip or even bzip2 archives. On such systems, it is equivalent to gunzip -c

    +
    +
    +
    +

    +9.2. more

    +

    In computing, more is a command to view (but not modify) the contents of a text file one screen at a time (terminal pager). It is available on Unix and Unix-like systems, DOS, OS/2 and Microsoft Windows. Programs of this sort are called pagers.

    +
    +

    +9.2.1. Usage

    +

    The command-syntax is:

    +
    $ more [options] [file_name]
    +

    If no file name is provided, more looks for input from stdin.

    +

    Once more has obtained input, it displays as much as can fit on the current screen and waits for user input to advance, with the exception that a form feed (^L) will also cause more to wait at that line, regardless of the amount of text on the screen. In the lower-left corner of the screen is displayed the text "--More--" and a percentage, representing the percent of the file that more has paged through. (This percentage includes the text displayed on the current screen.) When more reaches the end of a file (100%) it exits. The most common methods of navigating through a file are Enter, which advances the output by one line, and Space, which advances the output by one screen.

    +

    There are also other commands that can be used while navigating through the document; consult more's man page for more details.

    +

    Options are typically entered before the file name, but can also be entered in the environment variable $MORE. Options entered in the actual command line will override those entered in the $MORE environment variable. Available options may vary between Unix systems.

    +
    +
    +
    +

    +9.3. less

    +

    less is a terminal pager program on Unix, Windows and Unix-like systems used to view (but not change) the contents of a text file one screen at a time. It is similar to more, but has the extended capability of allowing both forward and backward navigation through the file. Unlike most Unix text editors/viewers, less does not need to read the entire file before starting, resulting in faster load times with large files.

    +
    +

    +9.3.1. Usage

    +

    less can be invoked with options to change its behaviour, for example, the number of lines to display on the screen. A few options vary depending on the operating system. While less is displaying the file, various commands can be used to navigate through the file. These commands are based on those used by both more and vi. It is also possible to search for character patterns in the file.

    +

    By default, less displays the contents of the file to the standard output (one screen at a time). If the file name argument is omitted, it displays the contents from standard input (usually the output of another command through a pipe). If the output is redirected to anything other than a terminal, for example a pipe to another command, less behaves like cat.

    +

    The command-syntax is:

    +
    $ less [options] file_name
    +
    +
    +

    +9.3.2. Frequently Used Options

    +<block_quote>
      +
    • -g: Highlights just the current match of any searched string.

    • +
    • -I: Case-insensitive searches.

    • +
    • -M: Shows more detailed prompt, including file position.

    • +
    • -N: Shows line numbers (useful for source code viewing).

    • +
    • -S: Disables line wrap ("chop long lines"). Long lines can be seen by side scrolling.

    • +
    • -?: Shows help.

    • +
    </block_quote>
    +
    +
    +

    +9.3.3. Frequently Used Commands

    +<block_quote>
      +
    • [Arrows]/[Page Up]/[Page Down]/[Home]/[End]: Navigation.

    • +
    • [Space bar]: Next page.

    • +
    • b: Previous page.

    • +
    • ng: Jump to line number n. Default is the start of the file.

    • +
    • nG: Jump to line number n. Default is the end of the file.

    • +
    • /pattern: Search for pattern. Regular expressions can be used.

    • +
    • '^ or g: Go to start of file.

    • +
    • '$ or G: Go to end of file.

    • +
    • s: Save current content (got from another program like grep) in a file.

    • +
    • =: File information.

    • +
    • h: Help.

    • +
    • q: Quit.

    • +
    </block_quote>
    +
    +
    +

    +9.3.4. Examples

    +
    $ less -M readme.txt                     #Read "readme.txt."
    +$ less +F /var/log/mail.log              #Follow mode for log
    +$ file * | less                          #Easier file analysis.
    +$ grep -i void *.c | less -I -p void     #Case insensitive search                                                         for "void" in all .c files
    +
    +
    +
    +
    +

    +10. Directory Structure

    +

    In the File Hierarchy Standard (FHS) all files and directories appear under the root directory "/", even if they are stored on different physical devices. Note however that some of these directories may or may not be present on a Unix system depending on whether certain subsystems, such as the X Window System, are installed.

    +

    The majority of these directories exist in all UNIX operating systems and are generally used in much the same way; however, the descriptions here are those used specifically for the FHS, and are not considered authoritative for platforms other than Linux.

    ++++
    +
    +

    +10.1. man hier

    +

    This is the manual page on the UNIX filesystem. The syntax for this is:

    +
    $ man hier
    +
    +
    +

    +10.2. ls -l

    +

    Shows you huge amounts of information (permissions, owners, size, and when last modified) for folders and files. The syntax is

    +
    $ ls -l
    +

    This can be done after entering the required directory.

    +
    +
    +
    +

    +11. Permissions and Ownership

    +
    +

    +11.1. chmod

    +

    The chmod command (abbreviated from 'change mode') is a shell command and C language function in Unix and Unix-like environments. When executed, it can change file system modes of files and directories. The modes include permissions and special modes.A chmod command first appeared in AT&T Unix version 1, and is still used today on Unix-like machines.

    +
    +

    +11.1.1. Usage

    +

    The chmod command options are specified like this:

    +
    $ chmod [options] mode[,mode] file1 [file2 ...]
    +

    To view what the permissions currently are, type:

    +
    $ ls -l file
    +
    +
    +

    +11.1.2. Command line options

    +

    The chmod command has a number of command line options that affect its behavior. The most common options are:

    +<block_quote>
      +
    • -R: Changes the modes of directories and files recursively

    • +
    • -v: Verbose mode; lists all files as they are being processed

    • +
    </block_quote>
    +
    +11.1.2.1. Symbolic modes
    +

    To the chmod utility, all permissions and special modes are represented by its mode parameter. One way to adjust the mode of files or directories is to specify a symbolic mode. The symbolic mode is composed of three components, which are combined to form a single string of text:

    +
    $ chmod [references][operator][modes] file1 ...
    +

    The references (or classes) are used to distinguish the users to whom the permissions apply. If no references are specified it defaults to “all” but modifies only the permissions allowed by the umask. The references are represented by one or more of the following letters:

    +++++
    +

    The chmod program uses an operator to specify how the modes of a file should be adjusted. The following operators are accepted:

    ++++
    +

    The modes indicate which permissions are to be granted or taken away from the specified classes. There are three basic modes which correspond to the basic permissions:

    +++++
    +

    The combination of these three components produces a string that is understood by the chmod command. Multiple changes can be specified by separating multiple symbolic modes with commas.

    +
    +
    +
    +11.1.2.2. Symbolic examples
    +

    Add the 'read' and 'write' permissions to the 'user' and 'group' classes of a directory:

    +
    $ chmod ug+rw mydir
    +$ ls -ld mydir
    +drw-rw----   2 starwars  yoda  96 Dec 8 12:53 mydir
    +

    For a file, remove write permissions for all classes:

    +
    $ chmod a-w myfile
    +$ ls -l myfile
    +-r-xr-xr-x   2 starwars  yoda 96 Dec 8 12:53 myfile
    +

    Set the permissions for the u*ser and the *g*roup to read and execute only (no write permission) on *mydir.

    +
    $ chmod ug=rx mydir
    +$ ls -ld mydir
    +dr-xr-x---   2 starwars  yoda 96 Dec 8 12:53 mydir
    +
    +
    +
    +11.1.2.3. Octal numbers
    +

    The chmod command also accepts three and four-digit octal numbers representing modes. Using a three-digit octal number to set the modes of a file named myfile :

    +
    $ chmod 664 myfile
    +$ ls -l myfile
    +-rw-rw-r--  1   57 Jul  3 10:13  myfile
    +

    Since the setuid, setgid and sticky bits are not set, this is equivalent to:

    +
    $ chmod 0664 myfile
    +
    +
    +
    +11.1.2.4. Special modes
    +

    The chmod command is also capable of changing the additional permissions or special modes of a file or directory. The symbolic modes use s to represent the setuid and setgid modes, and t to represent the sticky mode. The modes are only applied to the appropriate classes, regardless of whether or not other classes are specified.

    +

    Most operating systems support the specification of special modes using octal modes, but some do not. On these systems, only the symbolic modes can be used.

    +
    +
    +
    +
    +
    +

    +12. Redirection and Piping

    +

    In computing, redirection is a function common to most command-line interpreters, including the various Unix shells that can redirect standard streams to user-specified locations.

    +

    Programs do redirection with the dup2(2) system call, or its less-flexible but higher-level stdio analogues, freopen(3) and popen(3).

    +
    +

    +12.1. Redirecting standard input and standard output

    +

    Redirection is usually implemented by placing certain characters between commands. Typically, the syntax of these characters is as follows:

    +
    $ command1 > file1
    +

    executes command1, placing the output in file1. Note that this will truncate any existing data in file1. To append output to the end of the file, use the >> operator.:

    +
    $ command1 < file1
    +

    executes command1, using file1 as the source of input (as opposed to the keyboard).:

    +
    $ command1 < infile > outfile
    +

    combines the two capabilities: command1 reads from infile and writes to outfile

    +
    +
    +

    +12.2. Piping

    +

    Programs can be run together such that one program reads the output from another with no need for an explicit intermediate file: +A pipeline of three programs run on a text terminal:

    +
    $ command1 | command2
    +

    executes command1, using its output as the input for command2 (commonly called piping, since the "|" character is known as a "pipe").

    +

    This is equivalent to using two redirects and a temporary file:

    +
    $ command1 > tempfile
    +$ command2 < tempfile
    +$ rm tempfile
    +

    A good example for command piping is combining echo with another command to achieve something interactive in a non-interactive shell, e.g.:

    +
    $ echo -e "user\npass" | ftp localhost
    +

    This runs the ftp client with input user, press return, then pass.

    +
    +
    +

    +12.3. Redirecting to and from the standard file handles

    +

    In Unix shells derived from the original Bourne shell, the first two actions can be further modified by placing a number (the file descriptor) immediately before the character; this will affect which stream is used for the redirection. The Unix standard I/O streams are:

    +++++
    +

    For example:

    +
    $ command1 2> file1
    +

    executes command1, directing the standard error stream to file1.

    +

    In shells derived from csh (the C shell), the syntax instead appends the & character to the redirect characters, thus achieving a similar result.

    +

    Another useful capability is to redirect one standard file handle to another. The most popular variation is to merge standard error into standard output so error messages can be processed together with (or alternately to) the usual output. Example:

    +
    $ find / -name .profile > results 2>&1
    +

    will try to find all files named .profile. Executed without redirection, it will output hits to stdout and errors (e.g. for lack of privilege to traverse protected directories) to stderr. If standard output is directed to file results, error messages appear on the console. To see both hits and error messages in file results, merge stderr (handle 2) into stdout (handle 1) using 2>&1 .

    +

    It's possible use 2>&1 before ">" but it doesn't work. In fact, when the interpreter reads 2>&1, it doesn't know yet where standard output is redirected and then standard error isn't merged.

    +

    If the merged output is to be piped into another program, the file merge sequence 2>&1 must precede the pipe symbol, thus:

    +
    $ find / -name .profile 2>&1 | less
    +

    A simplified form of the command:

    +
    $ command > file 2>&1
    +

    is:

    +
    $ command &>file
    +

    or:

    +
    $command >&file
    +
    +
    +

    +12.4. Chained pipelines

    +

    The redirection and piping tokens can be chained together to create complex commands. For example:

    +
    $ ls | grep '\.sh' | sort > shlist
    +

    lists the contents of the current directory, where this output is filtered to only contain lines which contain .sh, sort this resultant output lexicographically, and place the final output in shlist. This type of construction is used very commonly in shell scripts and batch files.

    +
    +
    +

    +12.5. Redirect to multiple outputs

    +

    The standard command tee can redirect output from a command to several destinations.

    +
    $ ls -lrt | tee xyz
    +

    This directs the file list output to both standard output as well as to the file xyz.

    +
    +
    +
    +

    +13. More Text Processing

    +
    +

    +13.1. grep

    +

    grep is a command line text search utility originally written for Unix. The name is taken from the first letters in global / regular expression / print, a series of instructions for the ed text editor. The grep command searches files or standard input globally for lines matching a given regular expression, and prints them to the program's standard output.

    +
    +

    +13.1.1. Usage

    +

    This is an example of a common grep usage:

    +
    $ grep apple fruitlist.txt
    +

    In this case, grep prints all lines containing 'apple' from the file fruitlist.txt, regardless of word boundaries; therefore lines containing 'pineapple' or 'apples' are also printed. The grep command is case sensitive by default, so this example's output does not include lines containing 'Apple' (with a capital A) unless they also contain 'apple'.

    +

    Like most Unix commands, grep accepts command line arguments to change this and many other behaviors. For example:

    +
    $ grep -i apple fruitlist.txt
    +

    This prints all lines containing 'apple' regardless of capitalization. The '-i' argument tells grep to be case insensitive, or to ignore case.

    +

    To print all lines containing 'apple' as a word ('pineapple' and 'apples' will not match):

    +
    $ grep -w apple fruitlist.txt
    +

    Regular expressions can be used to match more complicated queries.

    +
    +
    +13.1.1.1. Variations
    +

    There are countless implementations and derivatives of grep available for many operating systems. Early variants of grep included egrep and fgrep. The former applies an extended regular expression syntax that was added to Unix after Ken Thompson's original regular expression implementation. The latter searches for any of a list of 'fixed' strings using the Aho-Corasick algorithm. These variants are embodied in most modern grep implementations as command-line switches (and standardized as -E and -F in POSIX). In such combined implementations, grep may also behave differently depending on the name by which it is invoked, allowing fgrep, egrep, and grep to be links to the same program.

    +

    pcregrep is an implementation of grep that uses Perl regular expression syntax.

    +

    Other commands contain the word 'grep' to indicate that they search (usually for regular expression matches). The pgrep utility, for instance, displays the processes whose names match a given regular expression.

    +
    +
    +
    +
    +

    +13.2. tr

    +

    tr (abbreviated from translate or transliterate) is a command in Unix-like operating systems.

    +

    When executed, the program reads from the standard input and writes to the standard output. It takes as parameters two sets of characters, and replaces occurrences of the characters in the first set with the corresponding elements from the other set. For example,

    +
    $ tr 'abcd' 'jkmn'
    +

    maps 'a' to 'j', 'b' to 'k', 'c' to 'm', and 'd' to 'n'.

    +

    Sets of characters may be abbreviated by using character ranges. The previous example could be written:

    +
    $ tr 'a-d' 'jkmn'
    +

    In POSIX compliant versions of tr the set represented by a character range depends on the locale's collating order, so it is safer to avoid character ranges in scripts that might be executed in a locale different from that in which they were written. Ranges can often be replaced with POSIX character sets such as [:alpha:].

    +

    The -c flag complements the first set of characters.

    +
    $ tr -cd '[:alnum:]'
    +

    therefore removes all non-alphanumeric characters.

    +

    The -s flag causes tr to compress sequences of identical adjacent characters in its output to a single token. For example,

    +
    $ tr -s '\n' '\n'
    +

    replaces sequences of one or more newline characters with a single newline.

    +

    The -d flag causes tr to delete all tokens of the specified set of characters from its input. In this case, only a single character set argument is used. The following command removes carriage return characters, thereby converting a file in DOS/Windows format to one in Unix format.

    +
    $ tr -d '\r'
    +

    Most versions of tr, including GNU tr and classic Unix tr, operate on single byte characters and are not Unicode compliant. An exception is the Heirloom Toolchest implementation, which provides basic Unicode support.

    +

    Ruby and Perl also have an internal tr operator, which operates analogously. Tcl's string map command is more general in that it maps strings to strings while tr maps characters to characters.

    +
    +
    +
    +

    +14. Elementary Regex

    +

    In computing, regular expressions provide a concise and flexible means for identifying strings of text of interest, such as particular characters, words, or patterns of characters. A regular expression (often shortened to regex or regexp) is written in a formal language that can be interpreted by a regular expression processor, a program that either serves as a parser generator or examines text and identifies parts that match the provided specification.

    +

    Regular expressions are used by many text editors, utilities, and programming languages to search and manipulate text based on patterns. For example, Perl, Ruby and Tcl have a powerful regular expression engine built directly into their syntax. Several utilities provided by Unix distributions—including the editor ed and the filter grep — were the first to popularize the concept of regular expressions.

    +

    Traditional Unix regular expression syntax followed common conventions but often differed from tool to tool. The IEEE POSIX Basic Regular Expressions (BRE) standard (released alongside an alternative flavor called Extended Regular Expressions or ERE) was designed mostly for backward compatibility with the traditional (Simple Regular Expression) syntax but provided a common standard which has since been adopted as the default syntax of many Unix regular expression tools, though there is often some variation or additional features. Many such tools also provide support for ERE syntax with command line arguments.

    +

    In the BRE syntax, most characters are treated as literals — they match only themselves (i.e., a matches "a"). The exceptions, listed below, are called metacharacters or metasequences.

    ++++
    +
    +

    +14.1. Lazy quantification

    +

    The standard quantifiers in regular expressions are greedy, meaning they match as much as they can, only giving back as necessary to match the remainder of the regex. For example, someone new to regexes wishing to find the first instance of an item between < and > symbols in this example:

    +
    Another whale explosion occurred on <January 26>, <2004>.
    +

    ...would likely come up with the pattern <.*>, or similar. However, this pattern will actually return "<January 26>, <2004>" instead of the "<January 26>" which might be expected, because the <title_reference>*</title_reference> quantifier is greedy — it will consume as many characters as possible from the input, and "January 26>, <2004" has more characters than "January 26".

    +

    Though this problem can be avoided in a number of ways (e.g., by specifying the text that is not to be matched: <[^>]*>), modern regular expression tools allow a quantifier to be specified as lazy (also known as non-greedy, reluctant, minimal, or ungreedy) by putting a question mark after the quantifier (e.g., <.*?>), or by using a modifier which reverses the greediness of quantifiers (though changing the meaning of the standard quantifiers can be confusing). By using a lazy quantifier, the expression tries the minimal match first. Though in the previous example lazy matching is used to select one of many matching results, in some cases it can also be used to improve performance when greedy matching would require more backtracking.

    +
    +
    +
    +

    +15. One Liners

    +

    A one-liner is textual input to the command-line of an operating system shell that performs some function in just one line of input.

    +

    The one liner can be

    +<block_quote>
      +
    1. An expression written in the language of the shell.

    2. +
    3. The invocation of an interpreter together with program source for the interpreter to run.

    4. +
    5. The invocation of a compiler together with source to compile and +instructions for executing the compiled program.

    6. +
    </block_quote>

    Certain dynamic scripting languages such as AWK, sed, and perl have traditionally been adept at expressing one-liners. Specialist shell interpreters such as these Unix shells or the Windows PowerShell, allow for the construction of powerful one-liners.

    +

    The use of the phrase one-liner has been widened to also include program-source for any language that does something useful in one line.

    +

    The word One-liner has two references in the index of the book The AWK Programming Language (the book is often referred to by the abbreviation TAPL). It explains the programming language AWK, which is part of the Unix operating system. The authors explain the birth of the One-liner paradigm with their daily work on early Unix machines:

    +
    “The 1977 version had only a few built-in variables and predefined functions. It was designed for writing short programs [...] Our model was that an invocation would be one or two lines long, typed in and used immediately. Defaults were chosen to match this style [...] We, being the authors, knew how the language was supposed to be used, and so we only wrote one-liners.”
    +

    Notice that this original definition of a One-liner implies immediate execution of the program without any compilation. So, in a strict sense, only source code for interpreted languages qualifies as a One-liner. But this strict understanding of a One-liner was broadened in 1985 when the IOCCC introduced the category of Best One Liner for C, which is a compiled language.

    +

    The TAPL book contains 20 examples of One-liners (A Handful of Useful awk One-Liners) at the end of the book's first chapter.

    +

    Here are the first few of them:

    +<block_quote>
      +
    1. +

      Print the total number of input lines:

      +

      END { print NR }

      +
    2. +
    3. +

      Print the tenth input line:

      +

      NR == 10

      +
    4. +
    5. +

      Print the last field of every input line:

      +

      { print $NF }

      +
    6. +
    </block_quote>

    One-liners are also used to show off the differential expressive power of programming languages. Frequently, one-liners are used to demonstrate programming ability. Contests are often held to see who can create the most exceptional one-liner.

    +

    The following example is a C program (a winning entry in the "Best one-liner" category of the IOCCC, here split to two lines for presentation).:

    +
    main(int c,char**v){return!m(v[1],v[2]);}m(char*s,char*t){return
    +*t-42?*s?63==*t|*s==*t&&m(s+1,t+1):!*t:m(s,t+1)||*s&&m(s+1,t);}
    +

    This one-liner program is a glob pattern matcher. It understands the glob characters '*' meaning 'zero or more characters' and '?' meaning exactly one character, just like most Unix shells.

    +

    Run it with two args, the string and the glob pattern. The exit status is 0 (shell true) when the pattern matches, 1 otherwise. The glob pattern must match the whole string, so you may want to use * at the beginning and end of the pattern if you are looking for something in the middle. Examples:

    +
    $ prog foo 'f??'; echo $?
    +
    +$ prog 'best short program' '??st*o**p?*'; echo $?
    +

    Here is a one line shell script to show directories:

    +
    $ ls -R | grep ":$" | sed -e 's/:$//' -e 's/[^-][^\/]*\//--/g' -e 's/^/   /' -e 's/-/|/'
    +
    +
    +
    + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/ch4strings_dicts.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/ch4strings_dicts.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1139 @@ + + + +Strings and Dicts  + + + + + + + +
    +
    + +
    +
    +

    +
    +
    + +
    +

    +1. Introducing Linux

    +

    (Attribution : A significant chunk of the content under this section is based on data from Wikipedia and the Linux Documentation Project)

    +

    Linux (usually pronounced ˈlɪnəks') is a generic term referring to Unix-like computer operating systems based on the Linux kernel, where a kernel is the intermediate layer between the hardware and the applications. The kernel is, on an abstract level, the core of (most) operating systems, that manages the various system resources. The development of the Linux OS is considered the basis for Free and Open Source Software (FOSS) collaboration since typically the underlying source code can be used, modified freely, and redistributed by anyone under the terms of the GNU (a recursive acronym for "GNU's Not Unix!") Global Public License (GPL) and other free software licences. This freedom to access and reuse various components of a system, is one of the primary reasons for the popularity of Linux.

    +

    Linux is installed on a variety of computer hardware, that include mobile phones, embedded devices and supercomputers, but is infamous for its use in servers.

    +

    The name "Linux" comes from the Linux kernel, originally written in 1991 by Linus Torvalds. The rest of the system usually comprises components such as the Apache HTTP Server, the X Window System, the GNOME and KDE desktop environments, and utilities and libraries from the GNU Project (announced in 1983 by Richard Stallman). Commonly-used applications with desktop Linux systems include the Mozilla Firefox web-browser and the OpenOffice.org office application suite. The GNU contribution is the basis for the Free Software Foundation's preferred name GNU/Linux. The kernel's mascot is a penguin named "Tux". Mozilla Firefox and OpenOffice.org are open-source projects which can be run on most Operating Systems, including proprietary ones.

    +
    +

    +1.1. Historical Background

    +
    +

    +1.1.1. Events leading to the creation

    +
      +
    • The Unix operating system was developed in the 1960s and released for public use in 1970. Its accessibility and portability caused it to be widely adopted, copied and modified by academic institutions and businesses. Its design became influential to authors of other systems. Other free operating systems include the Berkeley Software Distribution (BSD), developed at the University of California at Berkeley, and MINIX which was released by Andrew S. Tanenbaum. The development and adoption of BSD and MINIX were limited due to various reasons, and this lack of a widely-adopted and free kernel triggered Linus Torvalds into starting his project.

    • +
    • In 1983, Richard Stallman started the GNU project with the goal of creating a free UNIX-like operating system. As part of this work, he wrote the GNU General Public License (GPL). By the early 1990s there was almost enough available software to create a full operating system. However, the GNU kernel, called Hurd, failed to attract enough attention from developers leaving GNU incomplete.

    • +
    +
    +
    +

    +1.1.2. The Creation of Linux

    +

    In 1991, Linus Torvalds began a project at the University of Helsinki that later became the Linux kernel. It was initially a terminal (command-line) emulator, which Torvalds used to access the large UNIX servers of the university. He wrote the program targeting just the hardware he was using and independent of an operating system because he wanted to use the functions of his computer with an 80386 processor. Development was done on Minix using the GNU C compiler. This application is still the main choice for compiling Linux today (although the code can be built with other compilers, such as the Intel C Compiler).

    +

    Torvalds continues to direct the development of the kernel. Stallman heads the Free Software Foundation, which in turn supports the GNU components. Finally, individuals and corporations develop third-party non-GNU components, which constitute a vast body of work and including kernel modules, and user applications and libraries. Linux vendors and communities combine and distribute the kernel, GNU components, and non-GNU components, with additional package management software in the form of Linux distributions.

    +
    +
    +
    +

    +1.2. Design and Implications

    +

    A Linux-based system is a modular Unix-like operating system, deriving much of its basic design from principles established in Unix earlier. Such a system uses a monolithic kernel, called the Linux kernel, which handles process control, networking, and peripheral and file system access. Device drivers are integrated directly with the kernel. Separate projects that interface with the kernel provide much of the system's higher-level functionality. The GNU userland is an important part of most Linux-based systems, providing the most common implementation of the C library, a popular shell, and many of the common Unix tools which carry out many basic operating system tasks. The graphical user interface (or GUI) used by most Linux systems is based on the "X Window System".

    +
    +

    +1.2.1. User Interface

    +

    Users can control a Linux-based system through a command line interface (or CLI), a graphical user interface (or GUI), or through controls attached to the associated hardware (this is common for embedded systems). For desktop systems, the default mode is usually the GUI. On desktop machines, "KDE", "GNOME" and "Xfce" are the most popular user interfaces,though a variety of additional user interfaces exist. Most popular user interfaces run on top of the "X Window System" (or X), which enables a graphical application running on one machine to be displayed and controlled from another in a network.

    +

    A Linux system also provides a CLI of some sort through a shell, which is the traditional way of interacting with a Unix system. A Linux distribution specialized for servers may use the CLI as its only interface. A “headless system” (system run without even a monitor) can be controlled by the command line via a remote-control protocol such as SSH or telnet. The CLI is particularly suited for automation of repetitive or delayed tasks, and provides very simple inter-process communication. A graphical terminal emulator program is often used to access the CLI from a Linux desktop.

    +
    +
    +

    +1.2.2. Development

    +

    The primary difference between Linux and many other popular contemporary operating systems is that the Linux kernel and other components are free and open source software. Linux is not the only such operating system, although it is by far the most widely used. Some free and open source software licenses are based on the principle of "copyleft", a kind of reciprocity: any work derived from a copyleft piece of software must also be copyleft itself. The most common free software license, the GNU GPL, is a form of copyleft, and is used for the Linux kernel and many of the components from the GNU project.

    +

    Linux based distributions are intended by developers for interoperability with other operating systems and established computing standards. Linux systems adhere to POSIX, SUS, ISO and ANSI standards where possible, although to date only one Linux distribution has been POSIX.1 certified, Linux-FT.Free software projects, although developed in a collaborative fashion, are often produced independently of each other. The fact that the software licenses explicitly permit redistribution, however, provides a basis for larger scale projects that collect the software produced by stand-alone projects and make it available all at once in the form of a Linux distribution.

    +

    A Linux distribution, commonly called a "distro", is a project that manages a remote collection of system software and application software packages available for download and installation through a network connection. This allows the user to adapt the operating system to his/her specific needs. Distributions are maintained by individuals, loose-knit teams, volunteer organizations, and commercial entities. A distribution can be installed using a CD that contains distribution-specific software for initial system installation and configuration. A package manager such as Synaptic or YAST allows later package upgrades and installations. A distribution is responsible for the default configuration of the installed Linux kernel, general system security, and more generally integration of the different software packages into a coherent whole.

    +
    +
    +

    +1.2.3. Community

    +

    A distribution is largely driven by its developer and user communities. Some vendors develop and fund their distributions on a volunteer basis. Examples include Debian and the Debian-based, Ubuntu. Others maintain a community version of their commercial distributions, as Red Hat does with Fedora.

    +

    In many cities and regions, local associations known as Linux Users Groups (LUGs) seek to promote their preferred distribution and by extension free software. They hold meetings and provide free demonstrations, training, technical support, and operating system installation to new users. Many Internet communities also provide support to Linux users and developers. Most distributions and free software / open source projects have IRC (Internet Relay Chat) chatrooms or newsgroups. Online forums are another means for support. Linux distributions host mailing lists; commonly there will be a specific topic such as usage or development for a given list. All these can be found simply by running an appropriate search on Google.

    +

    Although Linux distributions are generally available without charge, several large corporations sell, support, and contribute to the development of the components of the system and of free software. These include Dell, IBM, HP, Oracle, Sun Microsystems, Novell, Nokia. A number of corporations, notably Red Hat, have built their entire business around Linux distributions.

    +
    +
    +

    +1.2.4. Can I make a profit out of running a business involving Linux?

    +

    The answer is, "Yes!". The free software licenses, on which the various software packages of a distribution built on the Linux kernel are based, explicitly accommodate and encourage commercialization; the relationship between a Linux distribution as a whole and individual vendors may be seen as symbiotic. One common business model of commercial suppliers is charging for support, especially for business users. A number of companies also offer a specialized business version of their distribution, which adds proprietary support packages and tools to administer higher numbers of installations or to simplify administrative tasks. Another business model is to give away the software in order to sell hardware. Examples of corporations that are extensively (and sometimes exclusively) open-source and Linux-powered , with successful revenue generation models involving these, are Google, SUN, Mozilla, etc.

    +
    +
    +

    +1.2.5. Programming on Linux

    +

    Most Linux distributions support dozens of programming languages. The most common collection of utilities for building both Linux applications and operating system programs is found within the GNU toolchain, which includes the GNU Compiler Collection (GCC) and the GNU build system. Amongst others, GCC provides compilers for Ada, C, C++, Java, and Fortran. The Linux kernel itself is written to be compiled with GCC. Proprietary compilers for Linux include the Intel C++ Compiler, Sun Studio, and IBM XL C/C++ Compiler.

    +

    Most distributions also include support for PHP, Perl, Ruby, Python and other dynamic languages. Examples of languages that are less common, but still supported, are C# via the Mono project, sponsored by Novell, and Scheme. A number of Java Virtual Machines and development kits run on Linux, including the original Sun Microsystems JVM (HotSpot), and IBM's J2SE RE, as well as many open-source projects like Kaffe.

    +

    The two main frameworks for developing graphical applications are those of GNOME and KDE. These projects are based on the GTK+ and Qt widget toolkits, respectively, which can also be used independently of the larger framework. Both support a wide variety of languages. There are a number of Integrated Development Environments (IDEs) available including Anjuta, Code::Blocks, Eclipse, KDevelop, Lazarus, MonoDevelop, NetBeans, and Omnis Studio while the long-established editors Vim and Emacs remain popular.

    +
    +
    +
    +

    +1.3. Reasons for Using Linux

    +
    • Linux is free:

    +

    As in "free beer". Linux can be downloaded in its entirety from the Internet completely for free. No registration fees, no costs per user, free updates, and freely available source code in case you want to change the behavior of your system. +Most of all, Linux is free as in "free speech": +The license commonly used is the GNU Public License (GPL). The license says that anybody who may want to do so, has the right to change Linux and eventually to redistribute a changed version, on the one condition that the code is still available after redistribution. In practice, you are free to grab a kernel image and sell the new code, as long as your customers can still have a copy of that code.

    +
    • Linux is portable to any hardware platform:

    +

    A vendor, who wants to sell a new type of computer and who does not know what kind of OS his/her new machine will run, can take a Linux kernel and make it work on his/her hardware, because documentation related to this activity is freely available.

    +
    • Linux was made to keep on running:

    +

    As with UNIX, a Linux system expects to run without rebooting all the time. That is why a lot of tasks are being executed at night or scheduled automatically for other times, resulting in higher availability during busier periods and a more balanced use of the hardware. This property allows for Linux to be applicable to environments where people do not have the time or the possibility to control their systems constantly.

    +
    • Linux is secure and versatile:

    +

    The security model used in Linux is based on the UNIX idea of security, which is known to be robust and of proven quality. But Linux is not only safe from attacks from the Internet: it will adapt equally to other situations, utilizing the same high standards for security.

    +
    • Linux is scalable:

    +

    From a Palmtop with 2 MB of memory to a petabyte storage cluster with hundreds of nodes: add or remove the appropriate packages and Linux fits all. One does not need a supercomputer anymore,because you can use Linux to do big things using the building blocks provided with the system. If one wants to do little things, such as making an operating system for an embedded processor or just recycling your old 486, Linux will do that as well.

    +
    • The Linux OS and Linux applications have very short debug−times:

    +

    Because Linux has been developed and tested by thousands of people, both errors and people to fix them are found very quickly. It often happens that there are only a couple of hours between discovery and fixing of a bug.

    +
    +
    +
    +

    +2. Getting Started

    +
    +

    +2.1. Logging in, activating the user interface and logging out

    +

    In order to work on a Linux system directly, one needs to provide a user name and password. You always need to authenticate to the system. Most PC−based Linux systems have two basic modes for a system to run in: either quick and clean in text console mode,which includes with mouse, multitasking and multi−user features, or in graphical console mode, which looks better but eats more system resources.

    +
    +

    +2.1.1. Graphical Mode

    +

    This is the default nowadays on most desktop computers. You know you will be connecting to the system using graphical mode when you are first asked for your user name, and then to type your password.

    +

    To log in, make sure the mouse pointer is in the login window, provide your user name and password to the system and click OK or press Enter. +It is generally considered a bad idea to connect (graphically) using the root user name, the system adminstrator's account, since the use of graphics includes running a lot of extra programs, in root's case with a lot of extra permissions. To keep all risks as low as possible, use a normal user account to connect graphically. But there are enough risks to keep this in mind as a general advice, for all use of the root account: only log in as root when extra privileges are required.

    +

    After entering your user name/password combination, it can take a little while before the graphical environment is started, depending on the CPU speed of your computer, on the software you use and on your personal settings.

    +

    To continue, you will need to open a terminal window or xterm for short (X being the name for the underlying software supporting the graphical environment). This program can be found in the Applications−>Utilities->System Tools or Internet menu, depending on what window manager you are using. There might be icons that you can use as a shortcut to get an xterm window as well, and clicking the right mouse button on the desktop background will usually present you with a menu containing a terminal window application.

    +

    While browsing the menus, you will notice that a lot of things can be done without entering commands via the keyboard. For most users, the good old point−n−click method of dealing with the computer will do. But for those who want to enter the "heart" of the system, a tool stronger than a mouse will be required to handle the various tasks. This tool is the shell, and when in graphical mode, we activate our shell by opening a terminal window.

    +

    A terminal window should always show a command prompt when you open one. This terminal shows a standard prompt, which displays the user's login name, and the current working directory, represented by the twiddle (~)

    +

    Another common form for a prompt is this one: +[

    +
    +

    user@host
    +

    dir]

    +

    In the above example, user will be your login name, hosts the name of the machine you are working on, and dir an indication of your current location in the file system. Prompts can display all kinds of information, but they are not part of the commands you are giving to your system. To disconnect from the system in graphical mode, you need to close all terminal windows and other applications. After that, hit the logout icon or find Log Out in the menu. Closing everything is not really necessary, and the system can do this for you, but session management might put all currently open applications back on your screen when you connect again, which takes longer and is not always the desired effect. However, this behavior is configurable.

    +

    When you see the login screen again, asking to enter user name and password, logout was successful.

    +
    +
    +

    +2.1.2. Text Mode

    +

    One is in text mode when the whole screen is black, showing (in most cases white) characters. A text mode login screen typically shows some information about the machine you are working on, the name of the machine and a prompt waiting for you to log in.

    +

    The login is different from a graphical login, in that you have to hit the Enter key after providing your user name, because there are no buttons on the screen that you can click with the mouse. Then you should type your password, followed by another Enter. You will not see any indication that you are entering something, not even an asterisk, and you won't see the cursor move. But this is normal on Linux and is done for security +reasons.

    +

    When the system has accepted you as a valid user, you may get some more information, called the message of the day, which can be anything. Additionally, it is popular on UNIX systems to display a fortune cookie, which contains some general wise or unwise (this is up to you) thoughts. After that, you will be given a shell, indicated with the same prompt that you would get in graphical mode.

    +

    Also in text mode: log in as root only to do setup and configuration that absolutely requires administrator privileges, such as adding users, installing software packages, and performing network and other system configuration. Once you are finished, immediately leave the special account and resume your work as a non−privileged user.

    +

    Logging out is done by entering the logout command, followed by Enter. You are successfully disconnected from the system when you see the login screen again.Don't power−off the computer after logging out. It is not meant to be shut off without application of the proper procedures for halting the system. Powering it off without going through the halting process might cause severe damage!

    +
    +
    +
    +
    +

    +3. Basic Commands

    +
    +

    +3.1. ls

    +

    When invoked without any arguments, ls lists the files in the current working directory. A directory that is not the current working directory can be specified and ls will list the files there. The user also may specify any list of files and directories. In this case, all files and all contents of specified directories will be listed. The name ls is derived from list segments which was used in earlier systems.

    +

    Files whose names start with "." are not listed, unless the -a flag is specified or the files are specified explicitly.

    +

    Without options, ls displays files in a bare format. This bare format however makes it difficult to establish the type, permissions, and size of the files. The most common options to reveal this information or change the list of files are:

    +<block_quote>
      +
    • -l long format, displaying Unix file types, permissions, number of hard links, owner, group, size, date, and filename

    • +
    • -F appends a character revealing the nature of a file, for example, * for an executable, or / for a directory. Regular files have no suffix.

    • +
    • -a lists all files in the given directory, including those whose names start with "." (which are hidden files in Unix). By default, these files are excluded from the list.

    • +
    • -R recursively lists subdirectories. The command ls -R / would therefore list all files.

    • +
    • -d shows information about a symbolic link or directory, rather than about the link's target or listing the contents of a directory.

    • +
    • -t sort the list of files by modification time.

    • +
    • -h print sizes in human readable format. (e.g., 1K, 234M, 2G, etc.)

    • +
    </block_quote>

    In some environments, providing the option --color (for GNU ls) or -G (FreeBSD ls) causes ls to highlight different types of files with different colors, instead of with characters as -F would. To determine what color to use for a file, GNU ls checks the Unix file type, the file permissions, and the file extension, while FreeBSD ls checks only the Unix file type and file permissions.:

    +
    $ ls
    +jeeves.rst psmith.html blandings.html
    +$ ls -l
    +drwxr--r--   1 plum  editors   4096  jeeves
    +-rw-r--r--   1 plum  editors  30405  psmith
    +-r-xr-xr-x   1 plum  plum      8460  blandings
    +

    Here "$" actually is the beginning of the prompt. This is typical in most Unix-based systems.

    +
    +
    +

    +3.2. date

    +

    The Unix date command displays the time and date. The super-user can use it to set the system clock.

    +

    With no options, the date command displays the current date and time, including the abbreviated day name, abbreviated month name, day of the month, the time separated by colons, the timezone name, and the year. For example:

    +
    $date
    +Tue Sep  8 12:01:45 IST 2009
    +

    On some systems to set the current date and time to September 8, 2004 01:22 you type:

    +
    $date --set="20040908 01:22"
    +

    In order to view the various options for the date command, type:

    +
    $man date
    +

    This will take you to the "Manual" page comprising of all the details on the date command. You can return to the terminal from the "man" page by pressing the Esc key in the keyboard and typing ":q" in that order.

    +
    +
    +

    +3.3. cd

    +

    This stands for "change directory". When one wants to go up to the parent directory, bypassing the tree of directories one has entered, “ cd ..” can be used.

    +

    One dot '.' represents the current directory while two dots '..' represent the parent directory.

    +

    “ cd -” will return you to the previous directory (a bit like an “undo”).

    +

    You can also use cd absolute path or cd relative path (see below):

    +

    Absolute paths:

    +<block_quote>

    An “ absolute path” is easily recognised from the leading forward slash, /. The / means that you start at the top level directory and continue down.

    </block_quote>

    For example to get to /boot/grub you would type:

    +
    $cd /boot/grub
    +

    This is an absolute path because you start at the top of the hierarchy and go downwards from there (it doesn't matter where in the filesystem you were when you typed the command).

    +

    Relative paths:

    +<block_quote>

    A “ relative path” doesn't have a preceding slash. Use a relative path when you start from a directory below the top level directory structure. This is dependent on where you are in the filesystem.

    +

    For example if you are in root's home directory and want to get to /root/music, you type:

    +
    $ cd music
    </block_quote>

    Please note that there is no / using the above cd command. Using a / would cause this to be an absolute path, working from the top of the hierarchy downward.

    +
    +
    +

    +3.4. who

    +

    The standard Unix command who displays a list of users who are currently logged into a computer.

    +

    The who command is related to the command w, which provides the same information but also displays additional data and statistics.:

    +
    $who
    +beeblebrox tty7         2009-09-08 10:50 (:0)
    +beeblebrox pts/0        2009-09-08 11:25 (:0.0)
    +dumbledore pts/1        2009-09-08 18:11 (potter.xyz.in)
    +beeblebrox pts/2        2009-09-08 18:53 (:0.0)
    +

    The command can be invoked with the arguments am i or am I (so it is invoked as who am i or * who am I*), showing information about the current terminal only (see the -m option below, of which this invocation is equivalent).

    +

    In order to find out the various options that can be appended to the who command, check the man page by typing out the following in the terminal:

    +
    $man who
    +

    This will take you to the "Manual" page containing details about the who command

    +
    +
    +

    +3.5. mkdir

    +

    This command is used to make a new directory. Normal usage is as straightforward as follows:

    +
    $mkdir name_of_directory
    +

    Where name_of_directory is the name of the directory one wants to create. When typed as above (ie. normal usage), the new directory would be created within the current directory. On Unix, multiple directories can be specified, and mkdir will try to create all of them.

    +
    +

    +3.5.1. Options

    +

    On Unix-like operating systems, mkdir takes options. Three of the most common options are:

    +<block_quote>
      +
    • -p: will also create all directories leading up to the given directory that do not exist already. If the given directory already exists, ignore the error.

    • +
    • -v: display each directory that mkdir creates. Most often used with -p.

    • +
    • -m: specify the octal permissions of directories created by mkdir.

    • +
    </block_quote>

    -p is most often used when using mkdir to build up complex directory hierarchies, in case a necessary directory is missing or already there. -m is commonly used to lock down temporary directories used by shell scripts.

    +
    +
    +

    +3.5.2. Examples

    +

    An example of -p in action is:

    +
    $mkdir -p /tmp/a/b/c
    +

    If /tmp/a exists but /tmp/a/b does not, mkdir will create /tmp/a/b before creating /tmp/a/b/c.

    +

    And an even more powerful command, creating a full tree at once (this however is a Shell extension, nothing mkdir does itself):

    +
    $mkdir -p tmpdir/{trunk/sources/{includes,docs},branches,tags}
    +

    This will create:

    +<definition_list><definition_list_item><term>tmpdir - branches</term><definition>
      +
    • tag

    • +
    • <definition_list><definition_list_item><term>trunk - sources - includes</term><definition>
      • docs

      </definition>
      </definition_list_item>
      </definition_list>
    • +
    </definition>
    </definition_list_item>
    </definition_list>
    +
    +
    +
    +
    +

    +4. Getting Help

    +
    +

    +4.1. apropos and whatis

    +

    This is a command to search the manual pages files in Unix and Unix-like operating systems.

    +
    $ apropos grep
    +egrep       egrep (1)       Search a file for a pattern using full regular expressions
    +fgrep       fgrep (1)       Search a file for a fixed-character string
    +fmlgrep     fmlgrep (1)     Search a file for a pattern
    +grep        grep (1)        Search a file for a pattern
    +gzgrep      gzgrep (1)      Search a possibly compressed file for a regular expression
    +nisgrep     nismatch (1)    Utilities for searching NIS+ tables
    +pgrep       pgrep (1)       Find or signal a process by name or other attribute
    +zgrep       zgrep (1)       Search a possibly compressed file for a regular expression
    +...
    +

    In this example, the user uses apropos to search for the string "grep", and apropos returns the indicated man pages that include the term "grep".

    +

    A short index of explanations for commands is available using the whatis command, like in the examples below:

    +
    $whatis ls
    +ls (1)           - list directory contents
    +

    This displays short information about a command, and the first section in the collection of man pages that contains an appropriate page.

    +

    If you don't know where to get started and which man page to read, apropos gives more information. Say that you do not know how to start a browser, then you could enter the following command:

    +
    $apropos browser
    +gmusicbrowser (1)    - Jukebox for large collections of audio files
    +infobrowser (1)      - read Info documents
    +libsmbclient (7)     - An extension library for browsers and that               can be used...
    +opera (1)            - a standards-compliant graphical Web browser
    +sensible-browser (1) - sensible editing, paging, and web browsing
    +smbtree (1)          - A text based smb network browser
    +tvtk_doc (1)         - A GUI based TVTK documentation search browser.
    +viewres (1)          - graphical class browser for Xt
    +w3m (1)              - a text based Web browser and pager
    +www-browser (1)      - a text based Web browser and pager
    +...
    +
    +
    +

    +4.2. man

    +

    Man pages (short for "manual pages") are the extensive documentation that comes preinstalled with almost all substantial Unix and Unix-like operating systems. The Unix command used to display them is man. Each page is a self-contained document.

    +

    To read a manual page for a Unix command, one can use:

    +
    $ man <command_name>
    +

    at a shell prompt; for example, "man ftp". In order to simplify navigation through the output, man generally uses the less terminal pager.

    +

    Pages are traditionally referred to using the notation "name(section)"; for example, ftp(1). The same page name may appear in more than one section of the manual, this can occur when the names of system calls, user commands, or macro packages coincide. Two examples are man(1) and man(7), or exit(2) and exit(3). The syntax for accessing the non-default manual section varies between different man implementations. On Linux and BSD, for example, the syntax for reading *printf(3) is:

    +
    $man 3 printf
    +

    Another example:

    +
    $man man
    +

    The previous example will take you to the "Manual" page entry about manual pages!

    +
    +

    +4.2.1. Layout

    +

    All man pages follow a common layout that is optimized for presentation on a simple ASCII text display, possibly without any form of highlighting or font control. Sections present may include:

    +<definition_list><definition_list_item><term>NAME</term><definition>

    The name of the command or function, followed by a one-line description of what it does.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SYNOPSIS</term><definition>

    In the case of a command, you get a formal description of how to run it and what command line options it takes. For program functions, a list of the parameters the function takes and which header file contains its definition. For experienced users, this may be all the documentation they need.

    </definition>
    </definition_list_item>
    <definition_list_item><term>DESCRIPTION</term><definition>

    A textual description of the functioning of the command or function.

    </definition>
    </definition_list_item>
    <definition_list_item><term>EXAMPLES</term><definition>

    Some examples of common usage.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SEE ALSO</term><definition>

    A list of related commands or functions.

    </definition>
    </definition_list_item>
    </definition_list>

    Other sections may be present, but these are not well standardized across man pages. Common examples include: OPTIONS, EXIT STATUS, ENVIRONMENT, KNOWN BUGS, FILES, AUTHOR, REPORTING BUGS, HISTORY and COPYRIGHT.

    +

    These days virtually every Unix command line application comes with its man page, and many Unix users perceive a lack of man pages as a sign of low quality; indeed, some projects, such as Debian, go out of their way to write man pages for programs lacking one. Few alternatives to man have enjoyed much popularity, with the possible exception of the GNU project's "info" system, an early and simple hypertext system.

    +

    However, the format of a single page for each application, the lack of classification within the sections and the relatively unsophisticated formatting facilities have motivated the development of alternative documentation systems, such as the previously mentioned "info" system.

    +

    Most Unix GUI applications (particularly those built using the GNOME and KDE development environments) now provide end-user documentation in HTML and include embedded HTML viewers such as yelp for reading the help within the application.

    +

    Usually the man pages are written in English. Translations into other languages can be also available on the system.

    +

    The default format of the man pages is troff, with either the macro package man (appearance oriented) or on some systems mdoc (semantic oriented). This makes it possible to typeset a man page to PostScript, PDF and various other formats for viewing or printing.

    +
    +
    +
    +

    +4.3. info

    +

    info is a software utility which forms a hypertextual, multipage documentation and help viewer working on a command line interface, useful when there is no GUI available.

    +

    The syntax is

    +
    $ info <command_name>
    +

    info processes info files, which are Texinfo formatted files, and presents the documentation as a tree, with simple commands to traverse the tree and to follow cross references. For instance

    +<block_quote>
      +
    • n goes to the next page.

    • +
    • p goes to the previous page.

    • +
    • u goes to the upper page.

    • +
    • l goes to the last(visited) node

    • +
    • To follow a cross reference, the cursor can be moved over a link (a word preceded by a <title_reference>*</title_reference>) and enter pressed.

    • +
    </block_quote>

    info was initially written for use with GNU/Linux and then ported to other Unix-like operating systems.

    +
    +
    +

    +4.4. --help

    +

    Most GNU commands support the --help, which gives a short explanation about how to use the command and a list of available options. Below is the output of this option with the cat command:

    +
    $ userprompt@host: cat --help
    +Usage: cat [OPTION] [FILE]...
    +Concatenate FILE(s), or standard input, to standard output.
    +
    +  -A, --show-all           equivalent to -vET
    +  -b, --number-nonblank    number nonempty output lines
    +  -e                       equivalent to -vE
    +  -E, --show-ends          display $ at end of each line
    +  -n, --number             number all output lines
    +  -s, --squeeze-blank      suppress repeated empty output lines
    +  -t                       equivalent to -vT
    +  -T, --show-tabs          display TAB characters as ^I
    +  -u                       (ignored)
    +  -v, --show-nonprinting   use ^ and M- notation, except for LFD and              TAB
    +  --help     display this help and exit
    +  --version  output version information and exit
    +
    +With no FILE, or when FILE is -, read standard input.
    +
    +Examples:
    +  cat f - g  Output f's contents, then standard input, then g's           contents.
    +  cat        Copy standard input to standard output.
    +
    +Report bugs to <bug-coreutils@gnu.org>.
    +
    +
    +
    +

    +5. Basic file handling

    +
    +

    +5.1. cp

    +

    cp is the command entered in a Unix shell to copy a file from one place to another, possibly on a different filesystem. The original file remains unchanged, and the new file may have the same or a different name.

    +
    +

    +5.1.1. Usage

    +

    To copy a file to another file:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ][ -- ] SourceFile TargetFile
    +

    To copy a file to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -r | -R ] [ -- ] SourceFile ...              TargetDirectory
    +

    To copy a directory to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -- ] { -r | -R }
    +SourceDirectory ... TargetDirectory
    +
    +
    +

    +5.1.2. Flags

    +

    -f (force) – specifies removal of the target file if it cannot be opened for write operations. The removal precedes any copying performed by the cp command.

    +

    -P – makes the cp command copy symbolic links. The default is to follow symbolic links, that is, to copy files to which symbolic links point.

    +

    -i (interactive) – prompts you with the name of a file to be overwritten. This occurs if the TargetDirectory or TargetFile parameter contains a file with the same name as a file specified in the SourceFile or SourceDirectory parameter. If you enter y or the locale's equivalent of y, the cp command continues. Any other answer prevents the cp command from overwriting the file.

    +

    -p (preserve) – duplicates the following characteristics of each SourceFile/SourceDirectory in the corresponding TargetFile and/or TargetDirectory:

    +<block_quote>
      +
    • The time of the last data modification and the time of the last access.

    • +
    • The user ID and group ID (only if it has permissions to do this)

    • +
    • The file permission bits and the SUID and SGID bits.

    • +
    </block_quote>

    -R (recursive) – copy directories (recursively copying all the contents)

    +
    +
    +

    +5.1.3. Examples

    +

    To make a copy of a file in the current directory, enter:

    +
    $ cp prog.c prog.bak
    +

    This copies prog.c to prog.bak. If the prog.bak file does not already exist, the cp command creates it. If it does exist, the cp command replaces it with a copy of the prog.c file.

    +

    To copy a file in your current directory into another directory, enter:

    +
    $ cp zaphod /home/books/hhgg
    +

    This copies the jones file to /home/books/hhgg/zaphod.

    +

    To copy a file to a new file and preserve the modification date, time, and access control list associated with the source file, enter:

    +
    $ cp -p martin_luther_king martin_luther_king.jr
    +

    This copies the martin_luther_king file to the martin_luther_king.jr file. Instead of creating the file with the current date and time stamp, the system gives the martin_luther_king.jr file the same date and time as the martin_luther_king file. The martin_luther_king.jr file also inherits the martin_luther_king file's access control protection.

    +

    To copy all the files in a directory to a new directory, enter:

    +
    $ cp /home/galactica/clients/* /home/hhgg/customers
    +

    This copies only the files in the clients directory to the customers directory.

    +

    To copy a directory, including all its files and subdirectories, to another directory, enter:

    +<block_quote>

    $ cp -R /home/hhgg/clients /home/hhgg/customers

    </block_quote>

    This copies the clients directory, including all its files, subdirectories, and the files in those subdirectories, to the customers/clients directory.

    +

    To copy a specific set of files of any extension to another directory, enter:

    +
    $ cp zaphod arthur ford /home/hhgg/clients
    +

    This copies the zaphod, arthur, and ford files in your current working directory to the /home/hhgg/clients directory.

    +

    To use pattern-matching characters to copy files, enter:

    +
    $ cp programs/*.py .
    +

    This copies the files in the programs directory that end with .py to the current directory, signified by the single "." (dot). You must type a space between the py and the final dot.

    +
    +
    +
    +

    +5.2. mv

    +

    mv (short for move) is a Unix command that moves one or more files or directories from one place to another. The original file is deleted, and the new file may have the same or a different name. If possible (i.e. when the original and new files are on the same file system), mv will rename the file instead. Write permission is required on all directories being modified.

    +
    +

    +5.2.1. Conflicting existing file

    +

    In all cases, when a file is moved to have the name of an existing file (in the same directory), the existing file is deleted. If the existing file is not writable but is in a directory that is writable, then the mv command asks for confirmation if possible (i.e. if run from a terminal) before proceeding, unless the -f (force) option is used.

    +
    +
    +

    +5.2.2. Differences with copy and delete

    +

    Note that, usually, when moving files within the same volume, moving (and/or renaming) is not the same as simply copying and then deleting the original. When moving a file, the link is simply removed from the old parent directory and added to the new parent directory. However, the file itself is untouched (i.e. it has the same inodes and resides at the same place on the disk). For example, you cannot copy a file you cannot read, but you can move (and/or rename) it (provided you have write permission to its old and new parent directories). Also, suppose there is a non-empty directory you do not have write permission to. You cannot delete this directory (since you cannot delete its contents); but you can move (and/or rename) it. Also, since moving between filenames on a single volume does not involve copying, it is faster and does not place strain of lots of reads and writes on the disk. Moving files across different volumes, however, does necessitate copying and deleting.

    +
    +
    +

    +5.2.3. Examples

    +
    $ mv myfile mynewfilename    renames a file
    +$ mv myfile otherfilename    renames a file and deletes the existing            file "myfile"
    +$ mv myfile /myfile          moves 'myfile' from the current            directory to the root directory
    +$ mv myfile dir/myfile       moves 'myfile' to 'dir/myfile' relative            to the current directory
    +$ mv myfile dir              same as the previous command (the          filename is implied to be the same)
    +$ mv myfile dir/myfile2      moves 'myfile' to dir and renames it to            'myfile2'
    +$ mv foo bar baz dir         moves multiple files to directory dir
    +$ mv --help                  shows a very concise help about the                syntax of the command
    +$ man mv                     prints an extensive user manual for                'mv' in the terminal
    +

    In all cases, the file or files being moved or renamed can be a directory.

    +

    Note that when the command is called with two arguments (as mv name1 name2 or mv name1 /dir/name2), it can have three different effects, depending on whether name2 does not exist, is an existing file, or is an existing directory. If the user intends to refer to an existing directory, /. (or in some Unix versions / is sufficient) may be appended to the name to force the system to check this. To move a file to a new directory, the directory must be created first.

    +
    +
    +
    +

    +5.3. rm

    +

    rm (short for "remove") is one of several basic Unix command lines that operates on files. It is used to delete files from a filesystem. The data is not actually destroyed. Only the index listing where the file is stored is destroyed, and the storage is made available for reuse. There are undelete utilities that will attempt to reconstruct the index and can bring the file back if the parts were not reused.

    +

    Here's example to remove a file named "foo" from a directory, here shown with the -i option:

    +
    $ rm -i foo
    +remove foo? y
    +
    +

    +5.3.1. Options

    +

    Common options that rm accepts include:

    +<block_quote>
      +
    • -r, which removes directories, removing the contents recursively beforehand (so as not to leave files without a directory to reside in) ("recursive")

    • +
    • -i, which asks for every deletion to be confirmed ("interactive")

    • +
    • -f, which ignores non-existent files and overrides any confirmation prompts ("force")

    • +
    • -v, which shows what is being removed as it happens ("verbose")

    • +
    </block_quote>

    rm is often aliased to "rm -i" so as to avoid accidental deletion of files. If a user still wishes to delete a large number of files without confirmation, they can manually cancel out the -i argument by adding the -f option (as the option specified later on the expanded command line "rm -i -f" takes precedence).

    +

    rm -rf (variously, rm -rf /, rm -rf <title_reference>*</title_reference>, and others) is frequently used in jokes and anecdotes about Unix disasters. The rm -rf variant of the command, if run by a superuser on the root directory, would cause the contents of every writable mounted filesystem on the computer to be deleted.

    +

    rm is often used in conjunction with xargs to supply a list of files to delete:

    +
    xargs rm < filelist
    +

    When rm is used on a symbolic link, it deletes the link, but does not affect the target of the link.

    +
    +
    +

    +5.3.2. Permissions

    +

    Usually, on most filesystems, deleting a file requires write permission on the parent directory (and execute permission, in order to enter the directory in the first place). (Note that, confusingly for beginners, permissions on the file itself are irrelevant. However, GNU rm asks for confirmation if a write-protected file is to be deleted, unless the -f option is used.)

    +

    To delete a directory (with rm -r), one must delete all of its contents recursively. This requires that one must have read and write and execute permission to that directory (if it's not empty) and all non-empty subdirectories recursively (if there are any). The read permissions are needed to list the contents of the directory in order to delete them. This sometimes leads to an odd situation where a non-empty directory cannot be deleted because one doesn't have write permission to it and so cannot delete its contents; but if the same directory were empty, one would be able to delete it.

    +

    If a file resides in a directory with the sticky bit set, then deleting the file requires one to be the owner of the file.

    +
    +
    +
    +
    +

    +6. Command Line Arguments

    +

    In computer command line interfaces, a command line argument is an argument sent to a program being called. In general, a program can take any number of command line arguments, which may be necessary for the program to run, or may even be ignored, depending on the function of that program.

    +

    For example, in Unix and Unix-like environments, an example of a command-line argument is:

    +
    rm file.s
    +

    "file.s" is a command line argument which tells the program rm to remove the file "file.s".

    +

    Programming languages such as C, C++ and Java allow a program to interpret the command line arguments by handling them as string parameters in the main function.

    +

    A command line option or simply option (also known as a command line parameter, flag, or a switch) is an indication by a user that a computer program should change its default output.

    +

    Long options are introduced via "--", and are typically whole words. For example, ls --long --classify --all. Arguments to long options are provided with "=", as ls --block-size=1024. Some Unix programs use long options with single dashes, for example MPlayer as in mplayer -nosound.

    +

    Linux also uses "--" to terminate option lists. For example, an attempt to delete a file called -file1 by using rm -file1 may produce an error, since rm may interpret -file1 as a command line switch. Using rm -- -file1 removes ambiguity.

    +
    +
    +

    +7. Basic Text Processing

    +
    +

    +7.1. head

    +

    head is a program on Unix and Unix-like systems used to display the first few lines of a text file or piped data. The command syntax is:

    +
    $ head [options] <file_name>
    +

    By default, head will print the first 10 lines of its input to the standard output. The number of lines printed may be changed with a command line option. The following example shows the first 20 lines of filename:

    +
    $ head -n 20 filename
    +

    This displays the first 5 lines of all files starting with foo:

    +
    $ head -n 5 foo*
    +

    Some versions omit the n and just let you say -5.

    +
    +

    +7.1.1. Flags

    +
    -c <x number of bytes> Copy first x number of bytes.
    +

    Other options: sed

    +

    Many early versions of Unix did not have this command, and so documentation and books had sed do this job:

    +
    sed 5q foo
    +

    This says to print every line (implicit), and quit after the fifth.

    +
    +
    +
    +

    +7.2. tail

    +

    tail is a program on Unix and Unix-like systems used to display the last few lines of a text file or piped data.

    +

    The command-syntax is:

    +
    $ tail [options] <file_name>
    +

    By default, tail will print the last 10 lines of its input to the standard output. With command line options the number of lines printed and the printing units (lines, blocks or bytes) may be changed. The following example shows the last 20 lines of filename:

    +
    $ tail -n 20 filename
    +

    This example shows the last 15 bytes of all files starting with foo:

    +
    $ tail -c 15 foo*
    +

    This example shows all lines of filename from the second line onwards:

    +
    $ tail -n +2 filename
    +

    Using an older syntax (still used in Sun Solaris as the -n option is not supported), the last 20 lines and the last 50 bytes of filename can be shown with the following command:

    +
    $ tail -20 filename
    +$ tail -50c filename
    +

    However this syntax is now obsolete and does not conform with the POSIX 1003.1-2001 standard. Even if still supported in current versions, when used with other options (like -f, see below), these switches could not work at all.

    +
    +

    +7.2.1. File monitoring

    +

    tail has a special command line option -f (follow) that allows a file to be monitored. Instead of displaying the last few lines and exiting, tail displays the lines and then monitors the file. As new lines are added to the file by another process, tail updates the display. This is particularly useful for monitoring log files. The following command will display the last 10 lines of messages and append new lines to the display as new lines are added to messages:

    +
    $ tail -f /var/adm/messages
    +

    To interrupt tail while it is monitoring, break-in with Ctrl+C. This command can be run "in the background" with &, see job control.

    +

    If you have a command's result to monitor, you can use the watch command.

    +
    +
    +
    +

    +7.3. cut

    +

    In computing, cut is a Unix command line utility which is used to extract sections from each line of input — usually from a file.

    +

    Extraction of line segments can typically be done by bytes (-b), characters (-c), or fields (-f) separated by a delimiter (-d — the tab character by default). A range must be provided in each case which consists of one of N, N-M, N- (N to the end of the line), or -M (beginning of the line to M), where N and M are counted from 1 (there is no zeroth value). Since version 6, an error is thrown if you include a zeroth value. Prior to this the value was ignored and assumed to be 1.

    +

    Assuming a file named file containing the lines:

    +
    foo:bar:baz:qux:quux
    +one:two:three:four:five:six:seven
    +alpha:beta:gamma:delta:epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +

    To output the fourth through tenth characters of each line:

    +
    $ cut -c 4-10 file
    +

    This gives the output:

    +
    :bar:ba
    +:two:th
    +ha:beta
    +

    To output the fifth field through the end of the line of each line using the colon character as the field delimiter:

    +
    $ cut -d : -f 5- file
    +

    This gives the output:

    +
    quux
    +five:six:seven
    +epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +
    +
    +

    +7.4. paste

    +

    paste is a Unix command line utility which is used to join files horizontally (parallel merging) by outputting lines consisting of the sequentially corresponding lines of each file specified, separated by tabs, to the standard output. It is effectively the horizontal equivalent to the utility cat command which operates on the vertical plane of two or more files.

    +

    To paste several columns of data together into the file www from files who, where, and when:

    +
    $ paste who where when > www
    +

    If the files contain:

    +++++
    +

    This creates the file named www containing:

    +
    Batman            GothamCity       January 3
    +Trillian          Andromeda        February 4
    +Jeeves            London           March 19
    +
    +
    +
    +

    +8. Shell Meta Characters

    +

    Unix recognizes certain special characters, called "meta characters," as command directives. The shell meta characters are recognized anywhere they appear in the command line, even if they are not surrounded by blank space. For that reason, it is safest to only use the characters A-Z, a-z, 0-9, and the period, dash, and underscore characters when naming files and directories on Unix. If your file or directory has a shell meta character in the name, you will find it difficult to use the name in a shell command.

    +

    The shell meta characters include:

    +

    / < > ! $ % ^ & * | { } [ ] " ' ` ~ ;

    +

    Different shells may differ in the meta characters recognized.

    +

    As an example,

    +
    $ ls file.*
    +

    run on a directory containing the files file, file.c, file.lst, and myfile would list the files file.c and file.lst. However,:

    +
    $ ls file.?
    +

    run on the same directory would only list file.c because the ? only matches one character, no more, no less. This can save you a great deal of typing time. For example, if there is a file called california_cornish_hens_with_wild_rice and no other files whose names begin with 'c', you could view the file without typing the whole name by typing this:

    +
    $ more c*
    +

    because the c* matches that long file name.

    +

    Filenames containing metacharacters can pose many problems and should never be intentionally created. If you do find that you've created a file with metacharacters, and you would like to remove it, you have three options. You may use wildcards to match metacharacter, use the to directly enter the filename, or put the command in double quotes (except in the case of double quotes within the file name, these must be captured with one of the first two methods). For example, deleting a file named <title_reference>"``*`|more</title_reference>"` can be accomplished with:

    +
    $ rm ??more
    +

    or:

    +
    $ rm $\backslash$*$\backslash$|more
    +

    or:

    +
    $ rm ''*|more''
    +
    +
    +

    +9. Looking At Files

    +
    +

    +9.1. cat

    +

    The cat command is a standard Unix program used to concatenate and display files. The name is from "catenate", a synonym of concatenate.

    +

    The Single Unix Specification specifies the behavior that the contents of each of the files given in sequence as arguments will be written to the standard output in the same sequence, and mandates one option, -u, where each byte is printed as it is read.

    +

    If the filename is specified as -, then cat will read from standard input at that point in the sequence. If no files are specified, cat will read from standard input entered.

    +
    +

    +9.1.1. Jargon File Definition

    +

    The Jargon File version 4.4.7 lists this as the definition of cat:

    +
    1. To spew an entire file to the screen or some other output sink without
    +     pause (syn. blast).
    +
    +2. By extension, to dump large amounts of data at an unprepared target or
    +     with no intention of browsing it carefully. Usage: considered silly.
    +     Rare outside Unix sites. See also dd, BLT.
    +
    +     Among Unix fans, *cat(1)* is considered an excellent example of
    +     user-interface design, because it delivers the file contents without
    +     such verbosity as spacing or headers between the files, and because
    +     it does not require the files to consist of lines of text, but works
    +     with any sort of data.
    +
    +     Among Unix critics, *cat(1)* is considered the canonical example of
    +     bad user-interface design, because of its woefully unobvious name.
    +     It is far more often used to blast a single file to standard output
    +     than to concatenate two or more files. The name cat for the former
    +     operation is just as unintuitive as, say, LISP's cdr.
    +
    +     Of such oppositions are holy wars made...
    +
    +
    +

    +9.1.2. Useless Use of 'cat'

    +

    UUOC (from comp.unix.shell on Usenet) stands for “Useless Use of cat”. As it is observed on comp.unix.shell, “The purpose of cat is to concatenate (or 'catenate') files. If it's only one file, concatenating it with nothing at all is a waste of time, and costs you a process.”

    +

    Nevertheless one sees people doing:

    +
    $ cat file | some_command and its args ...
    +

    instead of the equivalent and cheaper:

    +
    <file some_command and its args ...
    +

    or (equivalently and more classically):

    +
    some_command and its args ... <file
    +

    Since 1995, occasional awards for UUOC have been given out. The activity of fixing instances of UUOC is sometimes called 'demoggification'.

    +

    Amongst many, it is still considered safer to use cat for such cases given that the < and > keys are next to each other in many popular keyboard mappings. While the risk might be low, the impact of using > instead of < can be high and prohibitive.

    +
    +
    +

    +9.1.3. zcat

    +

    zcat is a Unix program similar to cat, that decompresses individual files and concatenates them to standard output. Traditionally zcat operated on files compressed by compress but today it is usually able to operate on gzip or even bzip2 archives. On such systems, it is equivalent to gunzip -c

    +
    +
    +
    +

    +9.2. more

    +

    In computing, more is a command to view (but not modify) the contents of a text file one screen at a time (terminal pager). It is available on Unix and Unix-like systems, DOS, OS/2 and Microsoft Windows. Programs of this sort are called pagers.

    +
    +

    +9.2.1. Usage

    +

    The command-syntax is:

    +
    $ more [options] [file_name]
    +

    If no file name is provided, more looks for input from stdin.

    +

    Once more has obtained input, it displays as much as can fit on the current screen and waits for user input to advance, with the exception that a form feed (^L) will also cause more to wait at that line, regardless of the amount of text on the screen. In the lower-left corner of the screen is displayed the text "--More--" and a percentage, representing the percent of the file that more has paged through. (This percentage includes the text displayed on the current screen.) When more reaches the end of a file (100%) it exits. The most common methods of navigating through a file are Enter, which advances the output by one line, and Space, which advances the output by one screen.

    +

    There are also other commands that can be used while navigating through the document; consult more's man page for more details.

    +

    Options are typically entered before the file name, but can also be entered in the environment variable $MORE. Options entered in the actual command line will override those entered in the $MORE environment variable. Available options may vary between Unix systems.

    +
    +
    +
    +

    +9.3. less

    +

    less is a terminal pager program on Unix, Windows and Unix-like systems used to view (but not change) the contents of a text file one screen at a time. It is similar to more, but has the extended capability of allowing both forward and backward navigation through the file. Unlike most Unix text editors/viewers, less does not need to read the entire file before starting, resulting in faster load times with large files.

    +
    +

    +9.3.1. Usage

    +

    less can be invoked with options to change its behaviour, for example, the number of lines to display on the screen. A few options vary depending on the operating system. While less is displaying the file, various commands can be used to navigate through the file. These commands are based on those used by both more and vi. It is also possible to search for character patterns in the file.

    +

    By default, less displays the contents of the file to the standard output (one screen at a time). If the file name argument is omitted, it displays the contents from standard input (usually the output of another command through a pipe). If the output is redirected to anything other than a terminal, for example a pipe to another command, less behaves like cat.

    +

    The command-syntax is:

    +
    $ less [options] file_name
    +
    +
    +

    +9.3.2. Frequently Used Options

    +<block_quote>
      +
    • -g: Highlights just the current match of any searched string.

    • +
    • -I: Case-insensitive searches.

    • +
    • -M: Shows more detailed prompt, including file position.

    • +
    • -N: Shows line numbers (useful for source code viewing).

    • +
    • -S: Disables line wrap ("chop long lines"). Long lines can be seen by side scrolling.

    • +
    • -?: Shows help.

    • +
    </block_quote>
    +
    +
    +

    +9.3.3. Frequently Used Commands

    +<block_quote>
      +
    • [Arrows]/[Page Up]/[Page Down]/[Home]/[End]: Navigation.

    • +
    • [Space bar]: Next page.

    • +
    • b: Previous page.

    • +
    • ng: Jump to line number n. Default is the start of the file.

    • +
    • nG: Jump to line number n. Default is the end of the file.

    • +
    • /pattern: Search for pattern. Regular expressions can be used.

    • +
    • '^ or g: Go to start of file.

    • +
    • '$ or G: Go to end of file.

    • +
    • s: Save current content (got from another program like grep) in a file.

    • +
    • =: File information.

    • +
    • h: Help.

    • +
    • q: Quit.

    • +
    </block_quote>
    +
    +
    +

    +9.3.4. Examples

    +
    $ less -M readme.txt                     #Read "readme.txt."
    +$ less +F /var/log/mail.log              #Follow mode for log
    +$ file * | less                          #Easier file analysis.
    +$ grep -i void *.c | less -I -p void     #Case insensitive search                                                         for "void" in all .c files
    +
    +
    +
    +
    +

    +10. Directory Structure

    +

    In the File Hierarchy Standard (FHS) all files and directories appear under the root directory "/", even if they are stored on different physical devices. Note however that some of these directories may or may not be present on a Unix system depending on whether certain subsystems, such as the X Window System, are installed.

    +

    The majority of these directories exist in all UNIX operating systems and are generally used in much the same way; however, the descriptions here are those used specifically for the FHS, and are not considered authoritative for platforms other than Linux.

    ++++
    +
    +

    +10.1. man hier

    +

    This is the manual page on the UNIX filesystem. The syntax for this is:

    +
    $ man hier
    +
    +
    +

    +10.2. ls -l

    +

    Shows you huge amounts of information (permissions, owners, size, and when last modified) for folders and files. The syntax is

    +
    $ ls -l
    +

    This can be done after entering the required directory.

    +
    +
    +
    +

    +11. Permissions and Ownership

    +
    +

    +11.1. chmod

    +

    The chmod command (abbreviated from 'change mode') is a shell command and C language function in Unix and Unix-like environments. When executed, it can change file system modes of files and directories. The modes include permissions and special modes.A chmod command first appeared in AT&T Unix version 1, and is still used today on Unix-like machines.

    +
    +

    +11.1.1. Usage

    +

    The chmod command options are specified like this:

    +
    $ chmod [options] mode[,mode] file1 [file2 ...]
    +

    To view what the permissions currently are, type:

    +
    $ ls -l file
    +
    +
    +

    +11.1.2. Command line options

    +

    The chmod command has a number of command line options that affect its behavior. The most common options are:

    +<block_quote>
      +
    • -R: Changes the modes of directories and files recursively

    • +
    • -v: Verbose mode; lists all files as they are being processed

    • +
    </block_quote>
    +
    +11.1.2.1. Symbolic modes
    +

    To the chmod utility, all permissions and special modes are represented by its mode parameter. One way to adjust the mode of files or directories is to specify a symbolic mode. The symbolic mode is composed of three components, which are combined to form a single string of text:

    +
    $ chmod [references][operator][modes] file1 ...
    +

    The references (or classes) are used to distinguish the users to whom the permissions apply. If no references are specified it defaults to “all” but modifies only the permissions allowed by the umask. The references are represented by one or more of the following letters:

    +++++
    +

    The chmod program uses an operator to specify how the modes of a file should be adjusted. The following operators are accepted:

    ++++
    +

    The modes indicate which permissions are to be granted or taken away from the specified classes. There are three basic modes which correspond to the basic permissions:

    +++++
    +

    The combination of these three components produces a string that is understood by the chmod command. Multiple changes can be specified by separating multiple symbolic modes with commas.

    +
    +
    +
    +11.1.2.2. Symbolic examples
    +

    Add the 'read' and 'write' permissions to the 'user' and 'group' classes of a directory:

    +
    $ chmod ug+rw mydir
    +$ ls -ld mydir
    +drw-rw----   2 starwars  yoda  96 Dec 8 12:53 mydir
    +

    For a file, remove write permissions for all classes:

    +
    $ chmod a-w myfile
    +$ ls -l myfile
    +-r-xr-xr-x   2 starwars  yoda 96 Dec 8 12:53 myfile
    +

    Set the permissions for the u*ser and the *g*roup to read and execute only (no write permission) on *mydir.

    +
    $ chmod ug=rx mydir
    +$ ls -ld mydir
    +dr-xr-x---   2 starwars  yoda 96 Dec 8 12:53 mydir
    +
    +
    +
    +11.1.2.3. Octal numbers
    +

    The chmod command also accepts three and four-digit octal numbers representing modes. Using a three-digit octal number to set the modes of a file named myfile :

    +
    $ chmod 664 myfile
    +$ ls -l myfile
    +-rw-rw-r--  1   57 Jul  3 10:13  myfile
    +

    Since the setuid, setgid and sticky bits are not set, this is equivalent to:

    +
    $ chmod 0664 myfile
    +
    +
    +
    +11.1.2.4. Special modes
    +

    The chmod command is also capable of changing the additional permissions or special modes of a file or directory. The symbolic modes use s to represent the setuid and setgid modes, and t to represent the sticky mode. The modes are only applied to the appropriate classes, regardless of whether or not other classes are specified.

    +

    Most operating systems support the specification of special modes using octal modes, but some do not. On these systems, only the symbolic modes can be used.

    +
    +
    +
    +
    +
    +

    +12. Redirection and Piping

    +

    In computing, redirection is a function common to most command-line interpreters, including the various Unix shells that can redirect standard streams to user-specified locations.

    +

    Programs do redirection with the dup2(2) system call, or its less-flexible but higher-level stdio analogues, freopen(3) and popen(3).

    +
    +

    +12.1. Redirecting standard input and standard output

    +

    Redirection is usually implemented by placing certain characters between commands. Typically, the syntax of these characters is as follows:

    +
    $ command1 > file1
    +

    executes command1, placing the output in file1. Note that this will truncate any existing data in file1. To append output to the end of the file, use the >> operator.:

    +
    $ command1 < file1
    +

    executes command1, using file1 as the source of input (as opposed to the keyboard).:

    +
    $ command1 < infile > outfile
    +

    combines the two capabilities: command1 reads from infile and writes to outfile

    +
    +
    +

    +12.2. Piping

    +

    Programs can be run together such that one program reads the output from another with no need for an explicit intermediate file: +A pipeline of three programs run on a text terminal:

    +
    $ command1 | command2
    +

    executes command1, using its output as the input for command2 (commonly called piping, since the "|" character is known as a "pipe").

    +

    This is equivalent to using two redirects and a temporary file:

    +
    $ command1 > tempfile
    +$ command2 < tempfile
    +$ rm tempfile
    +

    A good example for command piping is combining echo with another command to achieve something interactive in a non-interactive shell, e.g.:

    +
    $ echo -e "user\npass" | ftp localhost
    +

    This runs the ftp client with input user, press return, then pass.

    +
    +
    +

    +12.3. Redirecting to and from the standard file handles

    +

    In Unix shells derived from the original Bourne shell, the first two actions can be further modified by placing a number (the file descriptor) immediately before the character; this will affect which stream is used for the redirection. The Unix standard I/O streams are:

    +++++
    +

    For example:

    +
    $ command1 2> file1
    +

    executes command1, directing the standard error stream to file1.

    +

    In shells derived from csh (the C shell), the syntax instead appends the & character to the redirect characters, thus achieving a similar result.

    +

    Another useful capability is to redirect one standard file handle to another. The most popular variation is to merge standard error into standard output so error messages can be processed together with (or alternately to) the usual output. Example:

    +
    $ find / -name .profile > results 2>&1
    +

    will try to find all files named .profile. Executed without redirection, it will output hits to stdout and errors (e.g. for lack of privilege to traverse protected directories) to stderr. If standard output is directed to file results, error messages appear on the console. To see both hits and error messages in file results, merge stderr (handle 2) into stdout (handle 1) using 2>&1 .

    +

    It's possible use 2>&1 before ">" but it doesn't work. In fact, when the interpreter reads 2>&1, it doesn't know yet where standard output is redirected and then standard error isn't merged.

    +

    If the merged output is to be piped into another program, the file merge sequence 2>&1 must precede the pipe symbol, thus:

    +
    $ find / -name .profile 2>&1 | less
    +

    A simplified form of the command:

    +
    $ command > file 2>&1
    +

    is:

    +
    $ command &>file
    +

    or:

    +
    $command >&file
    +
    +
    +

    +12.4. Chained pipelines

    +

    The redirection and piping tokens can be chained together to create complex commands. For example:

    +
    $ ls | grep '\.sh' | sort > shlist
    +

    lists the contents of the current directory, where this output is filtered to only contain lines which contain .sh, sort this resultant output lexicographically, and place the final output in shlist. This type of construction is used very commonly in shell scripts and batch files.

    +
    +
    +

    +12.5. Redirect to multiple outputs

    +

    The standard command tee can redirect output from a command to several destinations.

    +
    $ ls -lrt | tee xyz
    +

    This directs the file list output to both standard output as well as to the file xyz.

    +
    +
    +
    +

    +13. More Text Processing

    +
    +

    +13.1. grep

    +

    grep is a command line text search utility originally written for Unix. The name is taken from the first letters in global / regular expression / print, a series of instructions for the ed text editor. The grep command searches files or standard input globally for lines matching a given regular expression, and prints them to the program's standard output.

    +
    +

    +13.1.1. Usage

    +

    This is an example of a common grep usage:

    +
    $ grep apple fruitlist.txt
    +

    In this case, grep prints all lines containing 'apple' from the file fruitlist.txt, regardless of word boundaries; therefore lines containing 'pineapple' or 'apples' are also printed. The grep command is case sensitive by default, so this example's output does not include lines containing 'Apple' (with a capital A) unless they also contain 'apple'.

    +

    Like most Unix commands, grep accepts command line arguments to change this and many other behaviors. For example:

    +
    $ grep -i apple fruitlist.txt
    +

    This prints all lines containing 'apple' regardless of capitalization. The '-i' argument tells grep to be case insensitive, or to ignore case.

    +

    To print all lines containing 'apple' as a word ('pineapple' and 'apples' will not match):

    +
    $ grep -w apple fruitlist.txt
    +

    Regular expressions can be used to match more complicated queries.

    +
    +
    +13.1.1.1. Variations
    +

    There are countless implementations and derivatives of grep available for many operating systems. Early variants of grep included egrep and fgrep. The former applies an extended regular expression syntax that was added to Unix after Ken Thompson's original regular expression implementation. The latter searches for any of a list of 'fixed' strings using the Aho-Corasick algorithm. These variants are embodied in most modern grep implementations as command-line switches (and standardized as -E and -F in POSIX). In such combined implementations, grep may also behave differently depending on the name by which it is invoked, allowing fgrep, egrep, and grep to be links to the same program.

    +

    pcregrep is an implementation of grep that uses Perl regular expression syntax.

    +

    Other commands contain the word 'grep' to indicate that they search (usually for regular expression matches). The pgrep utility, for instance, displays the processes whose names match a given regular expression.

    +
    +
    +
    +
    +

    +13.2. tr

    +

    tr (abbreviated from translate or transliterate) is a command in Unix-like operating systems.

    +

    When executed, the program reads from the standard input and writes to the standard output. It takes as parameters two sets of characters, and replaces occurrences of the characters in the first set with the corresponding elements from the other set. For example,

    +
    $ tr 'abcd' 'jkmn'
    +

    maps 'a' to 'j', 'b' to 'k', 'c' to 'm', and 'd' to 'n'.

    +

    Sets of characters may be abbreviated by using character ranges. The previous example could be written:

    +
    $ tr 'a-d' 'jkmn'
    +

    In POSIX compliant versions of tr the set represented by a character range depends on the locale's collating order, so it is safer to avoid character ranges in scripts that might be executed in a locale different from that in which they were written. Ranges can often be replaced with POSIX character sets such as [:alpha:].

    +

    The -c flag complements the first set of characters.

    +
    $ tr -cd '[:alnum:]'
    +

    therefore removes all non-alphanumeric characters.

    +

    The -s flag causes tr to compress sequences of identical adjacent characters in its output to a single token. For example,

    +
    $ tr -s '\n' '\n'
    +

    replaces sequences of one or more newline characters with a single newline.

    +

    The -d flag causes tr to delete all tokens of the specified set of characters from its input. In this case, only a single character set argument is used. The following command removes carriage return characters, thereby converting a file in DOS/Windows format to one in Unix format.

    +
    $ tr -d '\r'
    +

    Most versions of tr, including GNU tr and classic Unix tr, operate on single byte characters and are not Unicode compliant. An exception is the Heirloom Toolchest implementation, which provides basic Unicode support.

    +

    Ruby and Perl also have an internal tr operator, which operates analogously. Tcl's string map command is more general in that it maps strings to strings while tr maps characters to characters.

    +
    +
    +
    +

    +14. Elementary Regex

    +

    In computing, regular expressions provide a concise and flexible means for identifying strings of text of interest, such as particular characters, words, or patterns of characters. A regular expression (often shortened to regex or regexp) is written in a formal language that can be interpreted by a regular expression processor, a program that either serves as a parser generator or examines text and identifies parts that match the provided specification.

    +

    Regular expressions are used by many text editors, utilities, and programming languages to search and manipulate text based on patterns. For example, Perl, Ruby and Tcl have a powerful regular expression engine built directly into their syntax. Several utilities provided by Unix distributions—including the editor ed and the filter grep — were the first to popularize the concept of regular expressions.

    +

    Traditional Unix regular expression syntax followed common conventions but often differed from tool to tool. The IEEE POSIX Basic Regular Expressions (BRE) standard (released alongside an alternative flavor called Extended Regular Expressions or ERE) was designed mostly for backward compatibility with the traditional (Simple Regular Expression) syntax but provided a common standard which has since been adopted as the default syntax of many Unix regular expression tools, though there is often some variation or additional features. Many such tools also provide support for ERE syntax with command line arguments.

    +

    In the BRE syntax, most characters are treated as literals — they match only themselves (i.e., a matches "a"). The exceptions, listed below, are called metacharacters or metasequences.

    ++++
    +
    +

    +14.1. Lazy quantification

    +

    The standard quantifiers in regular expressions are greedy, meaning they match as much as they can, only giving back as necessary to match the remainder of the regex. For example, someone new to regexes wishing to find the first instance of an item between < and > symbols in this example:

    +
    Another whale explosion occurred on <January 26>, <2004>.
    +

    ...would likely come up with the pattern <.*>, or similar. However, this pattern will actually return "<January 26>, <2004>" instead of the "<January 26>" which might be expected, because the <title_reference>*</title_reference> quantifier is greedy — it will consume as many characters as possible from the input, and "January 26>, <2004" has more characters than "January 26".

    +

    Though this problem can be avoided in a number of ways (e.g., by specifying the text that is not to be matched: <[^>]*>), modern regular expression tools allow a quantifier to be specified as lazy (also known as non-greedy, reluctant, minimal, or ungreedy) by putting a question mark after the quantifier (e.g., <.*?>), or by using a modifier which reverses the greediness of quantifiers (though changing the meaning of the standard quantifiers can be confusing). By using a lazy quantifier, the expression tries the minimal match first. Though in the previous example lazy matching is used to select one of many matching results, in some cases it can also be used to improve performance when greedy matching would require more backtracking.

    +
    +
    +
    +

    +15. One Liners

    +

    A one-liner is textual input to the command-line of an operating system shell that performs some function in just one line of input.

    +

    The one liner can be

    +<block_quote>
      +
    1. An expression written in the language of the shell.

    2. +
    3. The invocation of an interpreter together with program source for the interpreter to run.

    4. +
    5. The invocation of a compiler together with source to compile and +instructions for executing the compiled program.

    6. +
    </block_quote>

    Certain dynamic scripting languages such as AWK, sed, and perl have traditionally been adept at expressing one-liners. Specialist shell interpreters such as these Unix shells or the Windows PowerShell, allow for the construction of powerful one-liners.

    +

    The use of the phrase one-liner has been widened to also include program-source for any language that does something useful in one line.

    +

    The word One-liner has two references in the index of the book The AWK Programming Language (the book is often referred to by the abbreviation TAPL). It explains the programming language AWK, which is part of the Unix operating system. The authors explain the birth of the One-liner paradigm with their daily work on early Unix machines:

    +
    “The 1977 version had only a few built-in variables and predefined functions. It was designed for writing short programs [...] Our model was that an invocation would be one or two lines long, typed in and used immediately. Defaults were chosen to match this style [...] We, being the authors, knew how the language was supposed to be used, and so we only wrote one-liners.”
    +

    Notice that this original definition of a One-liner implies immediate execution of the program without any compilation. So, in a strict sense, only source code for interpreted languages qualifies as a One-liner. But this strict understanding of a One-liner was broadened in 1985 when the IOCCC introduced the category of Best One Liner for C, which is a compiled language.

    +

    The TAPL book contains 20 examples of One-liners (A Handful of Useful awk One-Liners) at the end of the book's first chapter.

    +

    Here are the first few of them:

    +<block_quote>
      +
    1. +

      Print the total number of input lines:

      +

      END { print NR }

      +
    2. +
    3. +

      Print the tenth input line:

      +

      NR == 10

      +
    4. +
    5. +

      Print the last field of every input line:

      +

      { print $NF }

      +
    6. +
    </block_quote>

    One-liners are also used to show off the differential expressive power of programming languages. Frequently, one-liners are used to demonstrate programming ability. Contests are often held to see who can create the most exceptional one-liner.

    +

    The following example is a C program (a winning entry in the "Best one-liner" category of the IOCCC, here split to two lines for presentation).:

    +
    main(int c,char**v){return!m(v[1],v[2]);}m(char*s,char*t){return
    +*t-42?*s?63==*t|*s==*t&&m(s+1,t+1):!*t:m(s,t+1)||*s&&m(s+1,t);}
    +

    This one-liner program is a glob pattern matcher. It understands the glob characters '*' meaning 'zero or more characters' and '?' meaning exactly one character, just like most Unix shells.

    +

    Run it with two args, the string and the glob pattern. The exit status is 0 (shell true) when the pattern matches, 1 otherwise. The glob pattern must match the whole string, so you may want to use * at the beginning and end of the pattern if you are looking for something in the middle. Examples:

    +
    $ prog foo 'f??'; echo $?
    +
    +$ prog 'best short program' '??st*o**p?*'; echo $?
    +

    Here is a one line shell script to show directories:

    +
    $ ls -R | grep ":$" | sed -e 's/:$//' -e 's/[^-][^\/]*\//--/g' -e 's/^/   /' -e 's/-/|/'
    +
    +
    +
    + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/ch5func.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/ch5func.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1138 @@ + + + +Functions + + + + + + +
    +
    + +
    +
    +

    +
    +
    + +
    +

    +1. Introducing Linux

    +

    (Attribution : A significant chunk of the content under this section is based on data from Wikipedia and the Linux Documentation Project)

    +

    Linux (usually pronounced ˈlɪnəks') is a generic term referring to Unix-like computer operating systems based on the Linux kernel, where a kernel is the intermediate layer between the hardware and the applications. The kernel is, on an abstract level, the core of (most) operating systems, that manages the various system resources. The development of the Linux OS is considered the basis for Free and Open Source Software (FOSS) collaboration since typically the underlying source code can be used, modified freely, and redistributed by anyone under the terms of the GNU (a recursive acronym for "GNU's Not Unix!") Global Public License (GPL) and other free software licences. This freedom to access and reuse various components of a system, is one of the primary reasons for the popularity of Linux.

    +

    Linux is installed on a variety of computer hardware, that include mobile phones, embedded devices and supercomputers, but is infamous for its use in servers.

    +

    The name "Linux" comes from the Linux kernel, originally written in 1991 by Linus Torvalds. The rest of the system usually comprises components such as the Apache HTTP Server, the X Window System, the GNOME and KDE desktop environments, and utilities and libraries from the GNU Project (announced in 1983 by Richard Stallman). Commonly-used applications with desktop Linux systems include the Mozilla Firefox web-browser and the OpenOffice.org office application suite. The GNU contribution is the basis for the Free Software Foundation's preferred name GNU/Linux. The kernel's mascot is a penguin named "Tux". Mozilla Firefox and OpenOffice.org are open-source projects which can be run on most Operating Systems, including proprietary ones.

    +
    +

    +1.1. Historical Background

    +
    +

    +1.1.1. Events leading to the creation

    +
      +
    • The Unix operating system was developed in the 1960s and released for public use in 1970. Its accessibility and portability caused it to be widely adopted, copied and modified by academic institutions and businesses. Its design became influential to authors of other systems. Other free operating systems include the Berkeley Software Distribution (BSD), developed at the University of California at Berkeley, and MINIX which was released by Andrew S. Tanenbaum. The development and adoption of BSD and MINIX were limited due to various reasons, and this lack of a widely-adopted and free kernel triggered Linus Torvalds into starting his project.

    • +
    • In 1983, Richard Stallman started the GNU project with the goal of creating a free UNIX-like operating system. As part of this work, he wrote the GNU General Public License (GPL). By the early 1990s there was almost enough available software to create a full operating system. However, the GNU kernel, called Hurd, failed to attract enough attention from developers leaving GNU incomplete.

    • +
    +
    +
    +

    +1.1.2. The Creation of Linux

    +

    In 1991, Linus Torvalds began a project at the University of Helsinki that later became the Linux kernel. It was initially a terminal (command-line) emulator, which Torvalds used to access the large UNIX servers of the university. He wrote the program targeting just the hardware he was using and independent of an operating system because he wanted to use the functions of his computer with an 80386 processor. Development was done on Minix using the GNU C compiler. This application is still the main choice for compiling Linux today (although the code can be built with other compilers, such as the Intel C Compiler).

    +

    Torvalds continues to direct the development of the kernel. Stallman heads the Free Software Foundation, which in turn supports the GNU components. Finally, individuals and corporations develop third-party non-GNU components, which constitute a vast body of work and including kernel modules, and user applications and libraries. Linux vendors and communities combine and distribute the kernel, GNU components, and non-GNU components, with additional package management software in the form of Linux distributions.

    +
    +
    +
    +

    +1.2. Design and Implications

    +

    A Linux-based system is a modular Unix-like operating system, deriving much of its basic design from principles established in Unix earlier. Such a system uses a monolithic kernel, called the Linux kernel, which handles process control, networking, and peripheral and file system access. Device drivers are integrated directly with the kernel. Separate projects that interface with the kernel provide much of the system's higher-level functionality. The GNU userland is an important part of most Linux-based systems, providing the most common implementation of the C library, a popular shell, and many of the common Unix tools which carry out many basic operating system tasks. The graphical user interface (or GUI) used by most Linux systems is based on the "X Window System".

    +
    +

    +1.2.1. User Interface

    +

    Users can control a Linux-based system through a command line interface (or CLI), a graphical user interface (or GUI), or through controls attached to the associated hardware (this is common for embedded systems). For desktop systems, the default mode is usually the GUI. On desktop machines, "KDE", "GNOME" and "Xfce" are the most popular user interfaces,though a variety of additional user interfaces exist. Most popular user interfaces run on top of the "X Window System" (or X), which enables a graphical application running on one machine to be displayed and controlled from another in a network.

    +

    A Linux system also provides a CLI of some sort through a shell, which is the traditional way of interacting with a Unix system. A Linux distribution specialized for servers may use the CLI as its only interface. A “headless system” (system run without even a monitor) can be controlled by the command line via a remote-control protocol such as SSH or telnet. The CLI is particularly suited for automation of repetitive or delayed tasks, and provides very simple inter-process communication. A graphical terminal emulator program is often used to access the CLI from a Linux desktop.

    +
    +
    +

    +1.2.2. Development

    +

    The primary difference between Linux and many other popular contemporary operating systems is that the Linux kernel and other components are free and open source software. Linux is not the only such operating system, although it is by far the most widely used. Some free and open source software licenses are based on the principle of "copyleft", a kind of reciprocity: any work derived from a copyleft piece of software must also be copyleft itself. The most common free software license, the GNU GPL, is a form of copyleft, and is used for the Linux kernel and many of the components from the GNU project.

    +

    Linux based distributions are intended by developers for interoperability with other operating systems and established computing standards. Linux systems adhere to POSIX, SUS, ISO and ANSI standards where possible, although to date only one Linux distribution has been POSIX.1 certified, Linux-FT.Free software projects, although developed in a collaborative fashion, are often produced independently of each other. The fact that the software licenses explicitly permit redistribution, however, provides a basis for larger scale projects that collect the software produced by stand-alone projects and make it available all at once in the form of a Linux distribution.

    +

    A Linux distribution, commonly called a "distro", is a project that manages a remote collection of system software and application software packages available for download and installation through a network connection. This allows the user to adapt the operating system to his/her specific needs. Distributions are maintained by individuals, loose-knit teams, volunteer organizations, and commercial entities. A distribution can be installed using a CD that contains distribution-specific software for initial system installation and configuration. A package manager such as Synaptic or YAST allows later package upgrades and installations. A distribution is responsible for the default configuration of the installed Linux kernel, general system security, and more generally integration of the different software packages into a coherent whole.

    +
    +
    +

    +1.2.3. Community

    +

    A distribution is largely driven by its developer and user communities. Some vendors develop and fund their distributions on a volunteer basis. Examples include Debian and the Debian-based, Ubuntu. Others maintain a community version of their commercial distributions, as Red Hat does with Fedora.

    +

    In many cities and regions, local associations known as Linux Users Groups (LUGs) seek to promote their preferred distribution and by extension free software. They hold meetings and provide free demonstrations, training, technical support, and operating system installation to new users. Many Internet communities also provide support to Linux users and developers. Most distributions and free software / open source projects have IRC (Internet Relay Chat) chatrooms or newsgroups. Online forums are another means for support. Linux distributions host mailing lists; commonly there will be a specific topic such as usage or development for a given list. All these can be found simply by running an appropriate search on Google.

    +

    Although Linux distributions are generally available without charge, several large corporations sell, support, and contribute to the development of the components of the system and of free software. These include Dell, IBM, HP, Oracle, Sun Microsystems, Novell, Nokia. A number of corporations, notably Red Hat, have built their entire business around Linux distributions.

    +
    +
    +

    +1.2.4. Can I make a profit out of running a business involving Linux?

    +

    The answer is, "Yes!". The free software licenses, on which the various software packages of a distribution built on the Linux kernel are based, explicitly accommodate and encourage commercialization; the relationship between a Linux distribution as a whole and individual vendors may be seen as symbiotic. One common business model of commercial suppliers is charging for support, especially for business users. A number of companies also offer a specialized business version of their distribution, which adds proprietary support packages and tools to administer higher numbers of installations or to simplify administrative tasks. Another business model is to give away the software in order to sell hardware. Examples of corporations that are extensively (and sometimes exclusively) open-source and Linux-powered , with successful revenue generation models involving these, are Google, SUN, Mozilla, etc.

    +
    +
    +

    +1.2.5. Programming on Linux

    +

    Most Linux distributions support dozens of programming languages. The most common collection of utilities for building both Linux applications and operating system programs is found within the GNU toolchain, which includes the GNU Compiler Collection (GCC) and the GNU build system. Amongst others, GCC provides compilers for Ada, C, C++, Java, and Fortran. The Linux kernel itself is written to be compiled with GCC. Proprietary compilers for Linux include the Intel C++ Compiler, Sun Studio, and IBM XL C/C++ Compiler.

    +

    Most distributions also include support for PHP, Perl, Ruby, Python and other dynamic languages. Examples of languages that are less common, but still supported, are C# via the Mono project, sponsored by Novell, and Scheme. A number of Java Virtual Machines and development kits run on Linux, including the original Sun Microsystems JVM (HotSpot), and IBM's J2SE RE, as well as many open-source projects like Kaffe.

    +

    The two main frameworks for developing graphical applications are those of GNOME and KDE. These projects are based on the GTK+ and Qt widget toolkits, respectively, which can also be used independently of the larger framework. Both support a wide variety of languages. There are a number of Integrated Development Environments (IDEs) available including Anjuta, Code::Blocks, Eclipse, KDevelop, Lazarus, MonoDevelop, NetBeans, and Omnis Studio while the long-established editors Vim and Emacs remain popular.

    +
    +
    +
    +

    +1.3. Reasons for Using Linux

    +
    • Linux is free:

    +

    As in "free beer". Linux can be downloaded in its entirety from the Internet completely for free. No registration fees, no costs per user, free updates, and freely available source code in case you want to change the behavior of your system. +Most of all, Linux is free as in "free speech": +The license commonly used is the GNU Public License (GPL). The license says that anybody who may want to do so, has the right to change Linux and eventually to redistribute a changed version, on the one condition that the code is still available after redistribution. In practice, you are free to grab a kernel image and sell the new code, as long as your customers can still have a copy of that code.

    +
    • Linux is portable to any hardware platform:

    +

    A vendor, who wants to sell a new type of computer and who does not know what kind of OS his/her new machine will run, can take a Linux kernel and make it work on his/her hardware, because documentation related to this activity is freely available.

    +
    • Linux was made to keep on running:

    +

    As with UNIX, a Linux system expects to run without rebooting all the time. That is why a lot of tasks are being executed at night or scheduled automatically for other times, resulting in higher availability during busier periods and a more balanced use of the hardware. This property allows for Linux to be applicable to environments where people do not have the time or the possibility to control their systems constantly.

    +
    • Linux is secure and versatile:

    +

    The security model used in Linux is based on the UNIX idea of security, which is known to be robust and of proven quality. But Linux is not only safe from attacks from the Internet: it will adapt equally to other situations, utilizing the same high standards for security.

    +
    • Linux is scalable:

    +

    From a Palmtop with 2 MB of memory to a petabyte storage cluster with hundreds of nodes: add or remove the appropriate packages and Linux fits all. One does not need a supercomputer anymore,because you can use Linux to do big things using the building blocks provided with the system. If one wants to do little things, such as making an operating system for an embedded processor or just recycling your old 486, Linux will do that as well.

    +
    • The Linux OS and Linux applications have very short debug−times:

    +

    Because Linux has been developed and tested by thousands of people, both errors and people to fix them are found very quickly. It often happens that there are only a couple of hours between discovery and fixing of a bug.

    +
    +
    +
    +

    +2. Getting Started

    +
    +

    +2.1. Logging in, activating the user interface and logging out

    +

    In order to work on a Linux system directly, one needs to provide a user name and password. You always need to authenticate to the system. Most PC−based Linux systems have two basic modes for a system to run in: either quick and clean in text console mode,which includes with mouse, multitasking and multi−user features, or in graphical console mode, which looks better but eats more system resources.

    +
    +

    +2.1.1. Graphical Mode

    +

    This is the default nowadays on most desktop computers. You know you will be connecting to the system using graphical mode when you are first asked for your user name, and then to type your password.

    +

    To log in, make sure the mouse pointer is in the login window, provide your user name and password to the system and click OK or press Enter. +It is generally considered a bad idea to connect (graphically) using the root user name, the system adminstrator's account, since the use of graphics includes running a lot of extra programs, in root's case with a lot of extra permissions. To keep all risks as low as possible, use a normal user account to connect graphically. But there are enough risks to keep this in mind as a general advice, for all use of the root account: only log in as root when extra privileges are required.

    +

    After entering your user name/password combination, it can take a little while before the graphical environment is started, depending on the CPU speed of your computer, on the software you use and on your personal settings.

    +

    To continue, you will need to open a terminal window or xterm for short (X being the name for the underlying software supporting the graphical environment). This program can be found in the Applications−>Utilities->System Tools or Internet menu, depending on what window manager you are using. There might be icons that you can use as a shortcut to get an xterm window as well, and clicking the right mouse button on the desktop background will usually present you with a menu containing a terminal window application.

    +

    While browsing the menus, you will notice that a lot of things can be done without entering commands via the keyboard. For most users, the good old point−n−click method of dealing with the computer will do. But for those who want to enter the "heart" of the system, a tool stronger than a mouse will be required to handle the various tasks. This tool is the shell, and when in graphical mode, we activate our shell by opening a terminal window.

    +

    A terminal window should always show a command prompt when you open one. This terminal shows a standard prompt, which displays the user's login name, and the current working directory, represented by the twiddle (~)

    +

    Another common form for a prompt is this one: +[

    +
    +

    user@host
    +

    dir]

    +

    In the above example, user will be your login name, hosts the name of the machine you are working on, and dir an indication of your current location in the file system. Prompts can display all kinds of information, but they are not part of the commands you are giving to your system. To disconnect from the system in graphical mode, you need to close all terminal windows and other applications. After that, hit the logout icon or find Log Out in the menu. Closing everything is not really necessary, and the system can do this for you, but session management might put all currently open applications back on your screen when you connect again, which takes longer and is not always the desired effect. However, this behavior is configurable.

    +

    When you see the login screen again, asking to enter user name and password, logout was successful.

    +
    +
    +

    +2.1.2. Text Mode

    +

    One is in text mode when the whole screen is black, showing (in most cases white) characters. A text mode login screen typically shows some information about the machine you are working on, the name of the machine and a prompt waiting for you to log in.

    +

    The login is different from a graphical login, in that you have to hit the Enter key after providing your user name, because there are no buttons on the screen that you can click with the mouse. Then you should type your password, followed by another Enter. You will not see any indication that you are entering something, not even an asterisk, and you won't see the cursor move. But this is normal on Linux and is done for security +reasons.

    +

    When the system has accepted you as a valid user, you may get some more information, called the message of the day, which can be anything. Additionally, it is popular on UNIX systems to display a fortune cookie, which contains some general wise or unwise (this is up to you) thoughts. After that, you will be given a shell, indicated with the same prompt that you would get in graphical mode.

    +

    Also in text mode: log in as root only to do setup and configuration that absolutely requires administrator privileges, such as adding users, installing software packages, and performing network and other system configuration. Once you are finished, immediately leave the special account and resume your work as a non−privileged user.

    +

    Logging out is done by entering the logout command, followed by Enter. You are successfully disconnected from the system when you see the login screen again.Don't power−off the computer after logging out. It is not meant to be shut off without application of the proper procedures for halting the system. Powering it off without going through the halting process might cause severe damage!

    +
    +
    +
    +
    +

    +3. Basic Commands

    +
    +

    +3.1. ls

    +

    When invoked without any arguments, ls lists the files in the current working directory. A directory that is not the current working directory can be specified and ls will list the files there. The user also may specify any list of files and directories. In this case, all files and all contents of specified directories will be listed. The name ls is derived from list segments which was used in earlier systems.

    +

    Files whose names start with "." are not listed, unless the -a flag is specified or the files are specified explicitly.

    +

    Without options, ls displays files in a bare format. This bare format however makes it difficult to establish the type, permissions, and size of the files. The most common options to reveal this information or change the list of files are:

    +<block_quote>
      +
    • -l long format, displaying Unix file types, permissions, number of hard links, owner, group, size, date, and filename

    • +
    • -F appends a character revealing the nature of a file, for example, * for an executable, or / for a directory. Regular files have no suffix.

    • +
    • -a lists all files in the given directory, including those whose names start with "." (which are hidden files in Unix). By default, these files are excluded from the list.

    • +
    • -R recursively lists subdirectories. The command ls -R / would therefore list all files.

    • +
    • -d shows information about a symbolic link or directory, rather than about the link's target or listing the contents of a directory.

    • +
    • -t sort the list of files by modification time.

    • +
    • -h print sizes in human readable format. (e.g., 1K, 234M, 2G, etc.)

    • +
    </block_quote>

    In some environments, providing the option --color (for GNU ls) or -G (FreeBSD ls) causes ls to highlight different types of files with different colors, instead of with characters as -F would. To determine what color to use for a file, GNU ls checks the Unix file type, the file permissions, and the file extension, while FreeBSD ls checks only the Unix file type and file permissions.:

    +
    $ ls
    +jeeves.rst psmith.html blandings.html
    +$ ls -l
    +drwxr--r--   1 plum  editors   4096  jeeves
    +-rw-r--r--   1 plum  editors  30405  psmith
    +-r-xr-xr-x   1 plum  plum      8460  blandings
    +

    Here "$" actually is the beginning of the prompt. This is typical in most Unix-based systems.

    +
    +
    +

    +3.2. date

    +

    The Unix date command displays the time and date. The super-user can use it to set the system clock.

    +

    With no options, the date command displays the current date and time, including the abbreviated day name, abbreviated month name, day of the month, the time separated by colons, the timezone name, and the year. For example:

    +
    $date
    +Tue Sep  8 12:01:45 IST 2009
    +

    On some systems to set the current date and time to September 8, 2004 01:22 you type:

    +
    $date --set="20040908 01:22"
    +

    In order to view the various options for the date command, type:

    +
    $man date
    +

    This will take you to the "Manual" page comprising of all the details on the date command. You can return to the terminal from the "man" page by pressing the Esc key in the keyboard and typing ":q" in that order.

    +
    +
    +

    +3.3. cd

    +

    This stands for "change directory". When one wants to go up to the parent directory, bypassing the tree of directories one has entered, “ cd ..” can be used.

    +

    One dot '.' represents the current directory while two dots '..' represent the parent directory.

    +

    “ cd -” will return you to the previous directory (a bit like an “undo”).

    +

    You can also use cd absolute path or cd relative path (see below):

    +

    Absolute paths:

    +<block_quote>

    An “ absolute path” is easily recognised from the leading forward slash, /. The / means that you start at the top level directory and continue down.

    </block_quote>

    For example to get to /boot/grub you would type:

    +
    $cd /boot/grub
    +

    This is an absolute path because you start at the top of the hierarchy and go downwards from there (it doesn't matter where in the filesystem you were when you typed the command).

    +

    Relative paths:

    +<block_quote>

    A “ relative path” doesn't have a preceding slash. Use a relative path when you start from a directory below the top level directory structure. This is dependent on where you are in the filesystem.

    +

    For example if you are in root's home directory and want to get to /root/music, you type:

    +
    $ cd music
    </block_quote>

    Please note that there is no / using the above cd command. Using a / would cause this to be an absolute path, working from the top of the hierarchy downward.

    +
    +
    +

    +3.4. who

    +

    The standard Unix command who displays a list of users who are currently logged into a computer.

    +

    The who command is related to the command w, which provides the same information but also displays additional data and statistics.:

    +
    $who
    +beeblebrox tty7         2009-09-08 10:50 (:0)
    +beeblebrox pts/0        2009-09-08 11:25 (:0.0)
    +dumbledore pts/1        2009-09-08 18:11 (potter.xyz.in)
    +beeblebrox pts/2        2009-09-08 18:53 (:0.0)
    +

    The command can be invoked with the arguments am i or am I (so it is invoked as who am i or * who am I*), showing information about the current terminal only (see the -m option below, of which this invocation is equivalent).

    +

    In order to find out the various options that can be appended to the who command, check the man page by typing out the following in the terminal:

    +
    $man who
    +

    This will take you to the "Manual" page containing details about the who command

    +
    +
    +

    +3.5. mkdir

    +

    This command is used to make a new directory. Normal usage is as straightforward as follows:

    +
    $mkdir name_of_directory
    +

    Where name_of_directory is the name of the directory one wants to create. When typed as above (ie. normal usage), the new directory would be created within the current directory. On Unix, multiple directories can be specified, and mkdir will try to create all of them.

    +
    +

    +3.5.1. Options

    +

    On Unix-like operating systems, mkdir takes options. Three of the most common options are:

    +<block_quote>
      +
    • -p: will also create all directories leading up to the given directory that do not exist already. If the given directory already exists, ignore the error.

    • +
    • -v: display each directory that mkdir creates. Most often used with -p.

    • +
    • -m: specify the octal permissions of directories created by mkdir.

    • +
    </block_quote>

    -p is most often used when using mkdir to build up complex directory hierarchies, in case a necessary directory is missing or already there. -m is commonly used to lock down temporary directories used by shell scripts.

    +
    +
    +

    +3.5.2. Examples

    +

    An example of -p in action is:

    +
    $mkdir -p /tmp/a/b/c
    +

    If /tmp/a exists but /tmp/a/b does not, mkdir will create /tmp/a/b before creating /tmp/a/b/c.

    +

    And an even more powerful command, creating a full tree at once (this however is a Shell extension, nothing mkdir does itself):

    +
    $mkdir -p tmpdir/{trunk/sources/{includes,docs},branches,tags}
    +

    This will create:

    +<definition_list><definition_list_item><term>tmpdir - branches</term><definition>
      +
    • tag

    • +
    • <definition_list><definition_list_item><term>trunk - sources - includes</term><definition>
      • docs

      </definition>
      </definition_list_item>
      </definition_list>
    • +
    </definition>
    </definition_list_item>
    </definition_list>
    +
    +
    +
    +
    +

    +4. Getting Help

    +
    +

    +4.1. apropos and whatis

    +

    This is a command to search the manual pages files in Unix and Unix-like operating systems.

    +
    $ apropos grep
    +egrep       egrep (1)       Search a file for a pattern using full regular expressions
    +fgrep       fgrep (1)       Search a file for a fixed-character string
    +fmlgrep     fmlgrep (1)     Search a file for a pattern
    +grep        grep (1)        Search a file for a pattern
    +gzgrep      gzgrep (1)      Search a possibly compressed file for a regular expression
    +nisgrep     nismatch (1)    Utilities for searching NIS+ tables
    +pgrep       pgrep (1)       Find or signal a process by name or other attribute
    +zgrep       zgrep (1)       Search a possibly compressed file for a regular expression
    +...
    +

    In this example, the user uses apropos to search for the string "grep", and apropos returns the indicated man pages that include the term "grep".

    +

    A short index of explanations for commands is available using the whatis command, like in the examples below:

    +
    $whatis ls
    +ls (1)           - list directory contents
    +

    This displays short information about a command, and the first section in the collection of man pages that contains an appropriate page.

    +

    If you don't know where to get started and which man page to read, apropos gives more information. Say that you do not know how to start a browser, then you could enter the following command:

    +
    $apropos browser
    +gmusicbrowser (1)    - Jukebox for large collections of audio files
    +infobrowser (1)      - read Info documents
    +libsmbclient (7)     - An extension library for browsers and that               can be used...
    +opera (1)            - a standards-compliant graphical Web browser
    +sensible-browser (1) - sensible editing, paging, and web browsing
    +smbtree (1)          - A text based smb network browser
    +tvtk_doc (1)         - A GUI based TVTK documentation search browser.
    +viewres (1)          - graphical class browser for Xt
    +w3m (1)              - a text based Web browser and pager
    +www-browser (1)      - a text based Web browser and pager
    +...
    +
    +
    +

    +4.2. man

    +

    Man pages (short for "manual pages") are the extensive documentation that comes preinstalled with almost all substantial Unix and Unix-like operating systems. The Unix command used to display them is man. Each page is a self-contained document.

    +

    To read a manual page for a Unix command, one can use:

    +
    $ man <command_name>
    +

    at a shell prompt; for example, "man ftp". In order to simplify navigation through the output, man generally uses the less terminal pager.

    +

    Pages are traditionally referred to using the notation "name(section)"; for example, ftp(1). The same page name may appear in more than one section of the manual, this can occur when the names of system calls, user commands, or macro packages coincide. Two examples are man(1) and man(7), or exit(2) and exit(3). The syntax for accessing the non-default manual section varies between different man implementations. On Linux and BSD, for example, the syntax for reading *printf(3) is:

    +
    $man 3 printf
    +

    Another example:

    +
    $man man
    +

    The previous example will take you to the "Manual" page entry about manual pages!

    +
    +

    +4.2.1. Layout

    +

    All man pages follow a common layout that is optimized for presentation on a simple ASCII text display, possibly without any form of highlighting or font control. Sections present may include:

    +<definition_list><definition_list_item><term>NAME</term><definition>

    The name of the command or function, followed by a one-line description of what it does.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SYNOPSIS</term><definition>

    In the case of a command, you get a formal description of how to run it and what command line options it takes. For program functions, a list of the parameters the function takes and which header file contains its definition. For experienced users, this may be all the documentation they need.

    </definition>
    </definition_list_item>
    <definition_list_item><term>DESCRIPTION</term><definition>

    A textual description of the functioning of the command or function.

    </definition>
    </definition_list_item>
    <definition_list_item><term>EXAMPLES</term><definition>

    Some examples of common usage.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SEE ALSO</term><definition>

    A list of related commands or functions.

    </definition>
    </definition_list_item>
    </definition_list>

    Other sections may be present, but these are not well standardized across man pages. Common examples include: OPTIONS, EXIT STATUS, ENVIRONMENT, KNOWN BUGS, FILES, AUTHOR, REPORTING BUGS, HISTORY and COPYRIGHT.

    +

    These days virtually every Unix command line application comes with its man page, and many Unix users perceive a lack of man pages as a sign of low quality; indeed, some projects, such as Debian, go out of their way to write man pages for programs lacking one. Few alternatives to man have enjoyed much popularity, with the possible exception of the GNU project's "info" system, an early and simple hypertext system.

    +

    However, the format of a single page for each application, the lack of classification within the sections and the relatively unsophisticated formatting facilities have motivated the development of alternative documentation systems, such as the previously mentioned "info" system.

    +

    Most Unix GUI applications (particularly those built using the GNOME and KDE development environments) now provide end-user documentation in HTML and include embedded HTML viewers such as yelp for reading the help within the application.

    +

    Usually the man pages are written in English. Translations into other languages can be also available on the system.

    +

    The default format of the man pages is troff, with either the macro package man (appearance oriented) or on some systems mdoc (semantic oriented). This makes it possible to typeset a man page to PostScript, PDF and various other formats for viewing or printing.

    +
    +
    +
    +

    +4.3. info

    +

    info is a software utility which forms a hypertextual, multipage documentation and help viewer working on a command line interface, useful when there is no GUI available.

    +

    The syntax is

    +
    $ info <command_name>
    +

    info processes info files, which are Texinfo formatted files, and presents the documentation as a tree, with simple commands to traverse the tree and to follow cross references. For instance

    +<block_quote>
      +
    • n goes to the next page.

    • +
    • p goes to the previous page.

    • +
    • u goes to the upper page.

    • +
    • l goes to the last(visited) node

    • +
    • To follow a cross reference, the cursor can be moved over a link (a word preceded by a <title_reference>*</title_reference>) and enter pressed.

    • +
    </block_quote>

    info was initially written for use with GNU/Linux and then ported to other Unix-like operating systems.

    +
    +
    +

    +4.4. --help

    +

    Most GNU commands support the --help, which gives a short explanation about how to use the command and a list of available options. Below is the output of this option with the cat command:

    +
    $ userprompt@host: cat --help
    +Usage: cat [OPTION] [FILE]...
    +Concatenate FILE(s), or standard input, to standard output.
    +
    +  -A, --show-all           equivalent to -vET
    +  -b, --number-nonblank    number nonempty output lines
    +  -e                       equivalent to -vE
    +  -E, --show-ends          display $ at end of each line
    +  -n, --number             number all output lines
    +  -s, --squeeze-blank      suppress repeated empty output lines
    +  -t                       equivalent to -vT
    +  -T, --show-tabs          display TAB characters as ^I
    +  -u                       (ignored)
    +  -v, --show-nonprinting   use ^ and M- notation, except for LFD and              TAB
    +  --help     display this help and exit
    +  --version  output version information and exit
    +
    +With no FILE, or when FILE is -, read standard input.
    +
    +Examples:
    +  cat f - g  Output f's contents, then standard input, then g's           contents.
    +  cat        Copy standard input to standard output.
    +
    +Report bugs to <bug-coreutils@gnu.org>.
    +
    +
    +
    +

    +5. Basic file handling

    +
    +

    +5.1. cp

    +

    cp is the command entered in a Unix shell to copy a file from one place to another, possibly on a different filesystem. The original file remains unchanged, and the new file may have the same or a different name.

    +
    +

    +5.1.1. Usage

    +

    To copy a file to another file:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ][ -- ] SourceFile TargetFile
    +

    To copy a file to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -r | -R ] [ -- ] SourceFile ...              TargetDirectory
    +

    To copy a directory to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -- ] { -r | -R }
    +SourceDirectory ... TargetDirectory
    +
    +
    +

    +5.1.2. Flags

    +

    -f (force) – specifies removal of the target file if it cannot be opened for write operations. The removal precedes any copying performed by the cp command.

    +

    -P – makes the cp command copy symbolic links. The default is to follow symbolic links, that is, to copy files to which symbolic links point.

    +

    -i (interactive) – prompts you with the name of a file to be overwritten. This occurs if the TargetDirectory or TargetFile parameter contains a file with the same name as a file specified in the SourceFile or SourceDirectory parameter. If you enter y or the locale's equivalent of y, the cp command continues. Any other answer prevents the cp command from overwriting the file.

    +

    -p (preserve) – duplicates the following characteristics of each SourceFile/SourceDirectory in the corresponding TargetFile and/or TargetDirectory:

    +<block_quote>
      +
    • The time of the last data modification and the time of the last access.

    • +
    • The user ID and group ID (only if it has permissions to do this)

    • +
    • The file permission bits and the SUID and SGID bits.

    • +
    </block_quote>

    -R (recursive) – copy directories (recursively copying all the contents)

    +
    +
    +

    +5.1.3. Examples

    +

    To make a copy of a file in the current directory, enter:

    +
    $ cp prog.c prog.bak
    +

    This copies prog.c to prog.bak. If the prog.bak file does not already exist, the cp command creates it. If it does exist, the cp command replaces it with a copy of the prog.c file.

    +

    To copy a file in your current directory into another directory, enter:

    +
    $ cp zaphod /home/books/hhgg
    +

    This copies the jones file to /home/books/hhgg/zaphod.

    +

    To copy a file to a new file and preserve the modification date, time, and access control list associated with the source file, enter:

    +
    $ cp -p martin_luther_king martin_luther_king.jr
    +

    This copies the martin_luther_king file to the martin_luther_king.jr file. Instead of creating the file with the current date and time stamp, the system gives the martin_luther_king.jr file the same date and time as the martin_luther_king file. The martin_luther_king.jr file also inherits the martin_luther_king file's access control protection.

    +

    To copy all the files in a directory to a new directory, enter:

    +
    $ cp /home/galactica/clients/* /home/hhgg/customers
    +

    This copies only the files in the clients directory to the customers directory.

    +

    To copy a directory, including all its files and subdirectories, to another directory, enter:

    +<block_quote>

    $ cp -R /home/hhgg/clients /home/hhgg/customers

    </block_quote>

    This copies the clients directory, including all its files, subdirectories, and the files in those subdirectories, to the customers/clients directory.

    +

    To copy a specific set of files of any extension to another directory, enter:

    +
    $ cp zaphod arthur ford /home/hhgg/clients
    +

    This copies the zaphod, arthur, and ford files in your current working directory to the /home/hhgg/clients directory.

    +

    To use pattern-matching characters to copy files, enter:

    +
    $ cp programs/*.py .
    +

    This copies the files in the programs directory that end with .py to the current directory, signified by the single "." (dot). You must type a space between the py and the final dot.

    +
    +
    +
    +

    +5.2. mv

    +

    mv (short for move) is a Unix command that moves one or more files or directories from one place to another. The original file is deleted, and the new file may have the same or a different name. If possible (i.e. when the original and new files are on the same file system), mv will rename the file instead. Write permission is required on all directories being modified.

    +
    +

    +5.2.1. Conflicting existing file

    +

    In all cases, when a file is moved to have the name of an existing file (in the same directory), the existing file is deleted. If the existing file is not writable but is in a directory that is writable, then the mv command asks for confirmation if possible (i.e. if run from a terminal) before proceeding, unless the -f (force) option is used.

    +
    +
    +

    +5.2.2. Differences with copy and delete

    +

    Note that, usually, when moving files within the same volume, moving (and/or renaming) is not the same as simply copying and then deleting the original. When moving a file, the link is simply removed from the old parent directory and added to the new parent directory. However, the file itself is untouched (i.e. it has the same inodes and resides at the same place on the disk). For example, you cannot copy a file you cannot read, but you can move (and/or rename) it (provided you have write permission to its old and new parent directories). Also, suppose there is a non-empty directory you do not have write permission to. You cannot delete this directory (since you cannot delete its contents); but you can move (and/or rename) it. Also, since moving between filenames on a single volume does not involve copying, it is faster and does not place strain of lots of reads and writes on the disk. Moving files across different volumes, however, does necessitate copying and deleting.

    +
    +
    +

    +5.2.3. Examples

    +
    $ mv myfile mynewfilename    renames a file
    +$ mv myfile otherfilename    renames a file and deletes the existing            file "myfile"
    +$ mv myfile /myfile          moves 'myfile' from the current            directory to the root directory
    +$ mv myfile dir/myfile       moves 'myfile' to 'dir/myfile' relative            to the current directory
    +$ mv myfile dir              same as the previous command (the          filename is implied to be the same)
    +$ mv myfile dir/myfile2      moves 'myfile' to dir and renames it to            'myfile2'
    +$ mv foo bar baz dir         moves multiple files to directory dir
    +$ mv --help                  shows a very concise help about the                syntax of the command
    +$ man mv                     prints an extensive user manual for                'mv' in the terminal
    +

    In all cases, the file or files being moved or renamed can be a directory.

    +

    Note that when the command is called with two arguments (as mv name1 name2 or mv name1 /dir/name2), it can have three different effects, depending on whether name2 does not exist, is an existing file, or is an existing directory. If the user intends to refer to an existing directory, /. (or in some Unix versions / is sufficient) may be appended to the name to force the system to check this. To move a file to a new directory, the directory must be created first.

    +
    +
    +
    +

    +5.3. rm

    +

    rm (short for "remove") is one of several basic Unix command lines that operates on files. It is used to delete files from a filesystem. The data is not actually destroyed. Only the index listing where the file is stored is destroyed, and the storage is made available for reuse. There are undelete utilities that will attempt to reconstruct the index and can bring the file back if the parts were not reused.

    +

    Here's example to remove a file named "foo" from a directory, here shown with the -i option:

    +
    $ rm -i foo
    +remove foo? y
    +
    +

    +5.3.1. Options

    +

    Common options that rm accepts include:

    +<block_quote>
      +
    • -r, which removes directories, removing the contents recursively beforehand (so as not to leave files without a directory to reside in) ("recursive")

    • +
    • -i, which asks for every deletion to be confirmed ("interactive")

    • +
    • -f, which ignores non-existent files and overrides any confirmation prompts ("force")

    • +
    • -v, which shows what is being removed as it happens ("verbose")

    • +
    </block_quote>

    rm is often aliased to "rm -i" so as to avoid accidental deletion of files. If a user still wishes to delete a large number of files without confirmation, they can manually cancel out the -i argument by adding the -f option (as the option specified later on the expanded command line "rm -i -f" takes precedence).

    +

    rm -rf (variously, rm -rf /, rm -rf <title_reference>*</title_reference>, and others) is frequently used in jokes and anecdotes about Unix disasters. The rm -rf variant of the command, if run by a superuser on the root directory, would cause the contents of every writable mounted filesystem on the computer to be deleted.

    +

    rm is often used in conjunction with xargs to supply a list of files to delete:

    +
    xargs rm < filelist
    +

    When rm is used on a symbolic link, it deletes the link, but does not affect the target of the link.

    +
    +
    +

    +5.3.2. Permissions

    +

    Usually, on most filesystems, deleting a file requires write permission on the parent directory (and execute permission, in order to enter the directory in the first place). (Note that, confusingly for beginners, permissions on the file itself are irrelevant. However, GNU rm asks for confirmation if a write-protected file is to be deleted, unless the -f option is used.)

    +

    To delete a directory (with rm -r), one must delete all of its contents recursively. This requires that one must have read and write and execute permission to that directory (if it's not empty) and all non-empty subdirectories recursively (if there are any). The read permissions are needed to list the contents of the directory in order to delete them. This sometimes leads to an odd situation where a non-empty directory cannot be deleted because one doesn't have write permission to it and so cannot delete its contents; but if the same directory were empty, one would be able to delete it.

    +

    If a file resides in a directory with the sticky bit set, then deleting the file requires one to be the owner of the file.

    +
    +
    +
    +
    +

    +6. Command Line Arguments

    +

    In computer command line interfaces, a command line argument is an argument sent to a program being called. In general, a program can take any number of command line arguments, which may be necessary for the program to run, or may even be ignored, depending on the function of that program.

    +

    For example, in Unix and Unix-like environments, an example of a command-line argument is:

    +
    rm file.s
    +

    "file.s" is a command line argument which tells the program rm to remove the file "file.s".

    +

    Programming languages such as C, C++ and Java allow a program to interpret the command line arguments by handling them as string parameters in the main function.

    +

    A command line option or simply option (also known as a command line parameter, flag, or a switch) is an indication by a user that a computer program should change its default output.

    +

    Long options are introduced via "--", and are typically whole words. For example, ls --long --classify --all. Arguments to long options are provided with "=", as ls --block-size=1024. Some Unix programs use long options with single dashes, for example MPlayer as in mplayer -nosound.

    +

    Linux also uses "--" to terminate option lists. For example, an attempt to delete a file called -file1 by using rm -file1 may produce an error, since rm may interpret -file1 as a command line switch. Using rm -- -file1 removes ambiguity.

    +
    +
    +

    +7. Basic Text Processing

    +
    +

    +7.1. head

    +

    head is a program on Unix and Unix-like systems used to display the first few lines of a text file or piped data. The command syntax is:

    +
    $ head [options] <file_name>
    +

    By default, head will print the first 10 lines of its input to the standard output. The number of lines printed may be changed with a command line option. The following example shows the first 20 lines of filename:

    +
    $ head -n 20 filename
    +

    This displays the first 5 lines of all files starting with foo:

    +
    $ head -n 5 foo*
    +

    Some versions omit the n and just let you say -5.

    +
    +

    +7.1.1. Flags

    +
    -c <x number of bytes> Copy first x number of bytes.
    +

    Other options: sed

    +

    Many early versions of Unix did not have this command, and so documentation and books had sed do this job:

    +
    sed 5q foo
    +

    This says to print every line (implicit), and quit after the fifth.

    +
    +
    +
    +

    +7.2. tail

    +

    tail is a program on Unix and Unix-like systems used to display the last few lines of a text file or piped data.

    +

    The command-syntax is:

    +
    $ tail [options] <file_name>
    +

    By default, tail will print the last 10 lines of its input to the standard output. With command line options the number of lines printed and the printing units (lines, blocks or bytes) may be changed. The following example shows the last 20 lines of filename:

    +
    $ tail -n 20 filename
    +

    This example shows the last 15 bytes of all files starting with foo:

    +
    $ tail -c 15 foo*
    +

    This example shows all lines of filename from the second line onwards:

    +
    $ tail -n +2 filename
    +

    Using an older syntax (still used in Sun Solaris as the -n option is not supported), the last 20 lines and the last 50 bytes of filename can be shown with the following command:

    +
    $ tail -20 filename
    +$ tail -50c filename
    +

    However this syntax is now obsolete and does not conform with the POSIX 1003.1-2001 standard. Even if still supported in current versions, when used with other options (like -f, see below), these switches could not work at all.

    +
    +

    +7.2.1. File monitoring

    +

    tail has a special command line option -f (follow) that allows a file to be monitored. Instead of displaying the last few lines and exiting, tail displays the lines and then monitors the file. As new lines are added to the file by another process, tail updates the display. This is particularly useful for monitoring log files. The following command will display the last 10 lines of messages and append new lines to the display as new lines are added to messages:

    +
    $ tail -f /var/adm/messages
    +

    To interrupt tail while it is monitoring, break-in with Ctrl+C. This command can be run "in the background" with &, see job control.

    +

    If you have a command's result to monitor, you can use the watch command.

    +
    +
    +
    +

    +7.3. cut

    +

    In computing, cut is a Unix command line utility which is used to extract sections from each line of input — usually from a file.

    +

    Extraction of line segments can typically be done by bytes (-b), characters (-c), or fields (-f) separated by a delimiter (-d — the tab character by default). A range must be provided in each case which consists of one of N, N-M, N- (N to the end of the line), or -M (beginning of the line to M), where N and M are counted from 1 (there is no zeroth value). Since version 6, an error is thrown if you include a zeroth value. Prior to this the value was ignored and assumed to be 1.

    +

    Assuming a file named file containing the lines:

    +
    foo:bar:baz:qux:quux
    +one:two:three:four:five:six:seven
    +alpha:beta:gamma:delta:epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +

    To output the fourth through tenth characters of each line:

    +
    $ cut -c 4-10 file
    +

    This gives the output:

    +
    :bar:ba
    +:two:th
    +ha:beta
    +

    To output the fifth field through the end of the line of each line using the colon character as the field delimiter:

    +
    $ cut -d : -f 5- file
    +

    This gives the output:

    +
    quux
    +five:six:seven
    +epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +
    +
    +

    +7.4. paste

    +

    paste is a Unix command line utility which is used to join files horizontally (parallel merging) by outputting lines consisting of the sequentially corresponding lines of each file specified, separated by tabs, to the standard output. It is effectively the horizontal equivalent to the utility cat command which operates on the vertical plane of two or more files.

    +

    To paste several columns of data together into the file www from files who, where, and when:

    +
    $ paste who where when > www
    +

    If the files contain:

    +++++
    +

    This creates the file named www containing:

    +
    Batman            GothamCity       January 3
    +Trillian          Andromeda        February 4
    +Jeeves            London           March 19
    +
    +
    +
    +

    +8. Shell Meta Characters

    +

    Unix recognizes certain special characters, called "meta characters," as command directives. The shell meta characters are recognized anywhere they appear in the command line, even if they are not surrounded by blank space. For that reason, it is safest to only use the characters A-Z, a-z, 0-9, and the period, dash, and underscore characters when naming files and directories on Unix. If your file or directory has a shell meta character in the name, you will find it difficult to use the name in a shell command.

    +

    The shell meta characters include:

    +

    / < > ! $ % ^ & * | { } [ ] " ' ` ~ ;

    +

    Different shells may differ in the meta characters recognized.

    +

    As an example,

    +
    $ ls file.*
    +

    run on a directory containing the files file, file.c, file.lst, and myfile would list the files file.c and file.lst. However,:

    +
    $ ls file.?
    +

    run on the same directory would only list file.c because the ? only matches one character, no more, no less. This can save you a great deal of typing time. For example, if there is a file called california_cornish_hens_with_wild_rice and no other files whose names begin with 'c', you could view the file without typing the whole name by typing this:

    +
    $ more c*
    +

    because the c* matches that long file name.

    +

    Filenames containing metacharacters can pose many problems and should never be intentionally created. If you do find that you've created a file with metacharacters, and you would like to remove it, you have three options. You may use wildcards to match metacharacter, use the to directly enter the filename, or put the command in double quotes (except in the case of double quotes within the file name, these must be captured with one of the first two methods). For example, deleting a file named <title_reference>"``*`|more</title_reference>"` can be accomplished with:

    +
    $ rm ??more
    +

    or:

    +
    $ rm $\backslash$*$\backslash$|more
    +

    or:

    +
    $ rm ''*|more''
    +
    +
    +

    +9. Looking At Files

    +
    +

    +9.1. cat

    +

    The cat command is a standard Unix program used to concatenate and display files. The name is from "catenate", a synonym of concatenate.

    +

    The Single Unix Specification specifies the behavior that the contents of each of the files given in sequence as arguments will be written to the standard output in the same sequence, and mandates one option, -u, where each byte is printed as it is read.

    +

    If the filename is specified as -, then cat will read from standard input at that point in the sequence. If no files are specified, cat will read from standard input entered.

    +
    +

    +9.1.1. Jargon File Definition

    +

    The Jargon File version 4.4.7 lists this as the definition of cat:

    +
    1. To spew an entire file to the screen or some other output sink without
    +     pause (syn. blast).
    +
    +2. By extension, to dump large amounts of data at an unprepared target or
    +     with no intention of browsing it carefully. Usage: considered silly.
    +     Rare outside Unix sites. See also dd, BLT.
    +
    +     Among Unix fans, *cat(1)* is considered an excellent example of
    +     user-interface design, because it delivers the file contents without
    +     such verbosity as spacing or headers between the files, and because
    +     it does not require the files to consist of lines of text, but works
    +     with any sort of data.
    +
    +     Among Unix critics, *cat(1)* is considered the canonical example of
    +     bad user-interface design, because of its woefully unobvious name.
    +     It is far more often used to blast a single file to standard output
    +     than to concatenate two or more files. The name cat for the former
    +     operation is just as unintuitive as, say, LISP's cdr.
    +
    +     Of such oppositions are holy wars made...
    +
    +
    +

    +9.1.2. Useless Use of 'cat'

    +

    UUOC (from comp.unix.shell on Usenet) stands for “Useless Use of cat”. As it is observed on comp.unix.shell, “The purpose of cat is to concatenate (or 'catenate') files. If it's only one file, concatenating it with nothing at all is a waste of time, and costs you a process.”

    +

    Nevertheless one sees people doing:

    +
    $ cat file | some_command and its args ...
    +

    instead of the equivalent and cheaper:

    +
    <file some_command and its args ...
    +

    or (equivalently and more classically):

    +
    some_command and its args ... <file
    +

    Since 1995, occasional awards for UUOC have been given out. The activity of fixing instances of UUOC is sometimes called 'demoggification'.

    +

    Amongst many, it is still considered safer to use cat for such cases given that the < and > keys are next to each other in many popular keyboard mappings. While the risk might be low, the impact of using > instead of < can be high and prohibitive.

    +
    +
    +

    +9.1.3. zcat

    +

    zcat is a Unix program similar to cat, that decompresses individual files and concatenates them to standard output. Traditionally zcat operated on files compressed by compress but today it is usually able to operate on gzip or even bzip2 archives. On such systems, it is equivalent to gunzip -c

    +
    +
    +
    +

    +9.2. more

    +

    In computing, more is a command to view (but not modify) the contents of a text file one screen at a time (terminal pager). It is available on Unix and Unix-like systems, DOS, OS/2 and Microsoft Windows. Programs of this sort are called pagers.

    +
    +

    +9.2.1. Usage

    +

    The command-syntax is:

    +
    $ more [options] [file_name]
    +

    If no file name is provided, more looks for input from stdin.

    +

    Once more has obtained input, it displays as much as can fit on the current screen and waits for user input to advance, with the exception that a form feed (^L) will also cause more to wait at that line, regardless of the amount of text on the screen. In the lower-left corner of the screen is displayed the text "--More--" and a percentage, representing the percent of the file that more has paged through. (This percentage includes the text displayed on the current screen.) When more reaches the end of a file (100%) it exits. The most common methods of navigating through a file are Enter, which advances the output by one line, and Space, which advances the output by one screen.

    +

    There are also other commands that can be used while navigating through the document; consult more's man page for more details.

    +

    Options are typically entered before the file name, but can also be entered in the environment variable $MORE. Options entered in the actual command line will override those entered in the $MORE environment variable. Available options may vary between Unix systems.

    +
    +
    +
    +

    +9.3. less

    +

    less is a terminal pager program on Unix, Windows and Unix-like systems used to view (but not change) the contents of a text file one screen at a time. It is similar to more, but has the extended capability of allowing both forward and backward navigation through the file. Unlike most Unix text editors/viewers, less does not need to read the entire file before starting, resulting in faster load times with large files.

    +
    +

    +9.3.1. Usage

    +

    less can be invoked with options to change its behaviour, for example, the number of lines to display on the screen. A few options vary depending on the operating system. While less is displaying the file, various commands can be used to navigate through the file. These commands are based on those used by both more and vi. It is also possible to search for character patterns in the file.

    +

    By default, less displays the contents of the file to the standard output (one screen at a time). If the file name argument is omitted, it displays the contents from standard input (usually the output of another command through a pipe). If the output is redirected to anything other than a terminal, for example a pipe to another command, less behaves like cat.

    +

    The command-syntax is:

    +
    $ less [options] file_name
    +
    +
    +

    +9.3.2. Frequently Used Options

    +<block_quote>
      +
    • -g: Highlights just the current match of any searched string.

    • +
    • -I: Case-insensitive searches.

    • +
    • -M: Shows more detailed prompt, including file position.

    • +
    • -N: Shows line numbers (useful for source code viewing).

    • +
    • -S: Disables line wrap ("chop long lines"). Long lines can be seen by side scrolling.

    • +
    • -?: Shows help.

    • +
    </block_quote>
    +
    +
    +

    +9.3.3. Frequently Used Commands

    +<block_quote>
      +
    • [Arrows]/[Page Up]/[Page Down]/[Home]/[End]: Navigation.

    • +
    • [Space bar]: Next page.

    • +
    • b: Previous page.

    • +
    • ng: Jump to line number n. Default is the start of the file.

    • +
    • nG: Jump to line number n. Default is the end of the file.

    • +
    • /pattern: Search for pattern. Regular expressions can be used.

    • +
    • '^ or g: Go to start of file.

    • +
    • '$ or G: Go to end of file.

    • +
    • s: Save current content (got from another program like grep) in a file.

    • +
    • =: File information.

    • +
    • h: Help.

    • +
    • q: Quit.

    • +
    </block_quote>
    +
    +
    +

    +9.3.4. Examples

    +
    $ less -M readme.txt                     #Read "readme.txt."
    +$ less +F /var/log/mail.log              #Follow mode for log
    +$ file * | less                          #Easier file analysis.
    +$ grep -i void *.c | less -I -p void     #Case insensitive search                                                         for "void" in all .c files
    +
    +
    +
    +
    +

    +10. Directory Structure

    +

    In the File Hierarchy Standard (FHS) all files and directories appear under the root directory "/", even if they are stored on different physical devices. Note however that some of these directories may or may not be present on a Unix system depending on whether certain subsystems, such as the X Window System, are installed.

    +

    The majority of these directories exist in all UNIX operating systems and are generally used in much the same way; however, the descriptions here are those used specifically for the FHS, and are not considered authoritative for platforms other than Linux.

    ++++
    +
    +

    +10.1. man hier

    +

    This is the manual page on the UNIX filesystem. The syntax for this is:

    +
    $ man hier
    +
    +
    +

    +10.2. ls -l

    +

    Shows you huge amounts of information (permissions, owners, size, and when last modified) for folders and files. The syntax is

    +
    $ ls -l
    +

    This can be done after entering the required directory.

    +
    +
    +
    +

    +11. Permissions and Ownership

    +
    +

    +11.1. chmod

    +

    The chmod command (abbreviated from 'change mode') is a shell command and C language function in Unix and Unix-like environments. When executed, it can change file system modes of files and directories. The modes include permissions and special modes.A chmod command first appeared in AT&T Unix version 1, and is still used today on Unix-like machines.

    +
    +

    +11.1.1. Usage

    +

    The chmod command options are specified like this:

    +
    $ chmod [options] mode[,mode] file1 [file2 ...]
    +

    To view what the permissions currently are, type:

    +
    $ ls -l file
    +
    +
    +

    +11.1.2. Command line options

    +

    The chmod command has a number of command line options that affect its behavior. The most common options are:

    +<block_quote>
      +
    • -R: Changes the modes of directories and files recursively

    • +
    • -v: Verbose mode; lists all files as they are being processed

    • +
    </block_quote>
    +
    +11.1.2.1. Symbolic modes
    +

    To the chmod utility, all permissions and special modes are represented by its mode parameter. One way to adjust the mode of files or directories is to specify a symbolic mode. The symbolic mode is composed of three components, which are combined to form a single string of text:

    +
    $ chmod [references][operator][modes] file1 ...
    +

    The references (or classes) are used to distinguish the users to whom the permissions apply. If no references are specified it defaults to “all” but modifies only the permissions allowed by the umask. The references are represented by one or more of the following letters:

    +++++
    +

    The chmod program uses an operator to specify how the modes of a file should be adjusted. The following operators are accepted:

    ++++
    +

    The modes indicate which permissions are to be granted or taken away from the specified classes. There are three basic modes which correspond to the basic permissions:

    +++++
    +

    The combination of these three components produces a string that is understood by the chmod command. Multiple changes can be specified by separating multiple symbolic modes with commas.

    +
    +
    +
    +11.1.2.2. Symbolic examples
    +

    Add the 'read' and 'write' permissions to the 'user' and 'group' classes of a directory:

    +
    $ chmod ug+rw mydir
    +$ ls -ld mydir
    +drw-rw----   2 starwars  yoda  96 Dec 8 12:53 mydir
    +

    For a file, remove write permissions for all classes:

    +
    $ chmod a-w myfile
    +$ ls -l myfile
    +-r-xr-xr-x   2 starwars  yoda 96 Dec 8 12:53 myfile
    +

    Set the permissions for the u*ser and the *g*roup to read and execute only (no write permission) on *mydir.

    +
    $ chmod ug=rx mydir
    +$ ls -ld mydir
    +dr-xr-x---   2 starwars  yoda 96 Dec 8 12:53 mydir
    +
    +
    +
    +11.1.2.3. Octal numbers
    +

    The chmod command also accepts three and four-digit octal numbers representing modes. Using a three-digit octal number to set the modes of a file named myfile :

    +
    $ chmod 664 myfile
    +$ ls -l myfile
    +-rw-rw-r--  1   57 Jul  3 10:13  myfile
    +

    Since the setuid, setgid and sticky bits are not set, this is equivalent to:

    +
    $ chmod 0664 myfile
    +
    +
    +
    +11.1.2.4. Special modes
    +

    The chmod command is also capable of changing the additional permissions or special modes of a file or directory. The symbolic modes use s to represent the setuid and setgid modes, and t to represent the sticky mode. The modes are only applied to the appropriate classes, regardless of whether or not other classes are specified.

    +

    Most operating systems support the specification of special modes using octal modes, but some do not. On these systems, only the symbolic modes can be used.

    +
    +
    +
    +
    +
    +

    +12. Redirection and Piping

    +

    In computing, redirection is a function common to most command-line interpreters, including the various Unix shells that can redirect standard streams to user-specified locations.

    +

    Programs do redirection with the dup2(2) system call, or its less-flexible but higher-level stdio analogues, freopen(3) and popen(3).

    +
    +

    +12.1. Redirecting standard input and standard output

    +

    Redirection is usually implemented by placing certain characters between commands. Typically, the syntax of these characters is as follows:

    +
    $ command1 > file1
    +

    executes command1, placing the output in file1. Note that this will truncate any existing data in file1. To append output to the end of the file, use the >> operator.:

    +
    $ command1 < file1
    +

    executes command1, using file1 as the source of input (as opposed to the keyboard).:

    +
    $ command1 < infile > outfile
    +

    combines the two capabilities: command1 reads from infile and writes to outfile

    +
    +
    +

    +12.2. Piping

    +

    Programs can be run together such that one program reads the output from another with no need for an explicit intermediate file: +A pipeline of three programs run on a text terminal:

    +
    $ command1 | command2
    +

    executes command1, using its output as the input for command2 (commonly called piping, since the "|" character is known as a "pipe").

    +

    This is equivalent to using two redirects and a temporary file:

    +
    $ command1 > tempfile
    +$ command2 < tempfile
    +$ rm tempfile
    +

    A good example for command piping is combining echo with another command to achieve something interactive in a non-interactive shell, e.g.:

    +
    $ echo -e "user\npass" | ftp localhost
    +

    This runs the ftp client with input user, press return, then pass.

    +
    +
    +

    +12.3. Redirecting to and from the standard file handles

    +

    In Unix shells derived from the original Bourne shell, the first two actions can be further modified by placing a number (the file descriptor) immediately before the character; this will affect which stream is used for the redirection. The Unix standard I/O streams are:

    +++++
    +

    For example:

    +
    $ command1 2> file1
    +

    executes command1, directing the standard error stream to file1.

    +

    In shells derived from csh (the C shell), the syntax instead appends the & character to the redirect characters, thus achieving a similar result.

    +

    Another useful capability is to redirect one standard file handle to another. The most popular variation is to merge standard error into standard output so error messages can be processed together with (or alternately to) the usual output. Example:

    +
    $ find / -name .profile > results 2>&1
    +

    will try to find all files named .profile. Executed without redirection, it will output hits to stdout and errors (e.g. for lack of privilege to traverse protected directories) to stderr. If standard output is directed to file results, error messages appear on the console. To see both hits and error messages in file results, merge stderr (handle 2) into stdout (handle 1) using 2>&1 .

    +

    It's possible use 2>&1 before ">" but it doesn't work. In fact, when the interpreter reads 2>&1, it doesn't know yet where standard output is redirected and then standard error isn't merged.

    +

    If the merged output is to be piped into another program, the file merge sequence 2>&1 must precede the pipe symbol, thus:

    +
    $ find / -name .profile 2>&1 | less
    +

    A simplified form of the command:

    +
    $ command > file 2>&1
    +

    is:

    +
    $ command &>file
    +

    or:

    +
    $command >&file
    +
    +
    +

    +12.4. Chained pipelines

    +

    The redirection and piping tokens can be chained together to create complex commands. For example:

    +
    $ ls | grep '\.sh' | sort > shlist
    +

    lists the contents of the current directory, where this output is filtered to only contain lines which contain .sh, sort this resultant output lexicographically, and place the final output in shlist. This type of construction is used very commonly in shell scripts and batch files.

    +
    +
    +

    +12.5. Redirect to multiple outputs

    +

    The standard command tee can redirect output from a command to several destinations.

    +
    $ ls -lrt | tee xyz
    +

    This directs the file list output to both standard output as well as to the file xyz.

    +
    +
    +
    +

    +13. More Text Processing

    +
    +

    +13.1. grep

    +

    grep is a command line text search utility originally written for Unix. The name is taken from the first letters in global / regular expression / print, a series of instructions for the ed text editor. The grep command searches files or standard input globally for lines matching a given regular expression, and prints them to the program's standard output.

    +
    +

    +13.1.1. Usage

    +

    This is an example of a common grep usage:

    +
    $ grep apple fruitlist.txt
    +

    In this case, grep prints all lines containing 'apple' from the file fruitlist.txt, regardless of word boundaries; therefore lines containing 'pineapple' or 'apples' are also printed. The grep command is case sensitive by default, so this example's output does not include lines containing 'Apple' (with a capital A) unless they also contain 'apple'.

    +

    Like most Unix commands, grep accepts command line arguments to change this and many other behaviors. For example:

    +
    $ grep -i apple fruitlist.txt
    +

    This prints all lines containing 'apple' regardless of capitalization. The '-i' argument tells grep to be case insensitive, or to ignore case.

    +

    To print all lines containing 'apple' as a word ('pineapple' and 'apples' will not match):

    +
    $ grep -w apple fruitlist.txt
    +

    Regular expressions can be used to match more complicated queries.

    +
    +
    +13.1.1.1. Variations
    +

    There are countless implementations and derivatives of grep available for many operating systems. Early variants of grep included egrep and fgrep. The former applies an extended regular expression syntax that was added to Unix after Ken Thompson's original regular expression implementation. The latter searches for any of a list of 'fixed' strings using the Aho-Corasick algorithm. These variants are embodied in most modern grep implementations as command-line switches (and standardized as -E and -F in POSIX). In such combined implementations, grep may also behave differently depending on the name by which it is invoked, allowing fgrep, egrep, and grep to be links to the same program.

    +

    pcregrep is an implementation of grep that uses Perl regular expression syntax.

    +

    Other commands contain the word 'grep' to indicate that they search (usually for regular expression matches). The pgrep utility, for instance, displays the processes whose names match a given regular expression.

    +
    +
    +
    +
    +

    +13.2. tr

    +

    tr (abbreviated from translate or transliterate) is a command in Unix-like operating systems.

    +

    When executed, the program reads from the standard input and writes to the standard output. It takes as parameters two sets of characters, and replaces occurrences of the characters in the first set with the corresponding elements from the other set. For example,

    +
    $ tr 'abcd' 'jkmn'
    +

    maps 'a' to 'j', 'b' to 'k', 'c' to 'm', and 'd' to 'n'.

    +

    Sets of characters may be abbreviated by using character ranges. The previous example could be written:

    +
    $ tr 'a-d' 'jkmn'
    +

    In POSIX compliant versions of tr the set represented by a character range depends on the locale's collating order, so it is safer to avoid character ranges in scripts that might be executed in a locale different from that in which they were written. Ranges can often be replaced with POSIX character sets such as [:alpha:].

    +

    The -c flag complements the first set of characters.

    +
    $ tr -cd '[:alnum:]'
    +

    therefore removes all non-alphanumeric characters.

    +

    The -s flag causes tr to compress sequences of identical adjacent characters in its output to a single token. For example,

    +
    $ tr -s '\n' '\n'
    +

    replaces sequences of one or more newline characters with a single newline.

    +

    The -d flag causes tr to delete all tokens of the specified set of characters from its input. In this case, only a single character set argument is used. The following command removes carriage return characters, thereby converting a file in DOS/Windows format to one in Unix format.

    +
    $ tr -d '\r'
    +

    Most versions of tr, including GNU tr and classic Unix tr, operate on single byte characters and are not Unicode compliant. An exception is the Heirloom Toolchest implementation, which provides basic Unicode support.

    +

    Ruby and Perl also have an internal tr operator, which operates analogously. Tcl's string map command is more general in that it maps strings to strings while tr maps characters to characters.

    +
    +
    +
    +

    +14. Elementary Regex

    +

    In computing, regular expressions provide a concise and flexible means for identifying strings of text of interest, such as particular characters, words, or patterns of characters. A regular expression (often shortened to regex or regexp) is written in a formal language that can be interpreted by a regular expression processor, a program that either serves as a parser generator or examines text and identifies parts that match the provided specification.

    +

    Regular expressions are used by many text editors, utilities, and programming languages to search and manipulate text based on patterns. For example, Perl, Ruby and Tcl have a powerful regular expression engine built directly into their syntax. Several utilities provided by Unix distributions—including the editor ed and the filter grep — were the first to popularize the concept of regular expressions.

    +

    Traditional Unix regular expression syntax followed common conventions but often differed from tool to tool. The IEEE POSIX Basic Regular Expressions (BRE) standard (released alongside an alternative flavor called Extended Regular Expressions or ERE) was designed mostly for backward compatibility with the traditional (Simple Regular Expression) syntax but provided a common standard which has since been adopted as the default syntax of many Unix regular expression tools, though there is often some variation or additional features. Many such tools also provide support for ERE syntax with command line arguments.

    +

    In the BRE syntax, most characters are treated as literals — they match only themselves (i.e., a matches "a"). The exceptions, listed below, are called metacharacters or metasequences.

    ++++
    +
    +

    +14.1. Lazy quantification

    +

    The standard quantifiers in regular expressions are greedy, meaning they match as much as they can, only giving back as necessary to match the remainder of the regex. For example, someone new to regexes wishing to find the first instance of an item between < and > symbols in this example:

    +
    Another whale explosion occurred on <January 26>, <2004>.
    +

    ...would likely come up with the pattern <.*>, or similar. However, this pattern will actually return "<January 26>, <2004>" instead of the "<January 26>" which might be expected, because the <title_reference>*</title_reference> quantifier is greedy — it will consume as many characters as possible from the input, and "January 26>, <2004" has more characters than "January 26".

    +

    Though this problem can be avoided in a number of ways (e.g., by specifying the text that is not to be matched: <[^>]*>), modern regular expression tools allow a quantifier to be specified as lazy (also known as non-greedy, reluctant, minimal, or ungreedy) by putting a question mark after the quantifier (e.g., <.*?>), or by using a modifier which reverses the greediness of quantifiers (though changing the meaning of the standard quantifiers can be confusing). By using a lazy quantifier, the expression tries the minimal match first. Though in the previous example lazy matching is used to select one of many matching results, in some cases it can also be used to improve performance when greedy matching would require more backtracking.

    +
    +
    +
    +

    +15. One Liners

    +

    A one-liner is textual input to the command-line of an operating system shell that performs some function in just one line of input.

    +

    The one liner can be

    +<block_quote>
      +
    1. An expression written in the language of the shell.

    2. +
    3. The invocation of an interpreter together with program source for the interpreter to run.

    4. +
    5. The invocation of a compiler together with source to compile and +instructions for executing the compiled program.

    6. +
    </block_quote>

    Certain dynamic scripting languages such as AWK, sed, and perl have traditionally been adept at expressing one-liners. Specialist shell interpreters such as these Unix shells or the Windows PowerShell, allow for the construction of powerful one-liners.

    +

    The use of the phrase one-liner has been widened to also include program-source for any language that does something useful in one line.

    +

    The word One-liner has two references in the index of the book The AWK Programming Language (the book is often referred to by the abbreviation TAPL). It explains the programming language AWK, which is part of the Unix operating system. The authors explain the birth of the One-liner paradigm with their daily work on early Unix machines:

    +
    “The 1977 version had only a few built-in variables and predefined functions. It was designed for writing short programs [...] Our model was that an invocation would be one or two lines long, typed in and used immediately. Defaults were chosen to match this style [...] We, being the authors, knew how the language was supposed to be used, and so we only wrote one-liners.”
    +

    Notice that this original definition of a One-liner implies immediate execution of the program without any compilation. So, in a strict sense, only source code for interpreted languages qualifies as a One-liner. But this strict understanding of a One-liner was broadened in 1985 when the IOCCC introduced the category of Best One Liner for C, which is a compiled language.

    +

    The TAPL book contains 20 examples of One-liners (A Handful of Useful awk One-Liners) at the end of the book's first chapter.

    +

    Here are the first few of them:

    +<block_quote>
      +
    1. +

      Print the total number of input lines:

      +

      END { print NR }

      +
    2. +
    3. +

      Print the tenth input line:

      +

      NR == 10

      +
    4. +
    5. +

      Print the last field of every input line:

      +

      { print $NF }

      +
    6. +
    </block_quote>

    One-liners are also used to show off the differential expressive power of programming languages. Frequently, one-liners are used to demonstrate programming ability. Contests are often held to see who can create the most exceptional one-liner.

    +

    The following example is a C program (a winning entry in the "Best one-liner" category of the IOCCC, here split to two lines for presentation).:

    +
    main(int c,char**v){return!m(v[1],v[2]);}m(char*s,char*t){return
    +*t-42?*s?63==*t|*s==*t&&m(s+1,t+1):!*t:m(s,t+1)||*s&&m(s+1,t);}
    +

    This one-liner program is a glob pattern matcher. It understands the glob characters '*' meaning 'zero or more characters' and '?' meaning exactly one character, just like most Unix shells.

    +

    Run it with two args, the string and the glob pattern. The exit status is 0 (shell true) when the pattern matches, 1 otherwise. The glob pattern must match the whole string, so you may want to use * at the beginning and end of the pattern if you are looking for something in the middle. Examples:

    +
    $ prog foo 'f??'; echo $?
    +
    +$ prog 'best short program' '??st*o**p?*'; echo $?
    +

    Here is a one line shell script to show directories:

    +
    $ ls -R | grep ":$" | sed -e 's/:$//' -e 's/[^-][^\/]*\//--/g' -e 's/^/   /' -e 's/-/|/'
    +
    +
    +
    + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/ch6oop.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/ch6oop.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1138 @@ + + + +Strings and Dicts  + + + + + + +
    +
    + +
    +
    +

    +
    +
    + +
    +

    +1. Introducing Linux

    +

    (Attribution : A significant chunk of the content under this section is based on data from Wikipedia and the Linux Documentation Project)

    +

    Linux (usually pronounced ˈlɪnəks') is a generic term referring to Unix-like computer operating systems based on the Linux kernel, where a kernel is the intermediate layer between the hardware and the applications. The kernel is, on an abstract level, the core of (most) operating systems, that manages the various system resources. The development of the Linux OS is considered the basis for Free and Open Source Software (FOSS) collaboration since typically the underlying source code can be used, modified freely, and redistributed by anyone under the terms of the GNU (a recursive acronym for "GNU's Not Unix!") Global Public License (GPL) and other free software licences. This freedom to access and reuse various components of a system, is one of the primary reasons for the popularity of Linux.

    +

    Linux is installed on a variety of computer hardware, that include mobile phones, embedded devices and supercomputers, but is infamous for its use in servers.

    +

    The name "Linux" comes from the Linux kernel, originally written in 1991 by Linus Torvalds. The rest of the system usually comprises components such as the Apache HTTP Server, the X Window System, the GNOME and KDE desktop environments, and utilities and libraries from the GNU Project (announced in 1983 by Richard Stallman). Commonly-used applications with desktop Linux systems include the Mozilla Firefox web-browser and the OpenOffice.org office application suite. The GNU contribution is the basis for the Free Software Foundation's preferred name GNU/Linux. The kernel's mascot is a penguin named "Tux". Mozilla Firefox and OpenOffice.org are open-source projects which can be run on most Operating Systems, including proprietary ones.

    +
    +

    +1.1. Historical Background

    +
    +

    +1.1.1. Events leading to the creation

    +
      +
    • The Unix operating system was developed in the 1960s and released for public use in 1970. Its accessibility and portability caused it to be widely adopted, copied and modified by academic institutions and businesses. Its design became influential to authors of other systems. Other free operating systems include the Berkeley Software Distribution (BSD), developed at the University of California at Berkeley, and MINIX which was released by Andrew S. Tanenbaum. The development and adoption of BSD and MINIX were limited due to various reasons, and this lack of a widely-adopted and free kernel triggered Linus Torvalds into starting his project.

    • +
    • In 1983, Richard Stallman started the GNU project with the goal of creating a free UNIX-like operating system. As part of this work, he wrote the GNU General Public License (GPL). By the early 1990s there was almost enough available software to create a full operating system. However, the GNU kernel, called Hurd, failed to attract enough attention from developers leaving GNU incomplete.

    • +
    +
    +
    +

    +1.1.2. The Creation of Linux

    +

    In 1991, Linus Torvalds began a project at the University of Helsinki that later became the Linux kernel. It was initially a terminal (command-line) emulator, which Torvalds used to access the large UNIX servers of the university. He wrote the program targeting just the hardware he was using and independent of an operating system because he wanted to use the functions of his computer with an 80386 processor. Development was done on Minix using the GNU C compiler. This application is still the main choice for compiling Linux today (although the code can be built with other compilers, such as the Intel C Compiler).

    +

    Torvalds continues to direct the development of the kernel. Stallman heads the Free Software Foundation, which in turn supports the GNU components. Finally, individuals and corporations develop third-party non-GNU components, which constitute a vast body of work and including kernel modules, and user applications and libraries. Linux vendors and communities combine and distribute the kernel, GNU components, and non-GNU components, with additional package management software in the form of Linux distributions.

    +
    +
    +
    +

    +1.2. Design and Implications

    +

    A Linux-based system is a modular Unix-like operating system, deriving much of its basic design from principles established in Unix earlier. Such a system uses a monolithic kernel, called the Linux kernel, which handles process control, networking, and peripheral and file system access. Device drivers are integrated directly with the kernel. Separate projects that interface with the kernel provide much of the system's higher-level functionality. The GNU userland is an important part of most Linux-based systems, providing the most common implementation of the C library, a popular shell, and many of the common Unix tools which carry out many basic operating system tasks. The graphical user interface (or GUI) used by most Linux systems is based on the "X Window System".

    +
    +

    +1.2.1. User Interface

    +

    Users can control a Linux-based system through a command line interface (or CLI), a graphical user interface (or GUI), or through controls attached to the associated hardware (this is common for embedded systems). For desktop systems, the default mode is usually the GUI. On desktop machines, "KDE", "GNOME" and "Xfce" are the most popular user interfaces,though a variety of additional user interfaces exist. Most popular user interfaces run on top of the "X Window System" (or X), which enables a graphical application running on one machine to be displayed and controlled from another in a network.

    +

    A Linux system also provides a CLI of some sort through a shell, which is the traditional way of interacting with a Unix system. A Linux distribution specialized for servers may use the CLI as its only interface. A “headless system” (system run without even a monitor) can be controlled by the command line via a remote-control protocol such as SSH or telnet. The CLI is particularly suited for automation of repetitive or delayed tasks, and provides very simple inter-process communication. A graphical terminal emulator program is often used to access the CLI from a Linux desktop.

    +
    +
    +

    +1.2.2. Development

    +

    The primary difference between Linux and many other popular contemporary operating systems is that the Linux kernel and other components are free and open source software. Linux is not the only such operating system, although it is by far the most widely used. Some free and open source software licenses are based on the principle of "copyleft", a kind of reciprocity: any work derived from a copyleft piece of software must also be copyleft itself. The most common free software license, the GNU GPL, is a form of copyleft, and is used for the Linux kernel and many of the components from the GNU project.

    +

    Linux based distributions are intended by developers for interoperability with other operating systems and established computing standards. Linux systems adhere to POSIX, SUS, ISO and ANSI standards where possible, although to date only one Linux distribution has been POSIX.1 certified, Linux-FT.Free software projects, although developed in a collaborative fashion, are often produced independently of each other. The fact that the software licenses explicitly permit redistribution, however, provides a basis for larger scale projects that collect the software produced by stand-alone projects and make it available all at once in the form of a Linux distribution.

    +

    A Linux distribution, commonly called a "distro", is a project that manages a remote collection of system software and application software packages available for download and installation through a network connection. This allows the user to adapt the operating system to his/her specific needs. Distributions are maintained by individuals, loose-knit teams, volunteer organizations, and commercial entities. A distribution can be installed using a CD that contains distribution-specific software for initial system installation and configuration. A package manager such as Synaptic or YAST allows later package upgrades and installations. A distribution is responsible for the default configuration of the installed Linux kernel, general system security, and more generally integration of the different software packages into a coherent whole.

    +
    +
    +

    +1.2.3. Community

    +

    A distribution is largely driven by its developer and user communities. Some vendors develop and fund their distributions on a volunteer basis. Examples include Debian and the Debian-based, Ubuntu. Others maintain a community version of their commercial distributions, as Red Hat does with Fedora.

    +

    In many cities and regions, local associations known as Linux Users Groups (LUGs) seek to promote their preferred distribution and by extension free software. They hold meetings and provide free demonstrations, training, technical support, and operating system installation to new users. Many Internet communities also provide support to Linux users and developers. Most distributions and free software / open source projects have IRC (Internet Relay Chat) chatrooms or newsgroups. Online forums are another means for support. Linux distributions host mailing lists; commonly there will be a specific topic such as usage or development for a given list. All these can be found simply by running an appropriate search on Google.

    +

    Although Linux distributions are generally available without charge, several large corporations sell, support, and contribute to the development of the components of the system and of free software. These include Dell, IBM, HP, Oracle, Sun Microsystems, Novell, Nokia. A number of corporations, notably Red Hat, have built their entire business around Linux distributions.

    +
    +
    +

    +1.2.4. Can I make a profit out of running a business involving Linux?

    +

    The answer is, "Yes!". The free software licenses, on which the various software packages of a distribution built on the Linux kernel are based, explicitly accommodate and encourage commercialization; the relationship between a Linux distribution as a whole and individual vendors may be seen as symbiotic. One common business model of commercial suppliers is charging for support, especially for business users. A number of companies also offer a specialized business version of their distribution, which adds proprietary support packages and tools to administer higher numbers of installations or to simplify administrative tasks. Another business model is to give away the software in order to sell hardware. Examples of corporations that are extensively (and sometimes exclusively) open-source and Linux-powered , with successful revenue generation models involving these, are Google, SUN, Mozilla, etc.

    +
    +
    +

    +1.2.5. Programming on Linux

    +

    Most Linux distributions support dozens of programming languages. The most common collection of utilities for building both Linux applications and operating system programs is found within the GNU toolchain, which includes the GNU Compiler Collection (GCC) and the GNU build system. Amongst others, GCC provides compilers for Ada, C, C++, Java, and Fortran. The Linux kernel itself is written to be compiled with GCC. Proprietary compilers for Linux include the Intel C++ Compiler, Sun Studio, and IBM XL C/C++ Compiler.

    +

    Most distributions also include support for PHP, Perl, Ruby, Python and other dynamic languages. Examples of languages that are less common, but still supported, are C# via the Mono project, sponsored by Novell, and Scheme. A number of Java Virtual Machines and development kits run on Linux, including the original Sun Microsystems JVM (HotSpot), and IBM's J2SE RE, as well as many open-source projects like Kaffe.

    +

    The two main frameworks for developing graphical applications are those of GNOME and KDE. These projects are based on the GTK+ and Qt widget toolkits, respectively, which can also be used independently of the larger framework. Both support a wide variety of languages. There are a number of Integrated Development Environments (IDEs) available including Anjuta, Code::Blocks, Eclipse, KDevelop, Lazarus, MonoDevelop, NetBeans, and Omnis Studio while the long-established editors Vim and Emacs remain popular.

    +
    +
    +
    +

    +1.3. Reasons for Using Linux

    +
    • Linux is free:

    +

    As in "free beer". Linux can be downloaded in its entirety from the Internet completely for free. No registration fees, no costs per user, free updates, and freely available source code in case you want to change the behavior of your system. +Most of all, Linux is free as in "free speech": +The license commonly used is the GNU Public License (GPL). The license says that anybody who may want to do so, has the right to change Linux and eventually to redistribute a changed version, on the one condition that the code is still available after redistribution. In practice, you are free to grab a kernel image and sell the new code, as long as your customers can still have a copy of that code.

    +
    • Linux is portable to any hardware platform:

    +

    A vendor, who wants to sell a new type of computer and who does not know what kind of OS his/her new machine will run, can take a Linux kernel and make it work on his/her hardware, because documentation related to this activity is freely available.

    +
    • Linux was made to keep on running:

    +

    As with UNIX, a Linux system expects to run without rebooting all the time. That is why a lot of tasks are being executed at night or scheduled automatically for other times, resulting in higher availability during busier periods and a more balanced use of the hardware. This property allows for Linux to be applicable to environments where people do not have the time or the possibility to control their systems constantly.

    +
    • Linux is secure and versatile:

    +

    The security model used in Linux is based on the UNIX idea of security, which is known to be robust and of proven quality. But Linux is not only safe from attacks from the Internet: it will adapt equally to other situations, utilizing the same high standards for security.

    +
    • Linux is scalable:

    +

    From a Palmtop with 2 MB of memory to a petabyte storage cluster with hundreds of nodes: add or remove the appropriate packages and Linux fits all. One does not need a supercomputer anymore,because you can use Linux to do big things using the building blocks provided with the system. If one wants to do little things, such as making an operating system for an embedded processor or just recycling your old 486, Linux will do that as well.

    +
    • The Linux OS and Linux applications have very short debug−times:

    +

    Because Linux has been developed and tested by thousands of people, both errors and people to fix them are found very quickly. It often happens that there are only a couple of hours between discovery and fixing of a bug.

    +
    +
    +
    +

    +2. Getting Started

    +
    +

    +2.1. Logging in, activating the user interface and logging out

    +

    In order to work on a Linux system directly, one needs to provide a user name and password. You always need to authenticate to the system. Most PC−based Linux systems have two basic modes for a system to run in: either quick and clean in text console mode,which includes with mouse, multitasking and multi−user features, or in graphical console mode, which looks better but eats more system resources.

    +
    +

    +2.1.1. Graphical Mode

    +

    This is the default nowadays on most desktop computers. You know you will be connecting to the system using graphical mode when you are first asked for your user name, and then to type your password.

    +

    To log in, make sure the mouse pointer is in the login window, provide your user name and password to the system and click OK or press Enter. +It is generally considered a bad idea to connect (graphically) using the root user name, the system adminstrator's account, since the use of graphics includes running a lot of extra programs, in root's case with a lot of extra permissions. To keep all risks as low as possible, use a normal user account to connect graphically. But there are enough risks to keep this in mind as a general advice, for all use of the root account: only log in as root when extra privileges are required.

    +

    After entering your user name/password combination, it can take a little while before the graphical environment is started, depending on the CPU speed of your computer, on the software you use and on your personal settings.

    +

    To continue, you will need to open a terminal window or xterm for short (X being the name for the underlying software supporting the graphical environment). This program can be found in the Applications−>Utilities->System Tools or Internet menu, depending on what window manager you are using. There might be icons that you can use as a shortcut to get an xterm window as well, and clicking the right mouse button on the desktop background will usually present you with a menu containing a terminal window application.

    +

    While browsing the menus, you will notice that a lot of things can be done without entering commands via the keyboard. For most users, the good old point−n−click method of dealing with the computer will do. But for those who want to enter the "heart" of the system, a tool stronger than a mouse will be required to handle the various tasks. This tool is the shell, and when in graphical mode, we activate our shell by opening a terminal window.

    +

    A terminal window should always show a command prompt when you open one. This terminal shows a standard prompt, which displays the user's login name, and the current working directory, represented by the twiddle (~)

    +

    Another common form for a prompt is this one: +[

    +
    +

    user@host
    +

    dir]

    +

    In the above example, user will be your login name, hosts the name of the machine you are working on, and dir an indication of your current location in the file system. Prompts can display all kinds of information, but they are not part of the commands you are giving to your system. To disconnect from the system in graphical mode, you need to close all terminal windows and other applications. After that, hit the logout icon or find Log Out in the menu. Closing everything is not really necessary, and the system can do this for you, but session management might put all currently open applications back on your screen when you connect again, which takes longer and is not always the desired effect. However, this behavior is configurable.

    +

    When you see the login screen again, asking to enter user name and password, logout was successful.

    +
    +
    +

    +2.1.2. Text Mode

    +

    One is in text mode when the whole screen is black, showing (in most cases white) characters. A text mode login screen typically shows some information about the machine you are working on, the name of the machine and a prompt waiting for you to log in.

    +

    The login is different from a graphical login, in that you have to hit the Enter key after providing your user name, because there are no buttons on the screen that you can click with the mouse. Then you should type your password, followed by another Enter. You will not see any indication that you are entering something, not even an asterisk, and you won't see the cursor move. But this is normal on Linux and is done for security +reasons.

    +

    When the system has accepted you as a valid user, you may get some more information, called the message of the day, which can be anything. Additionally, it is popular on UNIX systems to display a fortune cookie, which contains some general wise or unwise (this is up to you) thoughts. After that, you will be given a shell, indicated with the same prompt that you would get in graphical mode.

    +

    Also in text mode: log in as root only to do setup and configuration that absolutely requires administrator privileges, such as adding users, installing software packages, and performing network and other system configuration. Once you are finished, immediately leave the special account and resume your work as a non−privileged user.

    +

    Logging out is done by entering the logout command, followed by Enter. You are successfully disconnected from the system when you see the login screen again.Don't power−off the computer after logging out. It is not meant to be shut off without application of the proper procedures for halting the system. Powering it off without going through the halting process might cause severe damage!

    +
    +
    +
    +
    +

    +3. Basic Commands

    +
    +

    +3.1. ls

    +

    When invoked without any arguments, ls lists the files in the current working directory. A directory that is not the current working directory can be specified and ls will list the files there. The user also may specify any list of files and directories. In this case, all files and all contents of specified directories will be listed. The name ls is derived from list segments which was used in earlier systems.

    +

    Files whose names start with "." are not listed, unless the -a flag is specified or the files are specified explicitly.

    +

    Without options, ls displays files in a bare format. This bare format however makes it difficult to establish the type, permissions, and size of the files. The most common options to reveal this information or change the list of files are:

    +<block_quote>
      +
    • -l long format, displaying Unix file types, permissions, number of hard links, owner, group, size, date, and filename

    • +
    • -F appends a character revealing the nature of a file, for example, * for an executable, or / for a directory. Regular files have no suffix.

    • +
    • -a lists all files in the given directory, including those whose names start with "." (which are hidden files in Unix). By default, these files are excluded from the list.

    • +
    • -R recursively lists subdirectories. The command ls -R / would therefore list all files.

    • +
    • -d shows information about a symbolic link or directory, rather than about the link's target or listing the contents of a directory.

    • +
    • -t sort the list of files by modification time.

    • +
    • -h print sizes in human readable format. (e.g., 1K, 234M, 2G, etc.)

    • +
    </block_quote>

    In some environments, providing the option --color (for GNU ls) or -G (FreeBSD ls) causes ls to highlight different types of files with different colors, instead of with characters as -F would. To determine what color to use for a file, GNU ls checks the Unix file type, the file permissions, and the file extension, while FreeBSD ls checks only the Unix file type and file permissions.:

    +
    $ ls
    +jeeves.rst psmith.html blandings.html
    +$ ls -l
    +drwxr--r--   1 plum  editors   4096  jeeves
    +-rw-r--r--   1 plum  editors  30405  psmith
    +-r-xr-xr-x   1 plum  plum      8460  blandings
    +

    Here "$" actually is the beginning of the prompt. This is typical in most Unix-based systems.

    +
    +
    +

    +3.2. date

    +

    The Unix date command displays the time and date. The super-user can use it to set the system clock.

    +

    With no options, the date command displays the current date and time, including the abbreviated day name, abbreviated month name, day of the month, the time separated by colons, the timezone name, and the year. For example:

    +
    $date
    +Tue Sep  8 12:01:45 IST 2009
    +

    On some systems to set the current date and time to September 8, 2004 01:22 you type:

    +
    $date --set="20040908 01:22"
    +

    In order to view the various options for the date command, type:

    +
    $man date
    +

    This will take you to the "Manual" page comprising of all the details on the date command. You can return to the terminal from the "man" page by pressing the Esc key in the keyboard and typing ":q" in that order.

    +
    +
    +

    +3.3. cd

    +

    This stands for "change directory". When one wants to go up to the parent directory, bypassing the tree of directories one has entered, “ cd ..” can be used.

    +

    One dot '.' represents the current directory while two dots '..' represent the parent directory.

    +

    “ cd -” will return you to the previous directory (a bit like an “undo”).

    +

    You can also use cd absolute path or cd relative path (see below):

    +

    Absolute paths:

    +<block_quote>

    An “ absolute path” is easily recognised from the leading forward slash, /. The / means that you start at the top level directory and continue down.

    </block_quote>

    For example to get to /boot/grub you would type:

    +
    $cd /boot/grub
    +

    This is an absolute path because you start at the top of the hierarchy and go downwards from there (it doesn't matter where in the filesystem you were when you typed the command).

    +

    Relative paths:

    +<block_quote>

    A “ relative path” doesn't have a preceding slash. Use a relative path when you start from a directory below the top level directory structure. This is dependent on where you are in the filesystem.

    +

    For example if you are in root's home directory and want to get to /root/music, you type:

    +
    $ cd music
    </block_quote>

    Please note that there is no / using the above cd command. Using a / would cause this to be an absolute path, working from the top of the hierarchy downward.

    +
    +
    +

    +3.4. who

    +

    The standard Unix command who displays a list of users who are currently logged into a computer.

    +

    The who command is related to the command w, which provides the same information but also displays additional data and statistics.:

    +
    $who
    +beeblebrox tty7         2009-09-08 10:50 (:0)
    +beeblebrox pts/0        2009-09-08 11:25 (:0.0)
    +dumbledore pts/1        2009-09-08 18:11 (potter.xyz.in)
    +beeblebrox pts/2        2009-09-08 18:53 (:0.0)
    +

    The command can be invoked with the arguments am i or am I (so it is invoked as who am i or * who am I*), showing information about the current terminal only (see the -m option below, of which this invocation is equivalent).

    +

    In order to find out the various options that can be appended to the who command, check the man page by typing out the following in the terminal:

    +
    $man who
    +

    This will take you to the "Manual" page containing details about the who command

    +
    +
    +

    +3.5. mkdir

    +

    This command is used to make a new directory. Normal usage is as straightforward as follows:

    +
    $mkdir name_of_directory
    +

    Where name_of_directory is the name of the directory one wants to create. When typed as above (ie. normal usage), the new directory would be created within the current directory. On Unix, multiple directories can be specified, and mkdir will try to create all of them.

    +
    +

    +3.5.1. Options

    +

    On Unix-like operating systems, mkdir takes options. Three of the most common options are:

    +<block_quote>
      +
    • -p: will also create all directories leading up to the given directory that do not exist already. If the given directory already exists, ignore the error.

    • +
    • -v: display each directory that mkdir creates. Most often used with -p.

    • +
    • -m: specify the octal permissions of directories created by mkdir.

    • +
    </block_quote>

    -p is most often used when using mkdir to build up complex directory hierarchies, in case a necessary directory is missing or already there. -m is commonly used to lock down temporary directories used by shell scripts.

    +
    +
    +

    +3.5.2. Examples

    +

    An example of -p in action is:

    +
    $mkdir -p /tmp/a/b/c
    +

    If /tmp/a exists but /tmp/a/b does not, mkdir will create /tmp/a/b before creating /tmp/a/b/c.

    +

    And an even more powerful command, creating a full tree at once (this however is a Shell extension, nothing mkdir does itself):

    +
    $mkdir -p tmpdir/{trunk/sources/{includes,docs},branches,tags}
    +

    This will create:

    +<definition_list><definition_list_item><term>tmpdir - branches</term><definition>
      +
    • tag

    • +
    • <definition_list><definition_list_item><term>trunk - sources - includes</term><definition>
      • docs

      </definition>
      </definition_list_item>
      </definition_list>
    • +
    </definition>
    </definition_list_item>
    </definition_list>
    +
    +
    +
    +
    +

    +4. Getting Help

    +
    +

    +4.1. apropos and whatis

    +

    This is a command to search the manual pages files in Unix and Unix-like operating systems.

    +
    $ apropos grep
    +egrep       egrep (1)       Search a file for a pattern using full regular expressions
    +fgrep       fgrep (1)       Search a file for a fixed-character string
    +fmlgrep     fmlgrep (1)     Search a file for a pattern
    +grep        grep (1)        Search a file for a pattern
    +gzgrep      gzgrep (1)      Search a possibly compressed file for a regular expression
    +nisgrep     nismatch (1)    Utilities for searching NIS+ tables
    +pgrep       pgrep (1)       Find or signal a process by name or other attribute
    +zgrep       zgrep (1)       Search a possibly compressed file for a regular expression
    +...
    +

    In this example, the user uses apropos to search for the string "grep", and apropos returns the indicated man pages that include the term "grep".

    +

    A short index of explanations for commands is available using the whatis command, like in the examples below:

    +
    $whatis ls
    +ls (1)           - list directory contents
    +

    This displays short information about a command, and the first section in the collection of man pages that contains an appropriate page.

    +

    If you don't know where to get started and which man page to read, apropos gives more information. Say that you do not know how to start a browser, then you could enter the following command:

    +
    $apropos browser
    +gmusicbrowser (1)    - Jukebox for large collections of audio files
    +infobrowser (1)      - read Info documents
    +libsmbclient (7)     - An extension library for browsers and that               can be used...
    +opera (1)            - a standards-compliant graphical Web browser
    +sensible-browser (1) - sensible editing, paging, and web browsing
    +smbtree (1)          - A text based smb network browser
    +tvtk_doc (1)         - A GUI based TVTK documentation search browser.
    +viewres (1)          - graphical class browser for Xt
    +w3m (1)              - a text based Web browser and pager
    +www-browser (1)      - a text based Web browser and pager
    +...
    +
    +
    +

    +4.2. man

    +

    Man pages (short for "manual pages") are the extensive documentation that comes preinstalled with almost all substantial Unix and Unix-like operating systems. The Unix command used to display them is man. Each page is a self-contained document.

    +

    To read a manual page for a Unix command, one can use:

    +
    $ man <command_name>
    +

    at a shell prompt; for example, "man ftp". In order to simplify navigation through the output, man generally uses the less terminal pager.

    +

    Pages are traditionally referred to using the notation "name(section)"; for example, ftp(1). The same page name may appear in more than one section of the manual, this can occur when the names of system calls, user commands, or macro packages coincide. Two examples are man(1) and man(7), or exit(2) and exit(3). The syntax for accessing the non-default manual section varies between different man implementations. On Linux and BSD, for example, the syntax for reading *printf(3) is:

    +
    $man 3 printf
    +

    Another example:

    +
    $man man
    +

    The previous example will take you to the "Manual" page entry about manual pages!

    +
    +

    +4.2.1. Layout

    +

    All man pages follow a common layout that is optimized for presentation on a simple ASCII text display, possibly without any form of highlighting or font control. Sections present may include:

    +<definition_list><definition_list_item><term>NAME</term><definition>

    The name of the command or function, followed by a one-line description of what it does.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SYNOPSIS</term><definition>

    In the case of a command, you get a formal description of how to run it and what command line options it takes. For program functions, a list of the parameters the function takes and which header file contains its definition. For experienced users, this may be all the documentation they need.

    </definition>
    </definition_list_item>
    <definition_list_item><term>DESCRIPTION</term><definition>

    A textual description of the functioning of the command or function.

    </definition>
    </definition_list_item>
    <definition_list_item><term>EXAMPLES</term><definition>

    Some examples of common usage.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SEE ALSO</term><definition>

    A list of related commands or functions.

    </definition>
    </definition_list_item>
    </definition_list>

    Other sections may be present, but these are not well standardized across man pages. Common examples include: OPTIONS, EXIT STATUS, ENVIRONMENT, KNOWN BUGS, FILES, AUTHOR, REPORTING BUGS, HISTORY and COPYRIGHT.

    +

    These days virtually every Unix command line application comes with its man page, and many Unix users perceive a lack of man pages as a sign of low quality; indeed, some projects, such as Debian, go out of their way to write man pages for programs lacking one. Few alternatives to man have enjoyed much popularity, with the possible exception of the GNU project's "info" system, an early and simple hypertext system.

    +

    However, the format of a single page for each application, the lack of classification within the sections and the relatively unsophisticated formatting facilities have motivated the development of alternative documentation systems, such as the previously mentioned "info" system.

    +

    Most Unix GUI applications (particularly those built using the GNOME and KDE development environments) now provide end-user documentation in HTML and include embedded HTML viewers such as yelp for reading the help within the application.

    +

    Usually the man pages are written in English. Translations into other languages can be also available on the system.

    +

    The default format of the man pages is troff, with either the macro package man (appearance oriented) or on some systems mdoc (semantic oriented). This makes it possible to typeset a man page to PostScript, PDF and various other formats for viewing or printing.

    +
    +
    +
    +

    +4.3. info

    +

    info is a software utility which forms a hypertextual, multipage documentation and help viewer working on a command line interface, useful when there is no GUI available.

    +

    The syntax is

    +
    $ info <command_name>
    +

    info processes info files, which are Texinfo formatted files, and presents the documentation as a tree, with simple commands to traverse the tree and to follow cross references. For instance

    +<block_quote>
      +
    • n goes to the next page.

    • +
    • p goes to the previous page.

    • +
    • u goes to the upper page.

    • +
    • l goes to the last(visited) node

    • +
    • To follow a cross reference, the cursor can be moved over a link (a word preceded by a <title_reference>*</title_reference>) and enter pressed.

    • +
    </block_quote>

    info was initially written for use with GNU/Linux and then ported to other Unix-like operating systems.

    +
    +
    +

    +4.4. --help

    +

    Most GNU commands support the --help, which gives a short explanation about how to use the command and a list of available options. Below is the output of this option with the cat command:

    +
    $ userprompt@host: cat --help
    +Usage: cat [OPTION] [FILE]...
    +Concatenate FILE(s), or standard input, to standard output.
    +
    +  -A, --show-all           equivalent to -vET
    +  -b, --number-nonblank    number nonempty output lines
    +  -e                       equivalent to -vE
    +  -E, --show-ends          display $ at end of each line
    +  -n, --number             number all output lines
    +  -s, --squeeze-blank      suppress repeated empty output lines
    +  -t                       equivalent to -vT
    +  -T, --show-tabs          display TAB characters as ^I
    +  -u                       (ignored)
    +  -v, --show-nonprinting   use ^ and M- notation, except for LFD and              TAB
    +  --help     display this help and exit
    +  --version  output version information and exit
    +
    +With no FILE, or when FILE is -, read standard input.
    +
    +Examples:
    +  cat f - g  Output f's contents, then standard input, then g's           contents.
    +  cat        Copy standard input to standard output.
    +
    +Report bugs to <bug-coreutils@gnu.org>.
    +
    +
    +
    +

    +5. Basic file handling

    +
    +

    +5.1. cp

    +

    cp is the command entered in a Unix shell to copy a file from one place to another, possibly on a different filesystem. The original file remains unchanged, and the new file may have the same or a different name.

    +
    +

    +5.1.1. Usage

    +

    To copy a file to another file:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ][ -- ] SourceFile TargetFile
    +

    To copy a file to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -r | -R ] [ -- ] SourceFile ...              TargetDirectory
    +

    To copy a directory to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -- ] { -r | -R }
    +SourceDirectory ... TargetDirectory
    +
    +
    +

    +5.1.2. Flags

    +

    -f (force) – specifies removal of the target file if it cannot be opened for write operations. The removal precedes any copying performed by the cp command.

    +

    -P – makes the cp command copy symbolic links. The default is to follow symbolic links, that is, to copy files to which symbolic links point.

    +

    -i (interactive) – prompts you with the name of a file to be overwritten. This occurs if the TargetDirectory or TargetFile parameter contains a file with the same name as a file specified in the SourceFile or SourceDirectory parameter. If you enter y or the locale's equivalent of y, the cp command continues. Any other answer prevents the cp command from overwriting the file.

    +

    -p (preserve) – duplicates the following characteristics of each SourceFile/SourceDirectory in the corresponding TargetFile and/or TargetDirectory:

    +<block_quote>
      +
    • The time of the last data modification and the time of the last access.

    • +
    • The user ID and group ID (only if it has permissions to do this)

    • +
    • The file permission bits and the SUID and SGID bits.

    • +
    </block_quote>

    -R (recursive) – copy directories (recursively copying all the contents)

    +
    +
    +

    +5.1.3. Examples

    +

    To make a copy of a file in the current directory, enter:

    +
    $ cp prog.c prog.bak
    +

    This copies prog.c to prog.bak. If the prog.bak file does not already exist, the cp command creates it. If it does exist, the cp command replaces it with a copy of the prog.c file.

    +

    To copy a file in your current directory into another directory, enter:

    +
    $ cp zaphod /home/books/hhgg
    +

    This copies the jones file to /home/books/hhgg/zaphod.

    +

    To copy a file to a new file and preserve the modification date, time, and access control list associated with the source file, enter:

    +
    $ cp -p martin_luther_king martin_luther_king.jr
    +

    This copies the martin_luther_king file to the martin_luther_king.jr file. Instead of creating the file with the current date and time stamp, the system gives the martin_luther_king.jr file the same date and time as the martin_luther_king file. The martin_luther_king.jr file also inherits the martin_luther_king file's access control protection.

    +

    To copy all the files in a directory to a new directory, enter:

    +
    $ cp /home/galactica/clients/* /home/hhgg/customers
    +

    This copies only the files in the clients directory to the customers directory.

    +

    To copy a directory, including all its files and subdirectories, to another directory, enter:

    +<block_quote>

    $ cp -R /home/hhgg/clients /home/hhgg/customers

    </block_quote>

    This copies the clients directory, including all its files, subdirectories, and the files in those subdirectories, to the customers/clients directory.

    +

    To copy a specific set of files of any extension to another directory, enter:

    +
    $ cp zaphod arthur ford /home/hhgg/clients
    +

    This copies the zaphod, arthur, and ford files in your current working directory to the /home/hhgg/clients directory.

    +

    To use pattern-matching characters to copy files, enter:

    +
    $ cp programs/*.py .
    +

    This copies the files in the programs directory that end with .py to the current directory, signified by the single "." (dot). You must type a space between the py and the final dot.

    +
    +
    +
    +

    +5.2. mv

    +

    mv (short for move) is a Unix command that moves one or more files or directories from one place to another. The original file is deleted, and the new file may have the same or a different name. If possible (i.e. when the original and new files are on the same file system), mv will rename the file instead. Write permission is required on all directories being modified.

    +
    +

    +5.2.1. Conflicting existing file

    +

    In all cases, when a file is moved to have the name of an existing file (in the same directory), the existing file is deleted. If the existing file is not writable but is in a directory that is writable, then the mv command asks for confirmation if possible (i.e. if run from a terminal) before proceeding, unless the -f (force) option is used.

    +
    +
    +

    +5.2.2. Differences with copy and delete

    +

    Note that, usually, when moving files within the same volume, moving (and/or renaming) is not the same as simply copying and then deleting the original. When moving a file, the link is simply removed from the old parent directory and added to the new parent directory. However, the file itself is untouched (i.e. it has the same inodes and resides at the same place on the disk). For example, you cannot copy a file you cannot read, but you can move (and/or rename) it (provided you have write permission to its old and new parent directories). Also, suppose there is a non-empty directory you do not have write permission to. You cannot delete this directory (since you cannot delete its contents); but you can move (and/or rename) it. Also, since moving between filenames on a single volume does not involve copying, it is faster and does not place strain of lots of reads and writes on the disk. Moving files across different volumes, however, does necessitate copying and deleting.

    +
    +
    +

    +5.2.3. Examples

    +
    $ mv myfile mynewfilename    renames a file
    +$ mv myfile otherfilename    renames a file and deletes the existing            file "myfile"
    +$ mv myfile /myfile          moves 'myfile' from the current            directory to the root directory
    +$ mv myfile dir/myfile       moves 'myfile' to 'dir/myfile' relative            to the current directory
    +$ mv myfile dir              same as the previous command (the          filename is implied to be the same)
    +$ mv myfile dir/myfile2      moves 'myfile' to dir and renames it to            'myfile2'
    +$ mv foo bar baz dir         moves multiple files to directory dir
    +$ mv --help                  shows a very concise help about the                syntax of the command
    +$ man mv                     prints an extensive user manual for                'mv' in the terminal
    +

    In all cases, the file or files being moved or renamed can be a directory.

    +

    Note that when the command is called with two arguments (as mv name1 name2 or mv name1 /dir/name2), it can have three different effects, depending on whether name2 does not exist, is an existing file, or is an existing directory. If the user intends to refer to an existing directory, /. (or in some Unix versions / is sufficient) may be appended to the name to force the system to check this. To move a file to a new directory, the directory must be created first.

    +
    +
    +
    +

    +5.3. rm

    +

    rm (short for "remove") is one of several basic Unix command lines that operates on files. It is used to delete files from a filesystem. The data is not actually destroyed. Only the index listing where the file is stored is destroyed, and the storage is made available for reuse. There are undelete utilities that will attempt to reconstruct the index and can bring the file back if the parts were not reused.

    +

    Here's example to remove a file named "foo" from a directory, here shown with the -i option:

    +
    $ rm -i foo
    +remove foo? y
    +
    +

    +5.3.1. Options

    +

    Common options that rm accepts include:

    +<block_quote>
      +
    • -r, which removes directories, removing the contents recursively beforehand (so as not to leave files without a directory to reside in) ("recursive")

    • +
    • -i, which asks for every deletion to be confirmed ("interactive")

    • +
    • -f, which ignores non-existent files and overrides any confirmation prompts ("force")

    • +
    • -v, which shows what is being removed as it happens ("verbose")

    • +
    </block_quote>

    rm is often aliased to "rm -i" so as to avoid accidental deletion of files. If a user still wishes to delete a large number of files without confirmation, they can manually cancel out the -i argument by adding the -f option (as the option specified later on the expanded command line "rm -i -f" takes precedence).

    +

    rm -rf (variously, rm -rf /, rm -rf <title_reference>*</title_reference>, and others) is frequently used in jokes and anecdotes about Unix disasters. The rm -rf variant of the command, if run by a superuser on the root directory, would cause the contents of every writable mounted filesystem on the computer to be deleted.

    +

    rm is often used in conjunction with xargs to supply a list of files to delete:

    +
    xargs rm < filelist
    +

    When rm is used on a symbolic link, it deletes the link, but does not affect the target of the link.

    +
    +
    +

    +5.3.2. Permissions

    +

    Usually, on most filesystems, deleting a file requires write permission on the parent directory (and execute permission, in order to enter the directory in the first place). (Note that, confusingly for beginners, permissions on the file itself are irrelevant. However, GNU rm asks for confirmation if a write-protected file is to be deleted, unless the -f option is used.)

    +

    To delete a directory (with rm -r), one must delete all of its contents recursively. This requires that one must have read and write and execute permission to that directory (if it's not empty) and all non-empty subdirectories recursively (if there are any). The read permissions are needed to list the contents of the directory in order to delete them. This sometimes leads to an odd situation where a non-empty directory cannot be deleted because one doesn't have write permission to it and so cannot delete its contents; but if the same directory were empty, one would be able to delete it.

    +

    If a file resides in a directory with the sticky bit set, then deleting the file requires one to be the owner of the file.

    +
    +
    +
    +
    +

    +6. Command Line Arguments

    +

    In computer command line interfaces, a command line argument is an argument sent to a program being called. In general, a program can take any number of command line arguments, which may be necessary for the program to run, or may even be ignored, depending on the function of that program.

    +

    For example, in Unix and Unix-like environments, an example of a command-line argument is:

    +
    rm file.s
    +

    "file.s" is a command line argument which tells the program rm to remove the file "file.s".

    +

    Programming languages such as C, C++ and Java allow a program to interpret the command line arguments by handling them as string parameters in the main function.

    +

    A command line option or simply option (also known as a command line parameter, flag, or a switch) is an indication by a user that a computer program should change its default output.

    +

    Long options are introduced via "--", and are typically whole words. For example, ls --long --classify --all. Arguments to long options are provided with "=", as ls --block-size=1024. Some Unix programs use long options with single dashes, for example MPlayer as in mplayer -nosound.

    +

    Linux also uses "--" to terminate option lists. For example, an attempt to delete a file called -file1 by using rm -file1 may produce an error, since rm may interpret -file1 as a command line switch. Using rm -- -file1 removes ambiguity.

    +
    +
    +

    +7. Basic Text Processing

    +
    +

    +7.1. head

    +

    head is a program on Unix and Unix-like systems used to display the first few lines of a text file or piped data. The command syntax is:

    +
    $ head [options] <file_name>
    +

    By default, head will print the first 10 lines of its input to the standard output. The number of lines printed may be changed with a command line option. The following example shows the first 20 lines of filename:

    +
    $ head -n 20 filename
    +

    This displays the first 5 lines of all files starting with foo:

    +
    $ head -n 5 foo*
    +

    Some versions omit the n and just let you say -5.

    +
    +

    +7.1.1. Flags

    +
    -c <x number of bytes> Copy first x number of bytes.
    +

    Other options: sed

    +

    Many early versions of Unix did not have this command, and so documentation and books had sed do this job:

    +
    sed 5q foo
    +

    This says to print every line (implicit), and quit after the fifth.

    +
    +
    +
    +

    +7.2. tail

    +

    tail is a program on Unix and Unix-like systems used to display the last few lines of a text file or piped data.

    +

    The command-syntax is:

    +
    $ tail [options] <file_name>
    +

    By default, tail will print the last 10 lines of its input to the standard output. With command line options the number of lines printed and the printing units (lines, blocks or bytes) may be changed. The following example shows the last 20 lines of filename:

    +
    $ tail -n 20 filename
    +

    This example shows the last 15 bytes of all files starting with foo:

    +
    $ tail -c 15 foo*
    +

    This example shows all lines of filename from the second line onwards:

    +
    $ tail -n +2 filename
    +

    Using an older syntax (still used in Sun Solaris as the -n option is not supported), the last 20 lines and the last 50 bytes of filename can be shown with the following command:

    +
    $ tail -20 filename
    +$ tail -50c filename
    +

    However this syntax is now obsolete and does not conform with the POSIX 1003.1-2001 standard. Even if still supported in current versions, when used with other options (like -f, see below), these switches could not work at all.

    +
    +

    +7.2.1. File monitoring

    +

    tail has a special command line option -f (follow) that allows a file to be monitored. Instead of displaying the last few lines and exiting, tail displays the lines and then monitors the file. As new lines are added to the file by another process, tail updates the display. This is particularly useful for monitoring log files. The following command will display the last 10 lines of messages and append new lines to the display as new lines are added to messages:

    +
    $ tail -f /var/adm/messages
    +

    To interrupt tail while it is monitoring, break-in with Ctrl+C. This command can be run "in the background" with &, see job control.

    +

    If you have a command's result to monitor, you can use the watch command.

    +
    +
    +
    +

    +7.3. cut

    +

    In computing, cut is a Unix command line utility which is used to extract sections from each line of input — usually from a file.

    +

    Extraction of line segments can typically be done by bytes (-b), characters (-c), or fields (-f) separated by a delimiter (-d — the tab character by default). A range must be provided in each case which consists of one of N, N-M, N- (N to the end of the line), or -M (beginning of the line to M), where N and M are counted from 1 (there is no zeroth value). Since version 6, an error is thrown if you include a zeroth value. Prior to this the value was ignored and assumed to be 1.

    +

    Assuming a file named file containing the lines:

    +
    foo:bar:baz:qux:quux
    +one:two:three:four:five:six:seven
    +alpha:beta:gamma:delta:epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +

    To output the fourth through tenth characters of each line:

    +
    $ cut -c 4-10 file
    +

    This gives the output:

    +
    :bar:ba
    +:two:th
    +ha:beta
    +

    To output the fifth field through the end of the line of each line using the colon character as the field delimiter:

    +
    $ cut -d : -f 5- file
    +

    This gives the output:

    +
    quux
    +five:six:seven
    +epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +
    +
    +

    +7.4. paste

    +

    paste is a Unix command line utility which is used to join files horizontally (parallel merging) by outputting lines consisting of the sequentially corresponding lines of each file specified, separated by tabs, to the standard output. It is effectively the horizontal equivalent to the utility cat command which operates on the vertical plane of two or more files.

    +

    To paste several columns of data together into the file www from files who, where, and when:

    +
    $ paste who where when > www
    +

    If the files contain:

    +++++
    +

    This creates the file named www containing:

    +
    Batman            GothamCity       January 3
    +Trillian          Andromeda        February 4
    +Jeeves            London           March 19
    +
    +
    +
    +

    +8. Shell Meta Characters

    +

    Unix recognizes certain special characters, called "meta characters," as command directives. The shell meta characters are recognized anywhere they appear in the command line, even if they are not surrounded by blank space. For that reason, it is safest to only use the characters A-Z, a-z, 0-9, and the period, dash, and underscore characters when naming files and directories on Unix. If your file or directory has a shell meta character in the name, you will find it difficult to use the name in a shell command.

    +

    The shell meta characters include:

    +

    / < > ! $ % ^ & * | { } [ ] " ' ` ~ ;

    +

    Different shells may differ in the meta characters recognized.

    +

    As an example,

    +
    $ ls file.*
    +

    run on a directory containing the files file, file.c, file.lst, and myfile would list the files file.c and file.lst. However,:

    +
    $ ls file.?
    +

    run on the same directory would only list file.c because the ? only matches one character, no more, no less. This can save you a great deal of typing time. For example, if there is a file called california_cornish_hens_with_wild_rice and no other files whose names begin with 'c', you could view the file without typing the whole name by typing this:

    +
    $ more c*
    +

    because the c* matches that long file name.

    +

    Filenames containing metacharacters can pose many problems and should never be intentionally created. If you do find that you've created a file with metacharacters, and you would like to remove it, you have three options. You may use wildcards to match metacharacter, use the to directly enter the filename, or put the command in double quotes (except in the case of double quotes within the file name, these must be captured with one of the first two methods). For example, deleting a file named <title_reference>"``*`|more</title_reference>"` can be accomplished with:

    +
    $ rm ??more
    +

    or:

    +
    $ rm $\backslash$*$\backslash$|more
    +

    or:

    +
    $ rm ''*|more''
    +
    +
    +

    +9. Looking At Files

    +
    +

    +9.1. cat

    +

    The cat command is a standard Unix program used to concatenate and display files. The name is from "catenate", a synonym of concatenate.

    +

    The Single Unix Specification specifies the behavior that the contents of each of the files given in sequence as arguments will be written to the standard output in the same sequence, and mandates one option, -u, where each byte is printed as it is read.

    +

    If the filename is specified as -, then cat will read from standard input at that point in the sequence. If no files are specified, cat will read from standard input entered.

    +
    +

    +9.1.1. Jargon File Definition

    +

    The Jargon File version 4.4.7 lists this as the definition of cat:

    +
    1. To spew an entire file to the screen or some other output sink without
    +     pause (syn. blast).
    +
    +2. By extension, to dump large amounts of data at an unprepared target or
    +     with no intention of browsing it carefully. Usage: considered silly.
    +     Rare outside Unix sites. See also dd, BLT.
    +
    +     Among Unix fans, *cat(1)* is considered an excellent example of
    +     user-interface design, because it delivers the file contents without
    +     such verbosity as spacing or headers between the files, and because
    +     it does not require the files to consist of lines of text, but works
    +     with any sort of data.
    +
    +     Among Unix critics, *cat(1)* is considered the canonical example of
    +     bad user-interface design, because of its woefully unobvious name.
    +     It is far more often used to blast a single file to standard output
    +     than to concatenate two or more files. The name cat for the former
    +     operation is just as unintuitive as, say, LISP's cdr.
    +
    +     Of such oppositions are holy wars made...
    +
    +
    +

    +9.1.2. Useless Use of 'cat'

    +

    UUOC (from comp.unix.shell on Usenet) stands for “Useless Use of cat”. As it is observed on comp.unix.shell, “The purpose of cat is to concatenate (or 'catenate') files. If it's only one file, concatenating it with nothing at all is a waste of time, and costs you a process.”

    +

    Nevertheless one sees people doing:

    +
    $ cat file | some_command and its args ...
    +

    instead of the equivalent and cheaper:

    +
    <file some_command and its args ...
    +

    or (equivalently and more classically):

    +
    some_command and its args ... <file
    +

    Since 1995, occasional awards for UUOC have been given out. The activity of fixing instances of UUOC is sometimes called 'demoggification'.

    +

    Amongst many, it is still considered safer to use cat for such cases given that the < and > keys are next to each other in many popular keyboard mappings. While the risk might be low, the impact of using > instead of < can be high and prohibitive.

    +
    +
    +

    +9.1.3. zcat

    +

    zcat is a Unix program similar to cat, that decompresses individual files and concatenates them to standard output. Traditionally zcat operated on files compressed by compress but today it is usually able to operate on gzip or even bzip2 archives. On such systems, it is equivalent to gunzip -c

    +
    +
    +
    +

    +9.2. more

    +

    In computing, more is a command to view (but not modify) the contents of a text file one screen at a time (terminal pager). It is available on Unix and Unix-like systems, DOS, OS/2 and Microsoft Windows. Programs of this sort are called pagers.

    +
    +

    +9.2.1. Usage

    +

    The command-syntax is:

    +
    $ more [options] [file_name]
    +

    If no file name is provided, more looks for input from stdin.

    +

    Once more has obtained input, it displays as much as can fit on the current screen and waits for user input to advance, with the exception that a form feed (^L) will also cause more to wait at that line, regardless of the amount of text on the screen. In the lower-left corner of the screen is displayed the text "--More--" and a percentage, representing the percent of the file that more has paged through. (This percentage includes the text displayed on the current screen.) When more reaches the end of a file (100%) it exits. The most common methods of navigating through a file are Enter, which advances the output by one line, and Space, which advances the output by one screen.

    +

    There are also other commands that can be used while navigating through the document; consult more's man page for more details.

    +

    Options are typically entered before the file name, but can also be entered in the environment variable $MORE. Options entered in the actual command line will override those entered in the $MORE environment variable. Available options may vary between Unix systems.

    +
    +
    +
    +

    +9.3. less

    +

    less is a terminal pager program on Unix, Windows and Unix-like systems used to view (but not change) the contents of a text file one screen at a time. It is similar to more, but has the extended capability of allowing both forward and backward navigation through the file. Unlike most Unix text editors/viewers, less does not need to read the entire file before starting, resulting in faster load times with large files.

    +
    +

    +9.3.1. Usage

    +

    less can be invoked with options to change its behaviour, for example, the number of lines to display on the screen. A few options vary depending on the operating system. While less is displaying the file, various commands can be used to navigate through the file. These commands are based on those used by both more and vi. It is also possible to search for character patterns in the file.

    +

    By default, less displays the contents of the file to the standard output (one screen at a time). If the file name argument is omitted, it displays the contents from standard input (usually the output of another command through a pipe). If the output is redirected to anything other than a terminal, for example a pipe to another command, less behaves like cat.

    +

    The command-syntax is:

    +
    $ less [options] file_name
    +
    +
    +

    +9.3.2. Frequently Used Options

    +<block_quote>
      +
    • -g: Highlights just the current match of any searched string.

    • +
    • -I: Case-insensitive searches.

    • +
    • -M: Shows more detailed prompt, including file position.

    • +
    • -N: Shows line numbers (useful for source code viewing).

    • +
    • -S: Disables line wrap ("chop long lines"). Long lines can be seen by side scrolling.

    • +
    • -?: Shows help.

    • +
    </block_quote>
    +
    +
    +

    +9.3.3. Frequently Used Commands

    +<block_quote>
      +
    • [Arrows]/[Page Up]/[Page Down]/[Home]/[End]: Navigation.

    • +
    • [Space bar]: Next page.

    • +
    • b: Previous page.

    • +
    • ng: Jump to line number n. Default is the start of the file.

    • +
    • nG: Jump to line number n. Default is the end of the file.

    • +
    • /pattern: Search for pattern. Regular expressions can be used.

    • +
    • '^ or g: Go to start of file.

    • +
    • '$ or G: Go to end of file.

    • +
    • s: Save current content (got from another program like grep) in a file.

    • +
    • =: File information.

    • +
    • h: Help.

    • +
    • q: Quit.

    • +
    </block_quote>
    +
    +
    +

    +9.3.4. Examples

    +
    $ less -M readme.txt                     #Read "readme.txt."
    +$ less +F /var/log/mail.log              #Follow mode for log
    +$ file * | less                          #Easier file analysis.
    +$ grep -i void *.c | less -I -p void     #Case insensitive search                                                         for "void" in all .c files
    +
    +
    +
    +
    +

    +10. Directory Structure

    +

    In the File Hierarchy Standard (FHS) all files and directories appear under the root directory "/", even if they are stored on different physical devices. Note however that some of these directories may or may not be present on a Unix system depending on whether certain subsystems, such as the X Window System, are installed.

    +

    The majority of these directories exist in all UNIX operating systems and are generally used in much the same way; however, the descriptions here are those used specifically for the FHS, and are not considered authoritative for platforms other than Linux.

    ++++
    +
    +

    +10.1. man hier

    +

    This is the manual page on the UNIX filesystem. The syntax for this is:

    +
    $ man hier
    +
    +
    +

    +10.2. ls -l

    +

    Shows you huge amounts of information (permissions, owners, size, and when last modified) for folders and files. The syntax is

    +
    $ ls -l
    +

    This can be done after entering the required directory.

    +
    +
    +
    +

    +11. Permissions and Ownership

    +
    +

    +11.1. chmod

    +

    The chmod command (abbreviated from 'change mode') is a shell command and C language function in Unix and Unix-like environments. When executed, it can change file system modes of files and directories. The modes include permissions and special modes.A chmod command first appeared in AT&T Unix version 1, and is still used today on Unix-like machines.

    +
    +

    +11.1.1. Usage

    +

    The chmod command options are specified like this:

    +
    $ chmod [options] mode[,mode] file1 [file2 ...]
    +

    To view what the permissions currently are, type:

    +
    $ ls -l file
    +
    +
    +

    +11.1.2. Command line options

    +

    The chmod command has a number of command line options that affect its behavior. The most common options are:

    +<block_quote>
      +
    • -R: Changes the modes of directories and files recursively

    • +
    • -v: Verbose mode; lists all files as they are being processed

    • +
    </block_quote>
    +
    +11.1.2.1. Symbolic modes
    +

    To the chmod utility, all permissions and special modes are represented by its mode parameter. One way to adjust the mode of files or directories is to specify a symbolic mode. The symbolic mode is composed of three components, which are combined to form a single string of text:

    +
    $ chmod [references][operator][modes] file1 ...
    +

    The references (or classes) are used to distinguish the users to whom the permissions apply. If no references are specified it defaults to “all” but modifies only the permissions allowed by the umask. The references are represented by one or more of the following letters:

    +++++
    +

    The chmod program uses an operator to specify how the modes of a file should be adjusted. The following operators are accepted:

    ++++
    +

    The modes indicate which permissions are to be granted or taken away from the specified classes. There are three basic modes which correspond to the basic permissions:

    +++++
    +

    The combination of these three components produces a string that is understood by the chmod command. Multiple changes can be specified by separating multiple symbolic modes with commas.

    +
    +
    +
    +11.1.2.2. Symbolic examples
    +

    Add the 'read' and 'write' permissions to the 'user' and 'group' classes of a directory:

    +
    $ chmod ug+rw mydir
    +$ ls -ld mydir
    +drw-rw----   2 starwars  yoda  96 Dec 8 12:53 mydir
    +

    For a file, remove write permissions for all classes:

    +
    $ chmod a-w myfile
    +$ ls -l myfile
    +-r-xr-xr-x   2 starwars  yoda 96 Dec 8 12:53 myfile
    +

    Set the permissions for the u*ser and the *g*roup to read and execute only (no write permission) on *mydir.

    +
    $ chmod ug=rx mydir
    +$ ls -ld mydir
    +dr-xr-x---   2 starwars  yoda 96 Dec 8 12:53 mydir
    +
    +
    +
    +11.1.2.3. Octal numbers
    +

    The chmod command also accepts three and four-digit octal numbers representing modes. Using a three-digit octal number to set the modes of a file named myfile :

    +
    $ chmod 664 myfile
    +$ ls -l myfile
    +-rw-rw-r--  1   57 Jul  3 10:13  myfile
    +

    Since the setuid, setgid and sticky bits are not set, this is equivalent to:

    +
    $ chmod 0664 myfile
    +
    +
    +
    +11.1.2.4. Special modes
    +

    The chmod command is also capable of changing the additional permissions or special modes of a file or directory. The symbolic modes use s to represent the setuid and setgid modes, and t to represent the sticky mode. The modes are only applied to the appropriate classes, regardless of whether or not other classes are specified.

    +

    Most operating systems support the specification of special modes using octal modes, but some do not. On these systems, only the symbolic modes can be used.

    +
    +
    +
    +
    +
    +

    +12. Redirection and Piping

    +

    In computing, redirection is a function common to most command-line interpreters, including the various Unix shells that can redirect standard streams to user-specified locations.

    +

    Programs do redirection with the dup2(2) system call, or its less-flexible but higher-level stdio analogues, freopen(3) and popen(3).

    +
    +

    +12.1. Redirecting standard input and standard output

    +

    Redirection is usually implemented by placing certain characters between commands. Typically, the syntax of these characters is as follows:

    +
    $ command1 > file1
    +

    executes command1, placing the output in file1. Note that this will truncate any existing data in file1. To append output to the end of the file, use the >> operator.:

    +
    $ command1 < file1
    +

    executes command1, using file1 as the source of input (as opposed to the keyboard).:

    +
    $ command1 < infile > outfile
    +

    combines the two capabilities: command1 reads from infile and writes to outfile

    +
    +
    +

    +12.2. Piping

    +

    Programs can be run together such that one program reads the output from another with no need for an explicit intermediate file: +A pipeline of three programs run on a text terminal:

    +
    $ command1 | command2
    +

    executes command1, using its output as the input for command2 (commonly called piping, since the "|" character is known as a "pipe").

    +

    This is equivalent to using two redirects and a temporary file:

    +
    $ command1 > tempfile
    +$ command2 < tempfile
    +$ rm tempfile
    +

    A good example for command piping is combining echo with another command to achieve something interactive in a non-interactive shell, e.g.:

    +
    $ echo -e "user\npass" | ftp localhost
    +

    This runs the ftp client with input user, press return, then pass.

    +
    +
    +

    +12.3. Redirecting to and from the standard file handles

    +

    In Unix shells derived from the original Bourne shell, the first two actions can be further modified by placing a number (the file descriptor) immediately before the character; this will affect which stream is used for the redirection. The Unix standard I/O streams are:

    +++++
    +

    For example:

    +
    $ command1 2> file1
    +

    executes command1, directing the standard error stream to file1.

    +

    In shells derived from csh (the C shell), the syntax instead appends the & character to the redirect characters, thus achieving a similar result.

    +

    Another useful capability is to redirect one standard file handle to another. The most popular variation is to merge standard error into standard output so error messages can be processed together with (or alternately to) the usual output. Example:

    +
    $ find / -name .profile > results 2>&1
    +

    will try to find all files named .profile. Executed without redirection, it will output hits to stdout and errors (e.g. for lack of privilege to traverse protected directories) to stderr. If standard output is directed to file results, error messages appear on the console. To see both hits and error messages in file results, merge stderr (handle 2) into stdout (handle 1) using 2>&1 .

    +

    It's possible use 2>&1 before ">" but it doesn't work. In fact, when the interpreter reads 2>&1, it doesn't know yet where standard output is redirected and then standard error isn't merged.

    +

    If the merged output is to be piped into another program, the file merge sequence 2>&1 must precede the pipe symbol, thus:

    +
    $ find / -name .profile 2>&1 | less
    +

    A simplified form of the command:

    +
    $ command > file 2>&1
    +

    is:

    +
    $ command &>file
    +

    or:

    +
    $command >&file
    +
    +
    +

    +12.4. Chained pipelines

    +

    The redirection and piping tokens can be chained together to create complex commands. For example:

    +
    $ ls | grep '\.sh' | sort > shlist
    +

    lists the contents of the current directory, where this output is filtered to only contain lines which contain .sh, sort this resultant output lexicographically, and place the final output in shlist. This type of construction is used very commonly in shell scripts and batch files.

    +
    +
    +

    +12.5. Redirect to multiple outputs

    +

    The standard command tee can redirect output from a command to several destinations.

    +
    $ ls -lrt | tee xyz
    +

    This directs the file list output to both standard output as well as to the file xyz.

    +
    +
    +
    +

    +13. More Text Processing

    +
    +

    +13.1. grep

    +

    grep is a command line text search utility originally written for Unix. The name is taken from the first letters in global / regular expression / print, a series of instructions for the ed text editor. The grep command searches files or standard input globally for lines matching a given regular expression, and prints them to the program's standard output.

    +
    +

    +13.1.1. Usage

    +

    This is an example of a common grep usage:

    +
    $ grep apple fruitlist.txt
    +

    In this case, grep prints all lines containing 'apple' from the file fruitlist.txt, regardless of word boundaries; therefore lines containing 'pineapple' or 'apples' are also printed. The grep command is case sensitive by default, so this example's output does not include lines containing 'Apple' (with a capital A) unless they also contain 'apple'.

    +

    Like most Unix commands, grep accepts command line arguments to change this and many other behaviors. For example:

    +
    $ grep -i apple fruitlist.txt
    +

    This prints all lines containing 'apple' regardless of capitalization. The '-i' argument tells grep to be case insensitive, or to ignore case.

    +

    To print all lines containing 'apple' as a word ('pineapple' and 'apples' will not match):

    +
    $ grep -w apple fruitlist.txt
    +

    Regular expressions can be used to match more complicated queries.

    +
    +
    +13.1.1.1. Variations
    +

    There are countless implementations and derivatives of grep available for many operating systems. Early variants of grep included egrep and fgrep. The former applies an extended regular expression syntax that was added to Unix after Ken Thompson's original regular expression implementation. The latter searches for any of a list of 'fixed' strings using the Aho-Corasick algorithm. These variants are embodied in most modern grep implementations as command-line switches (and standardized as -E and -F in POSIX). In such combined implementations, grep may also behave differently depending on the name by which it is invoked, allowing fgrep, egrep, and grep to be links to the same program.

    +

    pcregrep is an implementation of grep that uses Perl regular expression syntax.

    +

    Other commands contain the word 'grep' to indicate that they search (usually for regular expression matches). The pgrep utility, for instance, displays the processes whose names match a given regular expression.

    +
    +
    +
    +
    +

    +13.2. tr

    +

    tr (abbreviated from translate or transliterate) is a command in Unix-like operating systems.

    +

    When executed, the program reads from the standard input and writes to the standard output. It takes as parameters two sets of characters, and replaces occurrences of the characters in the first set with the corresponding elements from the other set. For example,

    +
    $ tr 'abcd' 'jkmn'
    +

    maps 'a' to 'j', 'b' to 'k', 'c' to 'm', and 'd' to 'n'.

    +

    Sets of characters may be abbreviated by using character ranges. The previous example could be written:

    +
    $ tr 'a-d' 'jkmn'
    +

    In POSIX compliant versions of tr the set represented by a character range depends on the locale's collating order, so it is safer to avoid character ranges in scripts that might be executed in a locale different from that in which they were written. Ranges can often be replaced with POSIX character sets such as [:alpha:].

    +

    The -c flag complements the first set of characters.

    +
    $ tr -cd '[:alnum:]'
    +

    therefore removes all non-alphanumeric characters.

    +

    The -s flag causes tr to compress sequences of identical adjacent characters in its output to a single token. For example,

    +
    $ tr -s '\n' '\n'
    +

    replaces sequences of one or more newline characters with a single newline.

    +

    The -d flag causes tr to delete all tokens of the specified set of characters from its input. In this case, only a single character set argument is used. The following command removes carriage return characters, thereby converting a file in DOS/Windows format to one in Unix format.

    +
    $ tr -d '\r'
    +

    Most versions of tr, including GNU tr and classic Unix tr, operate on single byte characters and are not Unicode compliant. An exception is the Heirloom Toolchest implementation, which provides basic Unicode support.

    +

    Ruby and Perl also have an internal tr operator, which operates analogously. Tcl's string map command is more general in that it maps strings to strings while tr maps characters to characters.

    +
    +
    +
    +

    +14. Elementary Regex

    +

    In computing, regular expressions provide a concise and flexible means for identifying strings of text of interest, such as particular characters, words, or patterns of characters. A regular expression (often shortened to regex or regexp) is written in a formal language that can be interpreted by a regular expression processor, a program that either serves as a parser generator or examines text and identifies parts that match the provided specification.

    +

    Regular expressions are used by many text editors, utilities, and programming languages to search and manipulate text based on patterns. For example, Perl, Ruby and Tcl have a powerful regular expression engine built directly into their syntax. Several utilities provided by Unix distributions—including the editor ed and the filter grep — were the first to popularize the concept of regular expressions.

    +

    Traditional Unix regular expression syntax followed common conventions but often differed from tool to tool. The IEEE POSIX Basic Regular Expressions (BRE) standard (released alongside an alternative flavor called Extended Regular Expressions or ERE) was designed mostly for backward compatibility with the traditional (Simple Regular Expression) syntax but provided a common standard which has since been adopted as the default syntax of many Unix regular expression tools, though there is often some variation or additional features. Many such tools also provide support for ERE syntax with command line arguments.

    +

    In the BRE syntax, most characters are treated as literals — they match only themselves (i.e., a matches "a"). The exceptions, listed below, are called metacharacters or metasequences.

    ++++
    +
    +

    +14.1. Lazy quantification

    +

    The standard quantifiers in regular expressions are greedy, meaning they match as much as they can, only giving back as necessary to match the remainder of the regex. For example, someone new to regexes wishing to find the first instance of an item between < and > symbols in this example:

    +
    Another whale explosion occurred on <January 26>, <2004>.
    +

    ...would likely come up with the pattern <.*>, or similar. However, this pattern will actually return "<January 26>, <2004>" instead of the "<January 26>" which might be expected, because the <title_reference>*</title_reference> quantifier is greedy — it will consume as many characters as possible from the input, and "January 26>, <2004" has more characters than "January 26".

    +

    Though this problem can be avoided in a number of ways (e.g., by specifying the text that is not to be matched: <[^>]*>), modern regular expression tools allow a quantifier to be specified as lazy (also known as non-greedy, reluctant, minimal, or ungreedy) by putting a question mark after the quantifier (e.g., <.*?>), or by using a modifier which reverses the greediness of quantifiers (though changing the meaning of the standard quantifiers can be confusing). By using a lazy quantifier, the expression tries the minimal match first. Though in the previous example lazy matching is used to select one of many matching results, in some cases it can also be used to improve performance when greedy matching would require more backtracking.

    +
    +
    +
    +

    +15. One Liners

    +

    A one-liner is textual input to the command-line of an operating system shell that performs some function in just one line of input.

    +

    The one liner can be

    +<block_quote>
      +
    1. An expression written in the language of the shell.

    2. +
    3. The invocation of an interpreter together with program source for the interpreter to run.

    4. +
    5. The invocation of a compiler together with source to compile and +instructions for executing the compiled program.

    6. +
    </block_quote>

    Certain dynamic scripting languages such as AWK, sed, and perl have traditionally been adept at expressing one-liners. Specialist shell interpreters such as these Unix shells or the Windows PowerShell, allow for the construction of powerful one-liners.

    +

    The use of the phrase one-liner has been widened to also include program-source for any language that does something useful in one line.

    +

    The word One-liner has two references in the index of the book The AWK Programming Language (the book is often referred to by the abbreviation TAPL). It explains the programming language AWK, which is part of the Unix operating system. The authors explain the birth of the One-liner paradigm with their daily work on early Unix machines:

    +
    “The 1977 version had only a few built-in variables and predefined functions. It was designed for writing short programs [...] Our model was that an invocation would be one or two lines long, typed in and used immediately. Defaults were chosen to match this style [...] We, being the authors, knew how the language was supposed to be used, and so we only wrote one-liners.”
    +

    Notice that this original definition of a One-liner implies immediate execution of the program without any compilation. So, in a strict sense, only source code for interpreted languages qualifies as a One-liner. But this strict understanding of a One-liner was broadened in 1985 when the IOCCC introduced the category of Best One Liner for C, which is a compiled language.

    +

    The TAPL book contains 20 examples of One-liners (A Handful of Useful awk One-Liners) at the end of the book's first chapter.

    +

    Here are the first few of them:

    +<block_quote>
      +
    1. +

      Print the total number of input lines:

      +

      END { print NR }

      +
    2. +
    3. +

      Print the tenth input line:

      +

      NR == 10

      +
    4. +
    5. +

      Print the last field of every input line:

      +

      { print $NF }

      +
    6. +
    </block_quote>

    One-liners are also used to show off the differential expressive power of programming languages. Frequently, one-liners are used to demonstrate programming ability. Contests are often held to see who can create the most exceptional one-liner.

    +

    The following example is a C program (a winning entry in the "Best one-liner" category of the IOCCC, here split to two lines for presentation).:

    +
    main(int c,char**v){return!m(v[1],v[2]);}m(char*s,char*t){return
    +*t-42?*s?63==*t|*s==*t&&m(s+1,t+1):!*t:m(s,t+1)||*s&&m(s+1,t);}
    +

    This one-liner program is a glob pattern matcher. It understands the glob characters '*' meaning 'zero or more characters' and '?' meaning exactly one character, just like most Unix shells.

    +

    Run it with two args, the string and the glob pattern. The exit status is 0 (shell true) when the pattern matches, 1 otherwise. The glob pattern must match the whole string, so you may want to use * at the beginning and end of the pattern if you are looking for something in the middle. Examples:

    +
    $ prog foo 'f??'; echo $?
    +
    +$ prog 'best short program' '??st*o**p?*'; echo $?
    +

    Here is a one line shell script to show directories:

    +
    $ ls -R | grep ":$" | sed -e 's/:$//' -e 's/[^-][^\/]*\//--/g' -e 's/^/   /' -e 's/-/|/'
    +
    +
    +
    + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/ch7interim_assessment.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/ch7interim_assessment.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1139 @@ + + + +Internal assesment  + + + + + + + +
    +
    + +
    +
    +

    +
    +
    + +
    +

    +1. Introducing Linux

    +

    (Attribution : A significant chunk of the content under this section is based on data from Wikipedia and the Linux Documentation Project)

    +

    Linux (usually pronounced ˈlɪnəks') is a generic term referring to Unix-like computer operating systems based on the Linux kernel, where a kernel is the intermediate layer between the hardware and the applications. The kernel is, on an abstract level, the core of (most) operating systems, that manages the various system resources. The development of the Linux OS is considered the basis for Free and Open Source Software (FOSS) collaboration since typically the underlying source code can be used, modified freely, and redistributed by anyone under the terms of the GNU (a recursive acronym for "GNU's Not Unix!") Global Public License (GPL) and other free software licences. This freedom to access and reuse various components of a system, is one of the primary reasons for the popularity of Linux.

    +

    Linux is installed on a variety of computer hardware, that include mobile phones, embedded devices and supercomputers, but is infamous for its use in servers.

    +

    The name "Linux" comes from the Linux kernel, originally written in 1991 by Linus Torvalds. The rest of the system usually comprises components such as the Apache HTTP Server, the X Window System, the GNOME and KDE desktop environments, and utilities and libraries from the GNU Project (announced in 1983 by Richard Stallman). Commonly-used applications with desktop Linux systems include the Mozilla Firefox web-browser and the OpenOffice.org office application suite. The GNU contribution is the basis for the Free Software Foundation's preferred name GNU/Linux. The kernel's mascot is a penguin named "Tux". Mozilla Firefox and OpenOffice.org are open-source projects which can be run on most Operating Systems, including proprietary ones.

    +
    +

    +1.1. Historical Background

    +
    +

    +1.1.1. Events leading to the creation

    +
      +
    • The Unix operating system was developed in the 1960s and released for public use in 1970. Its accessibility and portability caused it to be widely adopted, copied and modified by academic institutions and businesses. Its design became influential to authors of other systems. Other free operating systems include the Berkeley Software Distribution (BSD), developed at the University of California at Berkeley, and MINIX which was released by Andrew S. Tanenbaum. The development and adoption of BSD and MINIX were limited due to various reasons, and this lack of a widely-adopted and free kernel triggered Linus Torvalds into starting his project.

    • +
    • In 1983, Richard Stallman started the GNU project with the goal of creating a free UNIX-like operating system. As part of this work, he wrote the GNU General Public License (GPL). By the early 1990s there was almost enough available software to create a full operating system. However, the GNU kernel, called Hurd, failed to attract enough attention from developers leaving GNU incomplete.

    • +
    +
    +
    +

    +1.1.2. The Creation of Linux

    +

    In 1991, Linus Torvalds began a project at the University of Helsinki that later became the Linux kernel. It was initially a terminal (command-line) emulator, which Torvalds used to access the large UNIX servers of the university. He wrote the program targeting just the hardware he was using and independent of an operating system because he wanted to use the functions of his computer with an 80386 processor. Development was done on Minix using the GNU C compiler. This application is still the main choice for compiling Linux today (although the code can be built with other compilers, such as the Intel C Compiler).

    +

    Torvalds continues to direct the development of the kernel. Stallman heads the Free Software Foundation, which in turn supports the GNU components. Finally, individuals and corporations develop third-party non-GNU components, which constitute a vast body of work and including kernel modules, and user applications and libraries. Linux vendors and communities combine and distribute the kernel, GNU components, and non-GNU components, with additional package management software in the form of Linux distributions.

    +
    +
    +
    +

    +1.2. Design and Implications

    +

    A Linux-based system is a modular Unix-like operating system, deriving much of its basic design from principles established in Unix earlier. Such a system uses a monolithic kernel, called the Linux kernel, which handles process control, networking, and peripheral and file system access. Device drivers are integrated directly with the kernel. Separate projects that interface with the kernel provide much of the system's higher-level functionality. The GNU userland is an important part of most Linux-based systems, providing the most common implementation of the C library, a popular shell, and many of the common Unix tools which carry out many basic operating system tasks. The graphical user interface (or GUI) used by most Linux systems is based on the "X Window System".

    +
    +

    +1.2.1. User Interface

    +

    Users can control a Linux-based system through a command line interface (or CLI), a graphical user interface (or GUI), or through controls attached to the associated hardware (this is common for embedded systems). For desktop systems, the default mode is usually the GUI. On desktop machines, "KDE", "GNOME" and "Xfce" are the most popular user interfaces,though a variety of additional user interfaces exist. Most popular user interfaces run on top of the "X Window System" (or X), which enables a graphical application running on one machine to be displayed and controlled from another in a network.

    +

    A Linux system also provides a CLI of some sort through a shell, which is the traditional way of interacting with a Unix system. A Linux distribution specialized for servers may use the CLI as its only interface. A “headless system” (system run without even a monitor) can be controlled by the command line via a remote-control protocol such as SSH or telnet. The CLI is particularly suited for automation of repetitive or delayed tasks, and provides very simple inter-process communication. A graphical terminal emulator program is often used to access the CLI from a Linux desktop.

    +
    +
    +

    +1.2.2. Development

    +

    The primary difference between Linux and many other popular contemporary operating systems is that the Linux kernel and other components are free and open source software. Linux is not the only such operating system, although it is by far the most widely used. Some free and open source software licenses are based on the principle of "copyleft", a kind of reciprocity: any work derived from a copyleft piece of software must also be copyleft itself. The most common free software license, the GNU GPL, is a form of copyleft, and is used for the Linux kernel and many of the components from the GNU project.

    +

    Linux based distributions are intended by developers for interoperability with other operating systems and established computing standards. Linux systems adhere to POSIX, SUS, ISO and ANSI standards where possible, although to date only one Linux distribution has been POSIX.1 certified, Linux-FT.Free software projects, although developed in a collaborative fashion, are often produced independently of each other. The fact that the software licenses explicitly permit redistribution, however, provides a basis for larger scale projects that collect the software produced by stand-alone projects and make it available all at once in the form of a Linux distribution.

    +

    A Linux distribution, commonly called a "distro", is a project that manages a remote collection of system software and application software packages available for download and installation through a network connection. This allows the user to adapt the operating system to his/her specific needs. Distributions are maintained by individuals, loose-knit teams, volunteer organizations, and commercial entities. A distribution can be installed using a CD that contains distribution-specific software for initial system installation and configuration. A package manager such as Synaptic or YAST allows later package upgrades and installations. A distribution is responsible for the default configuration of the installed Linux kernel, general system security, and more generally integration of the different software packages into a coherent whole.

    +
    +
    +

    +1.2.3. Community

    +

    A distribution is largely driven by its developer and user communities. Some vendors develop and fund their distributions on a volunteer basis. Examples include Debian and the Debian-based, Ubuntu. Others maintain a community version of their commercial distributions, as Red Hat does with Fedora.

    +

    In many cities and regions, local associations known as Linux Users Groups (LUGs) seek to promote their preferred distribution and by extension free software. They hold meetings and provide free demonstrations, training, technical support, and operating system installation to new users. Many Internet communities also provide support to Linux users and developers. Most distributions and free software / open source projects have IRC (Internet Relay Chat) chatrooms or newsgroups. Online forums are another means for support. Linux distributions host mailing lists; commonly there will be a specific topic such as usage or development for a given list. All these can be found simply by running an appropriate search on Google.

    +

    Although Linux distributions are generally available without charge, several large corporations sell, support, and contribute to the development of the components of the system and of free software. These include Dell, IBM, HP, Oracle, Sun Microsystems, Novell, Nokia. A number of corporations, notably Red Hat, have built their entire business around Linux distributions.

    +
    +
    +

    +1.2.4. Can I make a profit out of running a business involving Linux?

    +

    The answer is, "Yes!". The free software licenses, on which the various software packages of a distribution built on the Linux kernel are based, explicitly accommodate and encourage commercialization; the relationship between a Linux distribution as a whole and individual vendors may be seen as symbiotic. One common business model of commercial suppliers is charging for support, especially for business users. A number of companies also offer a specialized business version of their distribution, which adds proprietary support packages and tools to administer higher numbers of installations or to simplify administrative tasks. Another business model is to give away the software in order to sell hardware. Examples of corporations that are extensively (and sometimes exclusively) open-source and Linux-powered , with successful revenue generation models involving these, are Google, SUN, Mozilla, etc.

    +
    +
    +

    +1.2.5. Programming on Linux

    +

    Most Linux distributions support dozens of programming languages. The most common collection of utilities for building both Linux applications and operating system programs is found within the GNU toolchain, which includes the GNU Compiler Collection (GCC) and the GNU build system. Amongst others, GCC provides compilers for Ada, C, C++, Java, and Fortran. The Linux kernel itself is written to be compiled with GCC. Proprietary compilers for Linux include the Intel C++ Compiler, Sun Studio, and IBM XL C/C++ Compiler.

    +

    Most distributions also include support for PHP, Perl, Ruby, Python and other dynamic languages. Examples of languages that are less common, but still supported, are C# via the Mono project, sponsored by Novell, and Scheme. A number of Java Virtual Machines and development kits run on Linux, including the original Sun Microsystems JVM (HotSpot), and IBM's J2SE RE, as well as many open-source projects like Kaffe.

    +

    The two main frameworks for developing graphical applications are those of GNOME and KDE. These projects are based on the GTK+ and Qt widget toolkits, respectively, which can also be used independently of the larger framework. Both support a wide variety of languages. There are a number of Integrated Development Environments (IDEs) available including Anjuta, Code::Blocks, Eclipse, KDevelop, Lazarus, MonoDevelop, NetBeans, and Omnis Studio while the long-established editors Vim and Emacs remain popular.

    +
    +
    +
    +

    +1.3. Reasons for Using Linux

    +
    • Linux is free:

    +

    As in "free beer". Linux can be downloaded in its entirety from the Internet completely for free. No registration fees, no costs per user, free updates, and freely available source code in case you want to change the behavior of your system. +Most of all, Linux is free as in "free speech": +The license commonly used is the GNU Public License (GPL). The license says that anybody who may want to do so, has the right to change Linux and eventually to redistribute a changed version, on the one condition that the code is still available after redistribution. In practice, you are free to grab a kernel image and sell the new code, as long as your customers can still have a copy of that code.

    +
    • Linux is portable to any hardware platform:

    +

    A vendor, who wants to sell a new type of computer and who does not know what kind of OS his/her new machine will run, can take a Linux kernel and make it work on his/her hardware, because documentation related to this activity is freely available.

    +
    • Linux was made to keep on running:

    +

    As with UNIX, a Linux system expects to run without rebooting all the time. That is why a lot of tasks are being executed at night or scheduled automatically for other times, resulting in higher availability during busier periods and a more balanced use of the hardware. This property allows for Linux to be applicable to environments where people do not have the time or the possibility to control their systems constantly.

    +
    • Linux is secure and versatile:

    +

    The security model used in Linux is based on the UNIX idea of security, which is known to be robust and of proven quality. But Linux is not only safe from attacks from the Internet: it will adapt equally to other situations, utilizing the same high standards for security.

    +
    • Linux is scalable:

    +

    From a Palmtop with 2 MB of memory to a petabyte storage cluster with hundreds of nodes: add or remove the appropriate packages and Linux fits all. One does not need a supercomputer anymore,because you can use Linux to do big things using the building blocks provided with the system. If one wants to do little things, such as making an operating system for an embedded processor or just recycling your old 486, Linux will do that as well.

    +
    • The Linux OS and Linux applications have very short debug−times:

    +

    Because Linux has been developed and tested by thousands of people, both errors and people to fix them are found very quickly. It often happens that there are only a couple of hours between discovery and fixing of a bug.

    +
    +
    +
    +

    +2. Getting Started

    +
    +

    +2.1. Logging in, activating the user interface and logging out

    +

    In order to work on a Linux system directly, one needs to provide a user name and password. You always need to authenticate to the system. Most PC−based Linux systems have two basic modes for a system to run in: either quick and clean in text console mode,which includes with mouse, multitasking and multi−user features, or in graphical console mode, which looks better but eats more system resources.

    +
    +

    +2.1.1. Graphical Mode

    +

    This is the default nowadays on most desktop computers. You know you will be connecting to the system using graphical mode when you are first asked for your user name, and then to type your password.

    +

    To log in, make sure the mouse pointer is in the login window, provide your user name and password to the system and click OK or press Enter. +It is generally considered a bad idea to connect (graphically) using the root user name, the system adminstrator's account, since the use of graphics includes running a lot of extra programs, in root's case with a lot of extra permissions. To keep all risks as low as possible, use a normal user account to connect graphically. But there are enough risks to keep this in mind as a general advice, for all use of the root account: only log in as root when extra privileges are required.

    +

    After entering your user name/password combination, it can take a little while before the graphical environment is started, depending on the CPU speed of your computer, on the software you use and on your personal settings.

    +

    To continue, you will need to open a terminal window or xterm for short (X being the name for the underlying software supporting the graphical environment). This program can be found in the Applications−>Utilities->System Tools or Internet menu, depending on what window manager you are using. There might be icons that you can use as a shortcut to get an xterm window as well, and clicking the right mouse button on the desktop background will usually present you with a menu containing a terminal window application.

    +

    While browsing the menus, you will notice that a lot of things can be done without entering commands via the keyboard. For most users, the good old point−n−click method of dealing with the computer will do. But for those who want to enter the "heart" of the system, a tool stronger than a mouse will be required to handle the various tasks. This tool is the shell, and when in graphical mode, we activate our shell by opening a terminal window.

    +

    A terminal window should always show a command prompt when you open one. This terminal shows a standard prompt, which displays the user's login name, and the current working directory, represented by the twiddle (~)

    +

    Another common form for a prompt is this one: +[

    +
    +

    user@host
    +

    dir]

    +

    In the above example, user will be your login name, hosts the name of the machine you are working on, and dir an indication of your current location in the file system. Prompts can display all kinds of information, but they are not part of the commands you are giving to your system. To disconnect from the system in graphical mode, you need to close all terminal windows and other applications. After that, hit the logout icon or find Log Out in the menu. Closing everything is not really necessary, and the system can do this for you, but session management might put all currently open applications back on your screen when you connect again, which takes longer and is not always the desired effect. However, this behavior is configurable.

    +

    When you see the login screen again, asking to enter user name and password, logout was successful.

    +
    +
    +

    +2.1.2. Text Mode

    +

    One is in text mode when the whole screen is black, showing (in most cases white) characters. A text mode login screen typically shows some information about the machine you are working on, the name of the machine and a prompt waiting for you to log in.

    +

    The login is different from a graphical login, in that you have to hit the Enter key after providing your user name, because there are no buttons on the screen that you can click with the mouse. Then you should type your password, followed by another Enter. You will not see any indication that you are entering something, not even an asterisk, and you won't see the cursor move. But this is normal on Linux and is done for security +reasons.

    +

    When the system has accepted you as a valid user, you may get some more information, called the message of the day, which can be anything. Additionally, it is popular on UNIX systems to display a fortune cookie, which contains some general wise or unwise (this is up to you) thoughts. After that, you will be given a shell, indicated with the same prompt that you would get in graphical mode.

    +

    Also in text mode: log in as root only to do setup and configuration that absolutely requires administrator privileges, such as adding users, installing software packages, and performing network and other system configuration. Once you are finished, immediately leave the special account and resume your work as a non−privileged user.

    +

    Logging out is done by entering the logout command, followed by Enter. You are successfully disconnected from the system when you see the login screen again.Don't power−off the computer after logging out. It is not meant to be shut off without application of the proper procedures for halting the system. Powering it off without going through the halting process might cause severe damage!

    +
    +
    +
    +
    +

    +3. Basic Commands

    +
    +

    +3.1. ls

    +

    When invoked without any arguments, ls lists the files in the current working directory. A directory that is not the current working directory can be specified and ls will list the files there. The user also may specify any list of files and directories. In this case, all files and all contents of specified directories will be listed. The name ls is derived from list segments which was used in earlier systems.

    +

    Files whose names start with "." are not listed, unless the -a flag is specified or the files are specified explicitly.

    +

    Without options, ls displays files in a bare format. This bare format however makes it difficult to establish the type, permissions, and size of the files. The most common options to reveal this information or change the list of files are:

    +<block_quote>
      +
    • -l long format, displaying Unix file types, permissions, number of hard links, owner, group, size, date, and filename

    • +
    • -F appends a character revealing the nature of a file, for example, * for an executable, or / for a directory. Regular files have no suffix.

    • +
    • -a lists all files in the given directory, including those whose names start with "." (which are hidden files in Unix). By default, these files are excluded from the list.

    • +
    • -R recursively lists subdirectories. The command ls -R / would therefore list all files.

    • +
    • -d shows information about a symbolic link or directory, rather than about the link's target or listing the contents of a directory.

    • +
    • -t sort the list of files by modification time.

    • +
    • -h print sizes in human readable format. (e.g., 1K, 234M, 2G, etc.)

    • +
    </block_quote>

    In some environments, providing the option --color (for GNU ls) or -G (FreeBSD ls) causes ls to highlight different types of files with different colors, instead of with characters as -F would. To determine what color to use for a file, GNU ls checks the Unix file type, the file permissions, and the file extension, while FreeBSD ls checks only the Unix file type and file permissions.:

    +
    $ ls
    +jeeves.rst psmith.html blandings.html
    +$ ls -l
    +drwxr--r--   1 plum  editors   4096  jeeves
    +-rw-r--r--   1 plum  editors  30405  psmith
    +-r-xr-xr-x   1 plum  plum      8460  blandings
    +

    Here "$" actually is the beginning of the prompt. This is typical in most Unix-based systems.

    +
    +
    +

    +3.2. date

    +

    The Unix date command displays the time and date. The super-user can use it to set the system clock.

    +

    With no options, the date command displays the current date and time, including the abbreviated day name, abbreviated month name, day of the month, the time separated by colons, the timezone name, and the year. For example:

    +
    $date
    +Tue Sep  8 12:01:45 IST 2009
    +

    On some systems to set the current date and time to September 8, 2004 01:22 you type:

    +
    $date --set="20040908 01:22"
    +

    In order to view the various options for the date command, type:

    +
    $man date
    +

    This will take you to the "Manual" page comprising of all the details on the date command. You can return to the terminal from the "man" page by pressing the Esc key in the keyboard and typing ":q" in that order.

    +
    +
    +

    +3.3. cd

    +

    This stands for "change directory". When one wants to go up to the parent directory, bypassing the tree of directories one has entered, “ cd ..” can be used.

    +

    One dot '.' represents the current directory while two dots '..' represent the parent directory.

    +

    “ cd -” will return you to the previous directory (a bit like an “undo”).

    +

    You can also use cd absolute path or cd relative path (see below):

    +

    Absolute paths:

    +<block_quote>

    An “ absolute path” is easily recognised from the leading forward slash, /. The / means that you start at the top level directory and continue down.

    </block_quote>

    For example to get to /boot/grub you would type:

    +
    $cd /boot/grub
    +

    This is an absolute path because you start at the top of the hierarchy and go downwards from there (it doesn't matter where in the filesystem you were when you typed the command).

    +

    Relative paths:

    +<block_quote>

    A “ relative path” doesn't have a preceding slash. Use a relative path when you start from a directory below the top level directory structure. This is dependent on where you are in the filesystem.

    +

    For example if you are in root's home directory and want to get to /root/music, you type:

    +
    $ cd music
    </block_quote>

    Please note that there is no / using the above cd command. Using a / would cause this to be an absolute path, working from the top of the hierarchy downward.

    +
    +
    +

    +3.4. who

    +

    The standard Unix command who displays a list of users who are currently logged into a computer.

    +

    The who command is related to the command w, which provides the same information but also displays additional data and statistics.:

    +
    $who
    +beeblebrox tty7         2009-09-08 10:50 (:0)
    +beeblebrox pts/0        2009-09-08 11:25 (:0.0)
    +dumbledore pts/1        2009-09-08 18:11 (potter.xyz.in)
    +beeblebrox pts/2        2009-09-08 18:53 (:0.0)
    +

    The command can be invoked with the arguments am i or am I (so it is invoked as who am i or * who am I*), showing information about the current terminal only (see the -m option below, of which this invocation is equivalent).

    +

    In order to find out the various options that can be appended to the who command, check the man page by typing out the following in the terminal:

    +
    $man who
    +

    This will take you to the "Manual" page containing details about the who command

    +
    +
    +

    +3.5. mkdir

    +

    This command is used to make a new directory. Normal usage is as straightforward as follows:

    +
    $mkdir name_of_directory
    +

    Where name_of_directory is the name of the directory one wants to create. When typed as above (ie. normal usage), the new directory would be created within the current directory. On Unix, multiple directories can be specified, and mkdir will try to create all of them.

    +
    +

    +3.5.1. Options

    +

    On Unix-like operating systems, mkdir takes options. Three of the most common options are:

    +<block_quote>
      +
    • -p: will also create all directories leading up to the given directory that do not exist already. If the given directory already exists, ignore the error.

    • +
    • -v: display each directory that mkdir creates. Most often used with -p.

    • +
    • -m: specify the octal permissions of directories created by mkdir.

    • +
    </block_quote>

    -p is most often used when using mkdir to build up complex directory hierarchies, in case a necessary directory is missing or already there. -m is commonly used to lock down temporary directories used by shell scripts.

    +
    +
    +

    +3.5.2. Examples

    +

    An example of -p in action is:

    +
    $mkdir -p /tmp/a/b/c
    +

    If /tmp/a exists but /tmp/a/b does not, mkdir will create /tmp/a/b before creating /tmp/a/b/c.

    +

    And an even more powerful command, creating a full tree at once (this however is a Shell extension, nothing mkdir does itself):

    +
    $mkdir -p tmpdir/{trunk/sources/{includes,docs},branches,tags}
    +

    This will create:

    +<definition_list><definition_list_item><term>tmpdir - branches</term><definition>
      +
    • tag

    • +
    • <definition_list><definition_list_item><term>trunk - sources - includes</term><definition>
      • docs

      </definition>
      </definition_list_item>
      </definition_list>
    • +
    </definition>
    </definition_list_item>
    </definition_list>
    +
    +
    +
    +
    +

    +4. Getting Help

    +
    +

    +4.1. apropos and whatis

    +

    This is a command to search the manual pages files in Unix and Unix-like operating systems.

    +
    $ apropos grep
    +egrep       egrep (1)       Search a file for a pattern using full regular expressions
    +fgrep       fgrep (1)       Search a file for a fixed-character string
    +fmlgrep     fmlgrep (1)     Search a file for a pattern
    +grep        grep (1)        Search a file for a pattern
    +gzgrep      gzgrep (1)      Search a possibly compressed file for a regular expression
    +nisgrep     nismatch (1)    Utilities for searching NIS+ tables
    +pgrep       pgrep (1)       Find or signal a process by name or other attribute
    +zgrep       zgrep (1)       Search a possibly compressed file for a regular expression
    +...
    +

    In this example, the user uses apropos to search for the string "grep", and apropos returns the indicated man pages that include the term "grep".

    +

    A short index of explanations for commands is available using the whatis command, like in the examples below:

    +
    $whatis ls
    +ls (1)           - list directory contents
    +

    This displays short information about a command, and the first section in the collection of man pages that contains an appropriate page.

    +

    If you don't know where to get started and which man page to read, apropos gives more information. Say that you do not know how to start a browser, then you could enter the following command:

    +
    $apropos browser
    +gmusicbrowser (1)    - Jukebox for large collections of audio files
    +infobrowser (1)      - read Info documents
    +libsmbclient (7)     - An extension library for browsers and that               can be used...
    +opera (1)            - a standards-compliant graphical Web browser
    +sensible-browser (1) - sensible editing, paging, and web browsing
    +smbtree (1)          - A text based smb network browser
    +tvtk_doc (1)         - A GUI based TVTK documentation search browser.
    +viewres (1)          - graphical class browser for Xt
    +w3m (1)              - a text based Web browser and pager
    +www-browser (1)      - a text based Web browser and pager
    +...
    +
    +
    +

    +4.2. man

    +

    Man pages (short for "manual pages") are the extensive documentation that comes preinstalled with almost all substantial Unix and Unix-like operating systems. The Unix command used to display them is man. Each page is a self-contained document.

    +

    To read a manual page for a Unix command, one can use:

    +
    $ man <command_name>
    +

    at a shell prompt; for example, "man ftp". In order to simplify navigation through the output, man generally uses the less terminal pager.

    +

    Pages are traditionally referred to using the notation "name(section)"; for example, ftp(1). The same page name may appear in more than one section of the manual, this can occur when the names of system calls, user commands, or macro packages coincide. Two examples are man(1) and man(7), or exit(2) and exit(3). The syntax for accessing the non-default manual section varies between different man implementations. On Linux and BSD, for example, the syntax for reading *printf(3) is:

    +
    $man 3 printf
    +

    Another example:

    +
    $man man
    +

    The previous example will take you to the "Manual" page entry about manual pages!

    +
    +

    +4.2.1. Layout

    +

    All man pages follow a common layout that is optimized for presentation on a simple ASCII text display, possibly without any form of highlighting or font control. Sections present may include:

    +<definition_list><definition_list_item><term>NAME</term><definition>

    The name of the command or function, followed by a one-line description of what it does.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SYNOPSIS</term><definition>

    In the case of a command, you get a formal description of how to run it and what command line options it takes. For program functions, a list of the parameters the function takes and which header file contains its definition. For experienced users, this may be all the documentation they need.

    </definition>
    </definition_list_item>
    <definition_list_item><term>DESCRIPTION</term><definition>

    A textual description of the functioning of the command or function.

    </definition>
    </definition_list_item>
    <definition_list_item><term>EXAMPLES</term><definition>

    Some examples of common usage.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SEE ALSO</term><definition>

    A list of related commands or functions.

    </definition>
    </definition_list_item>
    </definition_list>

    Other sections may be present, but these are not well standardized across man pages. Common examples include: OPTIONS, EXIT STATUS, ENVIRONMENT, KNOWN BUGS, FILES, AUTHOR, REPORTING BUGS, HISTORY and COPYRIGHT.

    +

    These days virtually every Unix command line application comes with its man page, and many Unix users perceive a lack of man pages as a sign of low quality; indeed, some projects, such as Debian, go out of their way to write man pages for programs lacking one. Few alternatives to man have enjoyed much popularity, with the possible exception of the GNU project's "info" system, an early and simple hypertext system.

    +

    However, the format of a single page for each application, the lack of classification within the sections and the relatively unsophisticated formatting facilities have motivated the development of alternative documentation systems, such as the previously mentioned "info" system.

    +

    Most Unix GUI applications (particularly those built using the GNOME and KDE development environments) now provide end-user documentation in HTML and include embedded HTML viewers such as yelp for reading the help within the application.

    +

    Usually the man pages are written in English. Translations into other languages can be also available on the system.

    +

    The default format of the man pages is troff, with either the macro package man (appearance oriented) or on some systems mdoc (semantic oriented). This makes it possible to typeset a man page to PostScript, PDF and various other formats for viewing or printing.

    +
    +
    +
    +

    +4.3. info

    +

    info is a software utility which forms a hypertextual, multipage documentation and help viewer working on a command line interface, useful when there is no GUI available.

    +

    The syntax is

    +
    $ info <command_name>
    +

    info processes info files, which are Texinfo formatted files, and presents the documentation as a tree, with simple commands to traverse the tree and to follow cross references. For instance

    +<block_quote>
      +
    • n goes to the next page.

    • +
    • p goes to the previous page.

    • +
    • u goes to the upper page.

    • +
    • l goes to the last(visited) node

    • +
    • To follow a cross reference, the cursor can be moved over a link (a word preceded by a <title_reference>*</title_reference>) and enter pressed.

    • +
    </block_quote>

    info was initially written for use with GNU/Linux and then ported to other Unix-like operating systems.

    +
    +
    +

    +4.4. --help

    +

    Most GNU commands support the --help, which gives a short explanation about how to use the command and a list of available options. Below is the output of this option with the cat command:

    +
    $ userprompt@host: cat --help
    +Usage: cat [OPTION] [FILE]...
    +Concatenate FILE(s), or standard input, to standard output.
    +
    +  -A, --show-all           equivalent to -vET
    +  -b, --number-nonblank    number nonempty output lines
    +  -e                       equivalent to -vE
    +  -E, --show-ends          display $ at end of each line
    +  -n, --number             number all output lines
    +  -s, --squeeze-blank      suppress repeated empty output lines
    +  -t                       equivalent to -vT
    +  -T, --show-tabs          display TAB characters as ^I
    +  -u                       (ignored)
    +  -v, --show-nonprinting   use ^ and M- notation, except for LFD and              TAB
    +  --help     display this help and exit
    +  --version  output version information and exit
    +
    +With no FILE, or when FILE is -, read standard input.
    +
    +Examples:
    +  cat f - g  Output f's contents, then standard input, then g's           contents.
    +  cat        Copy standard input to standard output.
    +
    +Report bugs to <bug-coreutils@gnu.org>.
    +
    +
    +
    +

    +5. Basic file handling

    +
    +

    +5.1. cp

    +

    cp is the command entered in a Unix shell to copy a file from one place to another, possibly on a different filesystem. The original file remains unchanged, and the new file may have the same or a different name.

    +
    +

    +5.1.1. Usage

    +

    To copy a file to another file:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ][ -- ] SourceFile TargetFile
    +

    To copy a file to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -r | -R ] [ -- ] SourceFile ...              TargetDirectory
    +

    To copy a directory to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -- ] { -r | -R }
    +SourceDirectory ... TargetDirectory
    +
    +
    +

    +5.1.2. Flags

    +

    -f (force) – specifies removal of the target file if it cannot be opened for write operations. The removal precedes any copying performed by the cp command.

    +

    -P – makes the cp command copy symbolic links. The default is to follow symbolic links, that is, to copy files to which symbolic links point.

    +

    -i (interactive) – prompts you with the name of a file to be overwritten. This occurs if the TargetDirectory or TargetFile parameter contains a file with the same name as a file specified in the SourceFile or SourceDirectory parameter. If you enter y or the locale's equivalent of y, the cp command continues. Any other answer prevents the cp command from overwriting the file.

    +

    -p (preserve) – duplicates the following characteristics of each SourceFile/SourceDirectory in the corresponding TargetFile and/or TargetDirectory:

    +<block_quote>
      +
    • The time of the last data modification and the time of the last access.

    • +
    • The user ID and group ID (only if it has permissions to do this)

    • +
    • The file permission bits and the SUID and SGID bits.

    • +
    </block_quote>

    -R (recursive) – copy directories (recursively copying all the contents)

    +
    +
    +

    +5.1.3. Examples

    +

    To make a copy of a file in the current directory, enter:

    +
    $ cp prog.c prog.bak
    +

    This copies prog.c to prog.bak. If the prog.bak file does not already exist, the cp command creates it. If it does exist, the cp command replaces it with a copy of the prog.c file.

    +

    To copy a file in your current directory into another directory, enter:

    +
    $ cp zaphod /home/books/hhgg
    +

    This copies the jones file to /home/books/hhgg/zaphod.

    +

    To copy a file to a new file and preserve the modification date, time, and access control list associated with the source file, enter:

    +
    $ cp -p martin_luther_king martin_luther_king.jr
    +

    This copies the martin_luther_king file to the martin_luther_king.jr file. Instead of creating the file with the current date and time stamp, the system gives the martin_luther_king.jr file the same date and time as the martin_luther_king file. The martin_luther_king.jr file also inherits the martin_luther_king file's access control protection.

    +

    To copy all the files in a directory to a new directory, enter:

    +
    $ cp /home/galactica/clients/* /home/hhgg/customers
    +

    This copies only the files in the clients directory to the customers directory.

    +

    To copy a directory, including all its files and subdirectories, to another directory, enter:

    +<block_quote>

    $ cp -R /home/hhgg/clients /home/hhgg/customers

    </block_quote>

    This copies the clients directory, including all its files, subdirectories, and the files in those subdirectories, to the customers/clients directory.

    +

    To copy a specific set of files of any extension to another directory, enter:

    +
    $ cp zaphod arthur ford /home/hhgg/clients
    +

    This copies the zaphod, arthur, and ford files in your current working directory to the /home/hhgg/clients directory.

    +

    To use pattern-matching characters to copy files, enter:

    +
    $ cp programs/*.py .
    +

    This copies the files in the programs directory that end with .py to the current directory, signified by the single "." (dot). You must type a space between the py and the final dot.

    +
    +
    +
    +

    +5.2. mv

    +

    mv (short for move) is a Unix command that moves one or more files or directories from one place to another. The original file is deleted, and the new file may have the same or a different name. If possible (i.e. when the original and new files are on the same file system), mv will rename the file instead. Write permission is required on all directories being modified.

    +
    +

    +5.2.1. Conflicting existing file

    +

    In all cases, when a file is moved to have the name of an existing file (in the same directory), the existing file is deleted. If the existing file is not writable but is in a directory that is writable, then the mv command asks for confirmation if possible (i.e. if run from a terminal) before proceeding, unless the -f (force) option is used.

    +
    +
    +

    +5.2.2. Differences with copy and delete

    +

    Note that, usually, when moving files within the same volume, moving (and/or renaming) is not the same as simply copying and then deleting the original. When moving a file, the link is simply removed from the old parent directory and added to the new parent directory. However, the file itself is untouched (i.e. it has the same inodes and resides at the same place on the disk). For example, you cannot copy a file you cannot read, but you can move (and/or rename) it (provided you have write permission to its old and new parent directories). Also, suppose there is a non-empty directory you do not have write permission to. You cannot delete this directory (since you cannot delete its contents); but you can move (and/or rename) it. Also, since moving between filenames on a single volume does not involve copying, it is faster and does not place strain of lots of reads and writes on the disk. Moving files across different volumes, however, does necessitate copying and deleting.

    +
    +
    +

    +5.2.3. Examples

    +
    $ mv myfile mynewfilename    renames a file
    +$ mv myfile otherfilename    renames a file and deletes the existing            file "myfile"
    +$ mv myfile /myfile          moves 'myfile' from the current            directory to the root directory
    +$ mv myfile dir/myfile       moves 'myfile' to 'dir/myfile' relative            to the current directory
    +$ mv myfile dir              same as the previous command (the          filename is implied to be the same)
    +$ mv myfile dir/myfile2      moves 'myfile' to dir and renames it to            'myfile2'
    +$ mv foo bar baz dir         moves multiple files to directory dir
    +$ mv --help                  shows a very concise help about the                syntax of the command
    +$ man mv                     prints an extensive user manual for                'mv' in the terminal
    +

    In all cases, the file or files being moved or renamed can be a directory.

    +

    Note that when the command is called with two arguments (as mv name1 name2 or mv name1 /dir/name2), it can have three different effects, depending on whether name2 does not exist, is an existing file, or is an existing directory. If the user intends to refer to an existing directory, /. (or in some Unix versions / is sufficient) may be appended to the name to force the system to check this. To move a file to a new directory, the directory must be created first.

    +
    +
    +
    +

    +5.3. rm

    +

    rm (short for "remove") is one of several basic Unix command lines that operates on files. It is used to delete files from a filesystem. The data is not actually destroyed. Only the index listing where the file is stored is destroyed, and the storage is made available for reuse. There are undelete utilities that will attempt to reconstruct the index and can bring the file back if the parts were not reused.

    +

    Here's example to remove a file named "foo" from a directory, here shown with the -i option:

    +
    $ rm -i foo
    +remove foo? y
    +
    +

    +5.3.1. Options

    +

    Common options that rm accepts include:

    +<block_quote>
      +
    • -r, which removes directories, removing the contents recursively beforehand (so as not to leave files without a directory to reside in) ("recursive")

    • +
    • -i, which asks for every deletion to be confirmed ("interactive")

    • +
    • -f, which ignores non-existent files and overrides any confirmation prompts ("force")

    • +
    • -v, which shows what is being removed as it happens ("verbose")

    • +
    </block_quote>

    rm is often aliased to "rm -i" so as to avoid accidental deletion of files. If a user still wishes to delete a large number of files without confirmation, they can manually cancel out the -i argument by adding the -f option (as the option specified later on the expanded command line "rm -i -f" takes precedence).

    +

    rm -rf (variously, rm -rf /, rm -rf <title_reference>*</title_reference>, and others) is frequently used in jokes and anecdotes about Unix disasters. The rm -rf variant of the command, if run by a superuser on the root directory, would cause the contents of every writable mounted filesystem on the computer to be deleted.

    +

    rm is often used in conjunction with xargs to supply a list of files to delete:

    +
    xargs rm < filelist
    +

    When rm is used on a symbolic link, it deletes the link, but does not affect the target of the link.

    +
    +
    +

    +5.3.2. Permissions

    +

    Usually, on most filesystems, deleting a file requires write permission on the parent directory (and execute permission, in order to enter the directory in the first place). (Note that, confusingly for beginners, permissions on the file itself are irrelevant. However, GNU rm asks for confirmation if a write-protected file is to be deleted, unless the -f option is used.)

    +

    To delete a directory (with rm -r), one must delete all of its contents recursively. This requires that one must have read and write and execute permission to that directory (if it's not empty) and all non-empty subdirectories recursively (if there are any). The read permissions are needed to list the contents of the directory in order to delete them. This sometimes leads to an odd situation where a non-empty directory cannot be deleted because one doesn't have write permission to it and so cannot delete its contents; but if the same directory were empty, one would be able to delete it.

    +

    If a file resides in a directory with the sticky bit set, then deleting the file requires one to be the owner of the file.

    +
    +
    +
    +
    +

    +6. Command Line Arguments

    +

    In computer command line interfaces, a command line argument is an argument sent to a program being called. In general, a program can take any number of command line arguments, which may be necessary for the program to run, or may even be ignored, depending on the function of that program.

    +

    For example, in Unix and Unix-like environments, an example of a command-line argument is:

    +
    rm file.s
    +

    "file.s" is a command line argument which tells the program rm to remove the file "file.s".

    +

    Programming languages such as C, C++ and Java allow a program to interpret the command line arguments by handling them as string parameters in the main function.

    +

    A command line option or simply option (also known as a command line parameter, flag, or a switch) is an indication by a user that a computer program should change its default output.

    +

    Long options are introduced via "--", and are typically whole words. For example, ls --long --classify --all. Arguments to long options are provided with "=", as ls --block-size=1024. Some Unix programs use long options with single dashes, for example MPlayer as in mplayer -nosound.

    +

    Linux also uses "--" to terminate option lists. For example, an attempt to delete a file called -file1 by using rm -file1 may produce an error, since rm may interpret -file1 as a command line switch. Using rm -- -file1 removes ambiguity.

    +
    +
    +

    +7. Basic Text Processing

    +
    +

    +7.1. head

    +

    head is a program on Unix and Unix-like systems used to display the first few lines of a text file or piped data. The command syntax is:

    +
    $ head [options] <file_name>
    +

    By default, head will print the first 10 lines of its input to the standard output. The number of lines printed may be changed with a command line option. The following example shows the first 20 lines of filename:

    +
    $ head -n 20 filename
    +

    This displays the first 5 lines of all files starting with foo:

    +
    $ head -n 5 foo*
    +

    Some versions omit the n and just let you say -5.

    +
    +

    +7.1.1. Flags

    +
    -c <x number of bytes> Copy first x number of bytes.
    +

    Other options: sed

    +

    Many early versions of Unix did not have this command, and so documentation and books had sed do this job:

    +
    sed 5q foo
    +

    This says to print every line (implicit), and quit after the fifth.

    +
    +
    +
    +

    +7.2. tail

    +

    tail is a program on Unix and Unix-like systems used to display the last few lines of a text file or piped data.

    +

    The command-syntax is:

    +
    $ tail [options] <file_name>
    +

    By default, tail will print the last 10 lines of its input to the standard output. With command line options the number of lines printed and the printing units (lines, blocks or bytes) may be changed. The following example shows the last 20 lines of filename:

    +
    $ tail -n 20 filename
    +

    This example shows the last 15 bytes of all files starting with foo:

    +
    $ tail -c 15 foo*
    +

    This example shows all lines of filename from the second line onwards:

    +
    $ tail -n +2 filename
    +

    Using an older syntax (still used in Sun Solaris as the -n option is not supported), the last 20 lines and the last 50 bytes of filename can be shown with the following command:

    +
    $ tail -20 filename
    +$ tail -50c filename
    +

    However this syntax is now obsolete and does not conform with the POSIX 1003.1-2001 standard. Even if still supported in current versions, when used with other options (like -f, see below), these switches could not work at all.

    +
    +

    +7.2.1. File monitoring

    +

    tail has a special command line option -f (follow) that allows a file to be monitored. Instead of displaying the last few lines and exiting, tail displays the lines and then monitors the file. As new lines are added to the file by another process, tail updates the display. This is particularly useful for monitoring log files. The following command will display the last 10 lines of messages and append new lines to the display as new lines are added to messages:

    +
    $ tail -f /var/adm/messages
    +

    To interrupt tail while it is monitoring, break-in with Ctrl+C. This command can be run "in the background" with &, see job control.

    +

    If you have a command's result to monitor, you can use the watch command.

    +
    +
    +
    +

    +7.3. cut

    +

    In computing, cut is a Unix command line utility which is used to extract sections from each line of input — usually from a file.

    +

    Extraction of line segments can typically be done by bytes (-b), characters (-c), or fields (-f) separated by a delimiter (-d — the tab character by default). A range must be provided in each case which consists of one of N, N-M, N- (N to the end of the line), or -M (beginning of the line to M), where N and M are counted from 1 (there is no zeroth value). Since version 6, an error is thrown if you include a zeroth value. Prior to this the value was ignored and assumed to be 1.

    +

    Assuming a file named file containing the lines:

    +
    foo:bar:baz:qux:quux
    +one:two:three:four:five:six:seven
    +alpha:beta:gamma:delta:epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +

    To output the fourth through tenth characters of each line:

    +
    $ cut -c 4-10 file
    +

    This gives the output:

    +
    :bar:ba
    +:two:th
    +ha:beta
    +

    To output the fifth field through the end of the line of each line using the colon character as the field delimiter:

    +
    $ cut -d : -f 5- file
    +

    This gives the output:

    +
    quux
    +five:six:seven
    +epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +
    +
    +

    +7.4. paste

    +

    paste is a Unix command line utility which is used to join files horizontally (parallel merging) by outputting lines consisting of the sequentially corresponding lines of each file specified, separated by tabs, to the standard output. It is effectively the horizontal equivalent to the utility cat command which operates on the vertical plane of two or more files.

    +

    To paste several columns of data together into the file www from files who, where, and when:

    +
    $ paste who where when > www
    +

    If the files contain:

    +++++
    +

    This creates the file named www containing:

    +
    Batman            GothamCity       January 3
    +Trillian          Andromeda        February 4
    +Jeeves            London           March 19
    +
    +
    +
    +

    +8. Shell Meta Characters

    +

    Unix recognizes certain special characters, called "meta characters," as command directives. The shell meta characters are recognized anywhere they appear in the command line, even if they are not surrounded by blank space. For that reason, it is safest to only use the characters A-Z, a-z, 0-9, and the period, dash, and underscore characters when naming files and directories on Unix. If your file or directory has a shell meta character in the name, you will find it difficult to use the name in a shell command.

    +

    The shell meta characters include:

    +

    / < > ! $ % ^ & * | { } [ ] " ' ` ~ ;

    +

    Different shells may differ in the meta characters recognized.

    +

    As an example,

    +
    $ ls file.*
    +

    run on a directory containing the files file, file.c, file.lst, and myfile would list the files file.c and file.lst. However,:

    +
    $ ls file.?
    +

    run on the same directory would only list file.c because the ? only matches one character, no more, no less. This can save you a great deal of typing time. For example, if there is a file called california_cornish_hens_with_wild_rice and no other files whose names begin with 'c', you could view the file without typing the whole name by typing this:

    +
    $ more c*
    +

    because the c* matches that long file name.

    +

    Filenames containing metacharacters can pose many problems and should never be intentionally created. If you do find that you've created a file with metacharacters, and you would like to remove it, you have three options. You may use wildcards to match metacharacter, use the to directly enter the filename, or put the command in double quotes (except in the case of double quotes within the file name, these must be captured with one of the first two methods). For example, deleting a file named <title_reference>"``*`|more</title_reference>"` can be accomplished with:

    +
    $ rm ??more
    +

    or:

    +
    $ rm $\backslash$*$\backslash$|more
    +

    or:

    +
    $ rm ''*|more''
    +
    +
    +

    +9. Looking At Files

    +
    +

    +9.1. cat

    +

    The cat command is a standard Unix program used to concatenate and display files. The name is from "catenate", a synonym of concatenate.

    +

    The Single Unix Specification specifies the behavior that the contents of each of the files given in sequence as arguments will be written to the standard output in the same sequence, and mandates one option, -u, where each byte is printed as it is read.

    +

    If the filename is specified as -, then cat will read from standard input at that point in the sequence. If no files are specified, cat will read from standard input entered.

    +
    +

    +9.1.1. Jargon File Definition

    +

    The Jargon File version 4.4.7 lists this as the definition of cat:

    +
    1. To spew an entire file to the screen or some other output sink without
    +     pause (syn. blast).
    +
    +2. By extension, to dump large amounts of data at an unprepared target or
    +     with no intention of browsing it carefully. Usage: considered silly.
    +     Rare outside Unix sites. See also dd, BLT.
    +
    +     Among Unix fans, *cat(1)* is considered an excellent example of
    +     user-interface design, because it delivers the file contents without
    +     such verbosity as spacing or headers between the files, and because
    +     it does not require the files to consist of lines of text, but works
    +     with any sort of data.
    +
    +     Among Unix critics, *cat(1)* is considered the canonical example of
    +     bad user-interface design, because of its woefully unobvious name.
    +     It is far more often used to blast a single file to standard output
    +     than to concatenate two or more files. The name cat for the former
    +     operation is just as unintuitive as, say, LISP's cdr.
    +
    +     Of such oppositions are holy wars made...
    +
    +
    +

    +9.1.2. Useless Use of 'cat'

    +

    UUOC (from comp.unix.shell on Usenet) stands for “Useless Use of cat”. As it is observed on comp.unix.shell, “The purpose of cat is to concatenate (or 'catenate') files. If it's only one file, concatenating it with nothing at all is a waste of time, and costs you a process.”

    +

    Nevertheless one sees people doing:

    +
    $ cat file | some_command and its args ...
    +

    instead of the equivalent and cheaper:

    +
    <file some_command and its args ...
    +

    or (equivalently and more classically):

    +
    some_command and its args ... <file
    +

    Since 1995, occasional awards for UUOC have been given out. The activity of fixing instances of UUOC is sometimes called 'demoggification'.

    +

    Amongst many, it is still considered safer to use cat for such cases given that the < and > keys are next to each other in many popular keyboard mappings. While the risk might be low, the impact of using > instead of < can be high and prohibitive.

    +
    +
    +

    +9.1.3. zcat

    +

    zcat is a Unix program similar to cat, that decompresses individual files and concatenates them to standard output. Traditionally zcat operated on files compressed by compress but today it is usually able to operate on gzip or even bzip2 archives. On such systems, it is equivalent to gunzip -c

    +
    +
    +
    +

    +9.2. more

    +

    In computing, more is a command to view (but not modify) the contents of a text file one screen at a time (terminal pager). It is available on Unix and Unix-like systems, DOS, OS/2 and Microsoft Windows. Programs of this sort are called pagers.

    +
    +

    +9.2.1. Usage

    +

    The command-syntax is:

    +
    $ more [options] [file_name]
    +

    If no file name is provided, more looks for input from stdin.

    +

    Once more has obtained input, it displays as much as can fit on the current screen and waits for user input to advance, with the exception that a form feed (^L) will also cause more to wait at that line, regardless of the amount of text on the screen. In the lower-left corner of the screen is displayed the text "--More--" and a percentage, representing the percent of the file that more has paged through. (This percentage includes the text displayed on the current screen.) When more reaches the end of a file (100%) it exits. The most common methods of navigating through a file are Enter, which advances the output by one line, and Space, which advances the output by one screen.

    +

    There are also other commands that can be used while navigating through the document; consult more's man page for more details.

    +

    Options are typically entered before the file name, but can also be entered in the environment variable $MORE. Options entered in the actual command line will override those entered in the $MORE environment variable. Available options may vary between Unix systems.

    +
    +
    +
    +

    +9.3. less

    +

    less is a terminal pager program on Unix, Windows and Unix-like systems used to view (but not change) the contents of a text file one screen at a time. It is similar to more, but has the extended capability of allowing both forward and backward navigation through the file. Unlike most Unix text editors/viewers, less does not need to read the entire file before starting, resulting in faster load times with large files.

    +
    +

    +9.3.1. Usage

    +

    less can be invoked with options to change its behaviour, for example, the number of lines to display on the screen. A few options vary depending on the operating system. While less is displaying the file, various commands can be used to navigate through the file. These commands are based on those used by both more and vi. It is also possible to search for character patterns in the file.

    +

    By default, less displays the contents of the file to the standard output (one screen at a time). If the file name argument is omitted, it displays the contents from standard input (usually the output of another command through a pipe). If the output is redirected to anything other than a terminal, for example a pipe to another command, less behaves like cat.

    +

    The command-syntax is:

    +
    $ less [options] file_name
    +
    +
    +

    +9.3.2. Frequently Used Options

    +<block_quote>
      +
    • -g: Highlights just the current match of any searched string.

    • +
    • -I: Case-insensitive searches.

    • +
    • -M: Shows more detailed prompt, including file position.

    • +
    • -N: Shows line numbers (useful for source code viewing).

    • +
    • -S: Disables line wrap ("chop long lines"). Long lines can be seen by side scrolling.

    • +
    • -?: Shows help.

    • +
    </block_quote>
    +
    +
    +

    +9.3.3. Frequently Used Commands

    +<block_quote>
      +
    • [Arrows]/[Page Up]/[Page Down]/[Home]/[End]: Navigation.

    • +
    • [Space bar]: Next page.

    • +
    • b: Previous page.

    • +
    • ng: Jump to line number n. Default is the start of the file.

    • +
    • nG: Jump to line number n. Default is the end of the file.

    • +
    • /pattern: Search for pattern. Regular expressions can be used.

    • +
    • '^ or g: Go to start of file.

    • +
    • '$ or G: Go to end of file.

    • +
    • s: Save current content (got from another program like grep) in a file.

    • +
    • =: File information.

    • +
    • h: Help.

    • +
    • q: Quit.

    • +
    </block_quote>
    +
    +
    +

    +9.3.4. Examples

    +
    $ less -M readme.txt                     #Read "readme.txt."
    +$ less +F /var/log/mail.log              #Follow mode for log
    +$ file * | less                          #Easier file analysis.
    +$ grep -i void *.c | less -I -p void     #Case insensitive search                                                         for "void" in all .c files
    +
    +
    +
    +
    +

    +10. Directory Structure

    +

    In the File Hierarchy Standard (FHS) all files and directories appear under the root directory "/", even if they are stored on different physical devices. Note however that some of these directories may or may not be present on a Unix system depending on whether certain subsystems, such as the X Window System, are installed.

    +

    The majority of these directories exist in all UNIX operating systems and are generally used in much the same way; however, the descriptions here are those used specifically for the FHS, and are not considered authoritative for platforms other than Linux.

    ++++
    +
    +

    +10.1. man hier

    +

    This is the manual page on the UNIX filesystem. The syntax for this is:

    +
    $ man hier
    +
    +
    +

    +10.2. ls -l

    +

    Shows you huge amounts of information (permissions, owners, size, and when last modified) for folders and files. The syntax is

    +
    $ ls -l
    +

    This can be done after entering the required directory.

    +
    +
    +
    +

    +11. Permissions and Ownership

    +
    +

    +11.1. chmod

    +

    The chmod command (abbreviated from 'change mode') is a shell command and C language function in Unix and Unix-like environments. When executed, it can change file system modes of files and directories. The modes include permissions and special modes.A chmod command first appeared in AT&T Unix version 1, and is still used today on Unix-like machines.

    +
    +

    +11.1.1. Usage

    +

    The chmod command options are specified like this:

    +
    $ chmod [options] mode[,mode] file1 [file2 ...]
    +

    To view what the permissions currently are, type:

    +
    $ ls -l file
    +
    +
    +

    +11.1.2. Command line options

    +

    The chmod command has a number of command line options that affect its behavior. The most common options are:

    +<block_quote>
      +
    • -R: Changes the modes of directories and files recursively

    • +
    • -v: Verbose mode; lists all files as they are being processed

    • +
    </block_quote>
    +
    +11.1.2.1. Symbolic modes
    +

    To the chmod utility, all permissions and special modes are represented by its mode parameter. One way to adjust the mode of files or directories is to specify a symbolic mode. The symbolic mode is composed of three components, which are combined to form a single string of text:

    +
    $ chmod [references][operator][modes] file1 ...
    +

    The references (or classes) are used to distinguish the users to whom the permissions apply. If no references are specified it defaults to “all” but modifies only the permissions allowed by the umask. The references are represented by one or more of the following letters:

    +++++
    +

    The chmod program uses an operator to specify how the modes of a file should be adjusted. The following operators are accepted:

    ++++
    +

    The modes indicate which permissions are to be granted or taken away from the specified classes. There are three basic modes which correspond to the basic permissions:

    +++++
    +

    The combination of these three components produces a string that is understood by the chmod command. Multiple changes can be specified by separating multiple symbolic modes with commas.

    +
    +
    +
    +11.1.2.2. Symbolic examples
    +

    Add the 'read' and 'write' permissions to the 'user' and 'group' classes of a directory:

    +
    $ chmod ug+rw mydir
    +$ ls -ld mydir
    +drw-rw----   2 starwars  yoda  96 Dec 8 12:53 mydir
    +

    For a file, remove write permissions for all classes:

    +
    $ chmod a-w myfile
    +$ ls -l myfile
    +-r-xr-xr-x   2 starwars  yoda 96 Dec 8 12:53 myfile
    +

    Set the permissions for the u*ser and the *g*roup to read and execute only (no write permission) on *mydir.

    +
    $ chmod ug=rx mydir
    +$ ls -ld mydir
    +dr-xr-x---   2 starwars  yoda 96 Dec 8 12:53 mydir
    +
    +
    +
    +11.1.2.3. Octal numbers
    +

    The chmod command also accepts three and four-digit octal numbers representing modes. Using a three-digit octal number to set the modes of a file named myfile :

    +
    $ chmod 664 myfile
    +$ ls -l myfile
    +-rw-rw-r--  1   57 Jul  3 10:13  myfile
    +

    Since the setuid, setgid and sticky bits are not set, this is equivalent to:

    +
    $ chmod 0664 myfile
    +
    +
    +
    +11.1.2.4. Special modes
    +

    The chmod command is also capable of changing the additional permissions or special modes of a file or directory. The symbolic modes use s to represent the setuid and setgid modes, and t to represent the sticky mode. The modes are only applied to the appropriate classes, regardless of whether or not other classes are specified.

    +

    Most operating systems support the specification of special modes using octal modes, but some do not. On these systems, only the symbolic modes can be used.

    +
    +
    +
    +
    +
    +

    +12. Redirection and Piping

    +

    In computing, redirection is a function common to most command-line interpreters, including the various Unix shells that can redirect standard streams to user-specified locations.

    +

    Programs do redirection with the dup2(2) system call, or its less-flexible but higher-level stdio analogues, freopen(3) and popen(3).

    +
    +

    +12.1. Redirecting standard input and standard output

    +

    Redirection is usually implemented by placing certain characters between commands. Typically, the syntax of these characters is as follows:

    +
    $ command1 > file1
    +

    executes command1, placing the output in file1. Note that this will truncate any existing data in file1. To append output to the end of the file, use the >> operator.:

    +
    $ command1 < file1
    +

    executes command1, using file1 as the source of input (as opposed to the keyboard).:

    +
    $ command1 < infile > outfile
    +

    combines the two capabilities: command1 reads from infile and writes to outfile

    +
    +
    +

    +12.2. Piping

    +

    Programs can be run together such that one program reads the output from another with no need for an explicit intermediate file: +A pipeline of three programs run on a text terminal:

    +
    $ command1 | command2
    +

    executes command1, using its output as the input for command2 (commonly called piping, since the "|" character is known as a "pipe").

    +

    This is equivalent to using two redirects and a temporary file:

    +
    $ command1 > tempfile
    +$ command2 < tempfile
    +$ rm tempfile
    +

    A good example for command piping is combining echo with another command to achieve something interactive in a non-interactive shell, e.g.:

    +
    $ echo -e "user\npass" | ftp localhost
    +

    This runs the ftp client with input user, press return, then pass.

    +
    +
    +

    +12.3. Redirecting to and from the standard file handles

    +

    In Unix shells derived from the original Bourne shell, the first two actions can be further modified by placing a number (the file descriptor) immediately before the character; this will affect which stream is used for the redirection. The Unix standard I/O streams are:

    +++++
    +

    For example:

    +
    $ command1 2> file1
    +

    executes command1, directing the standard error stream to file1.

    +

    In shells derived from csh (the C shell), the syntax instead appends the & character to the redirect characters, thus achieving a similar result.

    +

    Another useful capability is to redirect one standard file handle to another. The most popular variation is to merge standard error into standard output so error messages can be processed together with (or alternately to) the usual output. Example:

    +
    $ find / -name .profile > results 2>&1
    +

    will try to find all files named .profile. Executed without redirection, it will output hits to stdout and errors (e.g. for lack of privilege to traverse protected directories) to stderr. If standard output is directed to file results, error messages appear on the console. To see both hits and error messages in file results, merge stderr (handle 2) into stdout (handle 1) using 2>&1 .

    +

    It's possible use 2>&1 before ">" but it doesn't work. In fact, when the interpreter reads 2>&1, it doesn't know yet where standard output is redirected and then standard error isn't merged.

    +

    If the merged output is to be piped into another program, the file merge sequence 2>&1 must precede the pipe symbol, thus:

    +
    $ find / -name .profile 2>&1 | less
    +

    A simplified form of the command:

    +
    $ command > file 2>&1
    +

    is:

    +
    $ command &>file
    +

    or:

    +
    $command >&file
    +
    +
    +

    +12.4. Chained pipelines

    +

    The redirection and piping tokens can be chained together to create complex commands. For example:

    +
    $ ls | grep '\.sh' | sort > shlist
    +

    lists the contents of the current directory, where this output is filtered to only contain lines which contain .sh, sort this resultant output lexicographically, and place the final output in shlist. This type of construction is used very commonly in shell scripts and batch files.

    +
    +
    +

    +12.5. Redirect to multiple outputs

    +

    The standard command tee can redirect output from a command to several destinations.

    +
    $ ls -lrt | tee xyz
    +

    This directs the file list output to both standard output as well as to the file xyz.

    +
    +
    +
    +

    +13. More Text Processing

    +
    +

    +13.1. grep

    +

    grep is a command line text search utility originally written for Unix. The name is taken from the first letters in global / regular expression / print, a series of instructions for the ed text editor. The grep command searches files or standard input globally for lines matching a given regular expression, and prints them to the program's standard output.

    +
    +

    +13.1.1. Usage

    +

    This is an example of a common grep usage:

    +
    $ grep apple fruitlist.txt
    +

    In this case, grep prints all lines containing 'apple' from the file fruitlist.txt, regardless of word boundaries; therefore lines containing 'pineapple' or 'apples' are also printed. The grep command is case sensitive by default, so this example's output does not include lines containing 'Apple' (with a capital A) unless they also contain 'apple'.

    +

    Like most Unix commands, grep accepts command line arguments to change this and many other behaviors. For example:

    +
    $ grep -i apple fruitlist.txt
    +

    This prints all lines containing 'apple' regardless of capitalization. The '-i' argument tells grep to be case insensitive, or to ignore case.

    +

    To print all lines containing 'apple' as a word ('pineapple' and 'apples' will not match):

    +
    $ grep -w apple fruitlist.txt
    +

    Regular expressions can be used to match more complicated queries.

    +
    +
    +13.1.1.1. Variations
    +

    There are countless implementations and derivatives of grep available for many operating systems. Early variants of grep included egrep and fgrep. The former applies an extended regular expression syntax that was added to Unix after Ken Thompson's original regular expression implementation. The latter searches for any of a list of 'fixed' strings using the Aho-Corasick algorithm. These variants are embodied in most modern grep implementations as command-line switches (and standardized as -E and -F in POSIX). In such combined implementations, grep may also behave differently depending on the name by which it is invoked, allowing fgrep, egrep, and grep to be links to the same program.

    +

    pcregrep is an implementation of grep that uses Perl regular expression syntax.

    +

    Other commands contain the word 'grep' to indicate that they search (usually for regular expression matches). The pgrep utility, for instance, displays the processes whose names match a given regular expression.

    +
    +
    +
    +
    +

    +13.2. tr

    +

    tr (abbreviated from translate or transliterate) is a command in Unix-like operating systems.

    +

    When executed, the program reads from the standard input and writes to the standard output. It takes as parameters two sets of characters, and replaces occurrences of the characters in the first set with the corresponding elements from the other set. For example,

    +
    $ tr 'abcd' 'jkmn'
    +

    maps 'a' to 'j', 'b' to 'k', 'c' to 'm', and 'd' to 'n'.

    +

    Sets of characters may be abbreviated by using character ranges. The previous example could be written:

    +
    $ tr 'a-d' 'jkmn'
    +

    In POSIX compliant versions of tr the set represented by a character range depends on the locale's collating order, so it is safer to avoid character ranges in scripts that might be executed in a locale different from that in which they were written. Ranges can often be replaced with POSIX character sets such as [:alpha:].

    +

    The -c flag complements the first set of characters.

    +
    $ tr -cd '[:alnum:]'
    +

    therefore removes all non-alphanumeric characters.

    +

    The -s flag causes tr to compress sequences of identical adjacent characters in its output to a single token. For example,

    +
    $ tr -s '\n' '\n'
    +

    replaces sequences of one or more newline characters with a single newline.

    +

    The -d flag causes tr to delete all tokens of the specified set of characters from its input. In this case, only a single character set argument is used. The following command removes carriage return characters, thereby converting a file in DOS/Windows format to one in Unix format.

    +
    $ tr -d '\r'
    +

    Most versions of tr, including GNU tr and classic Unix tr, operate on single byte characters and are not Unicode compliant. An exception is the Heirloom Toolchest implementation, which provides basic Unicode support.

    +

    Ruby and Perl also have an internal tr operator, which operates analogously. Tcl's string map command is more general in that it maps strings to strings while tr maps characters to characters.

    +
    +
    +
    +

    +14. Elementary Regex

    +

    In computing, regular expressions provide a concise and flexible means for identifying strings of text of interest, such as particular characters, words, or patterns of characters. A regular expression (often shortened to regex or regexp) is written in a formal language that can be interpreted by a regular expression processor, a program that either serves as a parser generator or examines text and identifies parts that match the provided specification.

    +

    Regular expressions are used by many text editors, utilities, and programming languages to search and manipulate text based on patterns. For example, Perl, Ruby and Tcl have a powerful regular expression engine built directly into their syntax. Several utilities provided by Unix distributions—including the editor ed and the filter grep — were the first to popularize the concept of regular expressions.

    +

    Traditional Unix regular expression syntax followed common conventions but often differed from tool to tool. The IEEE POSIX Basic Regular Expressions (BRE) standard (released alongside an alternative flavor called Extended Regular Expressions or ERE) was designed mostly for backward compatibility with the traditional (Simple Regular Expression) syntax but provided a common standard which has since been adopted as the default syntax of many Unix regular expression tools, though there is often some variation or additional features. Many such tools also provide support for ERE syntax with command line arguments.

    +

    In the BRE syntax, most characters are treated as literals — they match only themselves (i.e., a matches "a"). The exceptions, listed below, are called metacharacters or metasequences.

    ++++
    +
    +

    +14.1. Lazy quantification

    +

    The standard quantifiers in regular expressions are greedy, meaning they match as much as they can, only giving back as necessary to match the remainder of the regex. For example, someone new to regexes wishing to find the first instance of an item between < and > symbols in this example:

    +
    Another whale explosion occurred on <January 26>, <2004>.
    +

    ...would likely come up with the pattern <.*>, or similar. However, this pattern will actually return "<January 26>, <2004>" instead of the "<January 26>" which might be expected, because the <title_reference>*</title_reference> quantifier is greedy — it will consume as many characters as possible from the input, and "January 26>, <2004" has more characters than "January 26".

    +

    Though this problem can be avoided in a number of ways (e.g., by specifying the text that is not to be matched: <[^>]*>), modern regular expression tools allow a quantifier to be specified as lazy (also known as non-greedy, reluctant, minimal, or ungreedy) by putting a question mark after the quantifier (e.g., <.*?>), or by using a modifier which reverses the greediness of quantifiers (though changing the meaning of the standard quantifiers can be confusing). By using a lazy quantifier, the expression tries the minimal match first. Though in the previous example lazy matching is used to select one of many matching results, in some cases it can also be used to improve performance when greedy matching would require more backtracking.

    +
    +
    +
    +

    +15. One Liners

    +

    A one-liner is textual input to the command-line of an operating system shell that performs some function in just one line of input.

    +

    The one liner can be

    +<block_quote>
      +
    1. An expression written in the language of the shell.

    2. +
    3. The invocation of an interpreter together with program source for the interpreter to run.

    4. +
    5. The invocation of a compiler together with source to compile and +instructions for executing the compiled program.

    6. +
    </block_quote>

    Certain dynamic scripting languages such as AWK, sed, and perl have traditionally been adept at expressing one-liners. Specialist shell interpreters such as these Unix shells or the Windows PowerShell, allow for the construction of powerful one-liners.

    +

    The use of the phrase one-liner has been widened to also include program-source for any language that does something useful in one line.

    +

    The word One-liner has two references in the index of the book The AWK Programming Language (the book is often referred to by the abbreviation TAPL). It explains the programming language AWK, which is part of the Unix operating system. The authors explain the birth of the One-liner paradigm with their daily work on early Unix machines:

    +
    “The 1977 version had only a few built-in variables and predefined functions. It was designed for writing short programs [...] Our model was that an invocation would be one or two lines long, typed in and used immediately. Defaults were chosen to match this style [...] We, being the authors, knew how the language was supposed to be used, and so we only wrote one-liners.”
    +

    Notice that this original definition of a One-liner implies immediate execution of the program without any compilation. So, in a strict sense, only source code for interpreted languages qualifies as a One-liner. But this strict understanding of a One-liner was broadened in 1985 when the IOCCC introduced the category of Best One Liner for C, which is a compiled language.

    +

    The TAPL book contains 20 examples of One-liners (A Handful of Useful awk One-Liners) at the end of the book's first chapter.

    +

    Here are the first few of them:

    +<block_quote>
      +
    1. +

      Print the total number of input lines:

      +

      END { print NR }

      +
    2. +
    3. +

      Print the tenth input line:

      +

      NR == 10

      +
    4. +
    5. +

      Print the last field of every input line:

      +

      { print $NF }

      +
    6. +
    </block_quote>

    One-liners are also used to show off the differential expressive power of programming languages. Frequently, one-liners are used to demonstrate programming ability. Contests are often held to see who can create the most exceptional one-liner.

    +

    The following example is a C program (a winning entry in the "Best one-liner" category of the IOCCC, here split to two lines for presentation).:

    +
    main(int c,char**v){return!m(v[1],v[2]);}m(char*s,char*t){return
    +*t-42?*s?63==*t|*s==*t&&m(s+1,t+1):!*t:m(s,t+1)||*s&&m(s+1,t);}
    +

    This one-liner program is a glob pattern matcher. It understands the glob characters '*' meaning 'zero or more characters' and '?' meaning exactly one character, just like most Unix shells.

    +

    Run it with two args, the string and the glob pattern. The exit status is 0 (shell true) when the pattern matches, 1 otherwise. The glob pattern must match the whole string, so you may want to use * at the beginning and end of the pattern if you are looking for something in the middle. Examples:

    +
    $ prog foo 'f??'; echo $?
    +
    +$ prog 'best short program' '??st*o**p?*'; echo $?
    +

    Here is a one line shell script to show directories:

    +
    $ ls -R | grep ":$" | sed -e 's/:$//' -e 's/[^-][^\/]*\//--/g' -e 's/^/   /' -e 's/-/|/'
    +
    +
    +
    + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/ch8ult_module_plan.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/ch8ult_module_plan.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1139 @@ + + + +ULT Module Plan  + + + + + + + +
    +
    + +
    +
    +

    +
    +
    + +
    +

    +1. Introducing Linux

    +

    (Attribution : A significant chunk of the content under this section is based on data from Wikipedia and the Linux Documentation Project)

    +

    Linux (usually pronounced ˈlɪnəks') is a generic term referring to Unix-like computer operating systems based on the Linux kernel, where a kernel is the intermediate layer between the hardware and the applications. The kernel is, on an abstract level, the core of (most) operating systems, that manages the various system resources. The development of the Linux OS is considered the basis for Free and Open Source Software (FOSS) collaboration since typically the underlying source code can be used, modified freely, and redistributed by anyone under the terms of the GNU (a recursive acronym for "GNU's Not Unix!") Global Public License (GPL) and other free software licences. This freedom to access and reuse various components of a system, is one of the primary reasons for the popularity of Linux.

    +

    Linux is installed on a variety of computer hardware, that include mobile phones, embedded devices and supercomputers, but is infamous for its use in servers.

    +

    The name "Linux" comes from the Linux kernel, originally written in 1991 by Linus Torvalds. The rest of the system usually comprises components such as the Apache HTTP Server, the X Window System, the GNOME and KDE desktop environments, and utilities and libraries from the GNU Project (announced in 1983 by Richard Stallman). Commonly-used applications with desktop Linux systems include the Mozilla Firefox web-browser and the OpenOffice.org office application suite. The GNU contribution is the basis for the Free Software Foundation's preferred name GNU/Linux. The kernel's mascot is a penguin named "Tux". Mozilla Firefox and OpenOffice.org are open-source projects which can be run on most Operating Systems, including proprietary ones.

    +
    +

    +1.1. Historical Background

    +
    +

    +1.1.1. Events leading to the creation

    +
      +
    • The Unix operating system was developed in the 1960s and released for public use in 1970. Its accessibility and portability caused it to be widely adopted, copied and modified by academic institutions and businesses. Its design became influential to authors of other systems. Other free operating systems include the Berkeley Software Distribution (BSD), developed at the University of California at Berkeley, and MINIX which was released by Andrew S. Tanenbaum. The development and adoption of BSD and MINIX were limited due to various reasons, and this lack of a widely-adopted and free kernel triggered Linus Torvalds into starting his project.

    • +
    • In 1983, Richard Stallman started the GNU project with the goal of creating a free UNIX-like operating system. As part of this work, he wrote the GNU General Public License (GPL). By the early 1990s there was almost enough available software to create a full operating system. However, the GNU kernel, called Hurd, failed to attract enough attention from developers leaving GNU incomplete.

    • +
    +
    +
    +

    +1.1.2. The Creation of Linux

    +

    In 1991, Linus Torvalds began a project at the University of Helsinki that later became the Linux kernel. It was initially a terminal (command-line) emulator, which Torvalds used to access the large UNIX servers of the university. He wrote the program targeting just the hardware he was using and independent of an operating system because he wanted to use the functions of his computer with an 80386 processor. Development was done on Minix using the GNU C compiler. This application is still the main choice for compiling Linux today (although the code can be built with other compilers, such as the Intel C Compiler).

    +

    Torvalds continues to direct the development of the kernel. Stallman heads the Free Software Foundation, which in turn supports the GNU components. Finally, individuals and corporations develop third-party non-GNU components, which constitute a vast body of work and including kernel modules, and user applications and libraries. Linux vendors and communities combine and distribute the kernel, GNU components, and non-GNU components, with additional package management software in the form of Linux distributions.

    +
    +
    +
    +

    +1.2. Design and Implications

    +

    A Linux-based system is a modular Unix-like operating system, deriving much of its basic design from principles established in Unix earlier. Such a system uses a monolithic kernel, called the Linux kernel, which handles process control, networking, and peripheral and file system access. Device drivers are integrated directly with the kernel. Separate projects that interface with the kernel provide much of the system's higher-level functionality. The GNU userland is an important part of most Linux-based systems, providing the most common implementation of the C library, a popular shell, and many of the common Unix tools which carry out many basic operating system tasks. The graphical user interface (or GUI) used by most Linux systems is based on the "X Window System".

    +
    +

    +1.2.1. User Interface

    +

    Users can control a Linux-based system through a command line interface (or CLI), a graphical user interface (or GUI), or through controls attached to the associated hardware (this is common for embedded systems). For desktop systems, the default mode is usually the GUI. On desktop machines, "KDE", "GNOME" and "Xfce" are the most popular user interfaces,though a variety of additional user interfaces exist. Most popular user interfaces run on top of the "X Window System" (or X), which enables a graphical application running on one machine to be displayed and controlled from another in a network.

    +

    A Linux system also provides a CLI of some sort through a shell, which is the traditional way of interacting with a Unix system. A Linux distribution specialized for servers may use the CLI as its only interface. A “headless system” (system run without even a monitor) can be controlled by the command line via a remote-control protocol such as SSH or telnet. The CLI is particularly suited for automation of repetitive or delayed tasks, and provides very simple inter-process communication. A graphical terminal emulator program is often used to access the CLI from a Linux desktop.

    +
    +
    +

    +1.2.2. Development

    +

    The primary difference between Linux and many other popular contemporary operating systems is that the Linux kernel and other components are free and open source software. Linux is not the only such operating system, although it is by far the most widely used. Some free and open source software licenses are based on the principle of "copyleft", a kind of reciprocity: any work derived from a copyleft piece of software must also be copyleft itself. The most common free software license, the GNU GPL, is a form of copyleft, and is used for the Linux kernel and many of the components from the GNU project.

    +

    Linux based distributions are intended by developers for interoperability with other operating systems and established computing standards. Linux systems adhere to POSIX, SUS, ISO and ANSI standards where possible, although to date only one Linux distribution has been POSIX.1 certified, Linux-FT.Free software projects, although developed in a collaborative fashion, are often produced independently of each other. The fact that the software licenses explicitly permit redistribution, however, provides a basis for larger scale projects that collect the software produced by stand-alone projects and make it available all at once in the form of a Linux distribution.

    +

    A Linux distribution, commonly called a "distro", is a project that manages a remote collection of system software and application software packages available for download and installation through a network connection. This allows the user to adapt the operating system to his/her specific needs. Distributions are maintained by individuals, loose-knit teams, volunteer organizations, and commercial entities. A distribution can be installed using a CD that contains distribution-specific software for initial system installation and configuration. A package manager such as Synaptic or YAST allows later package upgrades and installations. A distribution is responsible for the default configuration of the installed Linux kernel, general system security, and more generally integration of the different software packages into a coherent whole.

    +
    +
    +

    +1.2.3. Community

    +

    A distribution is largely driven by its developer and user communities. Some vendors develop and fund their distributions on a volunteer basis. Examples include Debian and the Debian-based, Ubuntu. Others maintain a community version of their commercial distributions, as Red Hat does with Fedora.

    +

    In many cities and regions, local associations known as Linux Users Groups (LUGs) seek to promote their preferred distribution and by extension free software. They hold meetings and provide free demonstrations, training, technical support, and operating system installation to new users. Many Internet communities also provide support to Linux users and developers. Most distributions and free software / open source projects have IRC (Internet Relay Chat) chatrooms or newsgroups. Online forums are another means for support. Linux distributions host mailing lists; commonly there will be a specific topic such as usage or development for a given list. All these can be found simply by running an appropriate search on Google.

    +

    Although Linux distributions are generally available without charge, several large corporations sell, support, and contribute to the development of the components of the system and of free software. These include Dell, IBM, HP, Oracle, Sun Microsystems, Novell, Nokia. A number of corporations, notably Red Hat, have built their entire business around Linux distributions.

    +
    +
    +

    +1.2.4. Can I make a profit out of running a business involving Linux?

    +

    The answer is, "Yes!". The free software licenses, on which the various software packages of a distribution built on the Linux kernel are based, explicitly accommodate and encourage commercialization; the relationship between a Linux distribution as a whole and individual vendors may be seen as symbiotic. One common business model of commercial suppliers is charging for support, especially for business users. A number of companies also offer a specialized business version of their distribution, which adds proprietary support packages and tools to administer higher numbers of installations or to simplify administrative tasks. Another business model is to give away the software in order to sell hardware. Examples of corporations that are extensively (and sometimes exclusively) open-source and Linux-powered , with successful revenue generation models involving these, are Google, SUN, Mozilla, etc.

    +
    +
    +

    +1.2.5. Programming on Linux

    +

    Most Linux distributions support dozens of programming languages. The most common collection of utilities for building both Linux applications and operating system programs is found within the GNU toolchain, which includes the GNU Compiler Collection (GCC) and the GNU build system. Amongst others, GCC provides compilers for Ada, C, C++, Java, and Fortran. The Linux kernel itself is written to be compiled with GCC. Proprietary compilers for Linux include the Intel C++ Compiler, Sun Studio, and IBM XL C/C++ Compiler.

    +

    Most distributions also include support for PHP, Perl, Ruby, Python and other dynamic languages. Examples of languages that are less common, but still supported, are C# via the Mono project, sponsored by Novell, and Scheme. A number of Java Virtual Machines and development kits run on Linux, including the original Sun Microsystems JVM (HotSpot), and IBM's J2SE RE, as well as many open-source projects like Kaffe.

    +

    The two main frameworks for developing graphical applications are those of GNOME and KDE. These projects are based on the GTK+ and Qt widget toolkits, respectively, which can also be used independently of the larger framework. Both support a wide variety of languages. There are a number of Integrated Development Environments (IDEs) available including Anjuta, Code::Blocks, Eclipse, KDevelop, Lazarus, MonoDevelop, NetBeans, and Omnis Studio while the long-established editors Vim and Emacs remain popular.

    +
    +
    +
    +

    +1.3. Reasons for Using Linux

    +
    • Linux is free:

    +

    As in "free beer". Linux can be downloaded in its entirety from the Internet completely for free. No registration fees, no costs per user, free updates, and freely available source code in case you want to change the behavior of your system. +Most of all, Linux is free as in "free speech": +The license commonly used is the GNU Public License (GPL). The license says that anybody who may want to do so, has the right to change Linux and eventually to redistribute a changed version, on the one condition that the code is still available after redistribution. In practice, you are free to grab a kernel image and sell the new code, as long as your customers can still have a copy of that code.

    +
    • Linux is portable to any hardware platform:

    +

    A vendor, who wants to sell a new type of computer and who does not know what kind of OS his/her new machine will run, can take a Linux kernel and make it work on his/her hardware, because documentation related to this activity is freely available.

    +
    • Linux was made to keep on running:

    +

    As with UNIX, a Linux system expects to run without rebooting all the time. That is why a lot of tasks are being executed at night or scheduled automatically for other times, resulting in higher availability during busier periods and a more balanced use of the hardware. This property allows for Linux to be applicable to environments where people do not have the time or the possibility to control their systems constantly.

    +
    • Linux is secure and versatile:

    +

    The security model used in Linux is based on the UNIX idea of security, which is known to be robust and of proven quality. But Linux is not only safe from attacks from the Internet: it will adapt equally to other situations, utilizing the same high standards for security.

    +
    • Linux is scalable:

    +

    From a Palmtop with 2 MB of memory to a petabyte storage cluster with hundreds of nodes: add or remove the appropriate packages and Linux fits all. One does not need a supercomputer anymore,because you can use Linux to do big things using the building blocks provided with the system. If one wants to do little things, such as making an operating system for an embedded processor or just recycling your old 486, Linux will do that as well.

    +
    • The Linux OS and Linux applications have very short debug−times:

    +

    Because Linux has been developed and tested by thousands of people, both errors and people to fix them are found very quickly. It often happens that there are only a couple of hours between discovery and fixing of a bug.

    +
    +
    +
    +

    +2. Getting Started

    +
    +

    +2.1. Logging in, activating the user interface and logging out

    +

    In order to work on a Linux system directly, one needs to provide a user name and password. You always need to authenticate to the system. Most PC−based Linux systems have two basic modes for a system to run in: either quick and clean in text console mode,which includes with mouse, multitasking and multi−user features, or in graphical console mode, which looks better but eats more system resources.

    +
    +

    +2.1.1. Graphical Mode

    +

    This is the default nowadays on most desktop computers. You know you will be connecting to the system using graphical mode when you are first asked for your user name, and then to type your password.

    +

    To log in, make sure the mouse pointer is in the login window, provide your user name and password to the system and click OK or press Enter. +It is generally considered a bad idea to connect (graphically) using the root user name, the system adminstrator's account, since the use of graphics includes running a lot of extra programs, in root's case with a lot of extra permissions. To keep all risks as low as possible, use a normal user account to connect graphically. But there are enough risks to keep this in mind as a general advice, for all use of the root account: only log in as root when extra privileges are required.

    +

    After entering your user name/password combination, it can take a little while before the graphical environment is started, depending on the CPU speed of your computer, on the software you use and on your personal settings.

    +

    To continue, you will need to open a terminal window or xterm for short (X being the name for the underlying software supporting the graphical environment). This program can be found in the Applications−>Utilities->System Tools or Internet menu, depending on what window manager you are using. There might be icons that you can use as a shortcut to get an xterm window as well, and clicking the right mouse button on the desktop background will usually present you with a menu containing a terminal window application.

    +

    While browsing the menus, you will notice that a lot of things can be done without entering commands via the keyboard. For most users, the good old point−n−click method of dealing with the computer will do. But for those who want to enter the "heart" of the system, a tool stronger than a mouse will be required to handle the various tasks. This tool is the shell, and when in graphical mode, we activate our shell by opening a terminal window.

    +

    A terminal window should always show a command prompt when you open one. This terminal shows a standard prompt, which displays the user's login name, and the current working directory, represented by the twiddle (~)

    +

    Another common form for a prompt is this one: +[

    +
    +

    user@host
    +

    dir]

    +

    In the above example, user will be your login name, hosts the name of the machine you are working on, and dir an indication of your current location in the file system. Prompts can display all kinds of information, but they are not part of the commands you are giving to your system. To disconnect from the system in graphical mode, you need to close all terminal windows and other applications. After that, hit the logout icon or find Log Out in the menu. Closing everything is not really necessary, and the system can do this for you, but session management might put all currently open applications back on your screen when you connect again, which takes longer and is not always the desired effect. However, this behavior is configurable.

    +

    When you see the login screen again, asking to enter user name and password, logout was successful.

    +
    +
    +

    +2.1.2. Text Mode

    +

    One is in text mode when the whole screen is black, showing (in most cases white) characters. A text mode login screen typically shows some information about the machine you are working on, the name of the machine and a prompt waiting for you to log in.

    +

    The login is different from a graphical login, in that you have to hit the Enter key after providing your user name, because there are no buttons on the screen that you can click with the mouse. Then you should type your password, followed by another Enter. You will not see any indication that you are entering something, not even an asterisk, and you won't see the cursor move. But this is normal on Linux and is done for security +reasons.

    +

    When the system has accepted you as a valid user, you may get some more information, called the message of the day, which can be anything. Additionally, it is popular on UNIX systems to display a fortune cookie, which contains some general wise or unwise (this is up to you) thoughts. After that, you will be given a shell, indicated with the same prompt that you would get in graphical mode.

    +

    Also in text mode: log in as root only to do setup and configuration that absolutely requires administrator privileges, such as adding users, installing software packages, and performing network and other system configuration. Once you are finished, immediately leave the special account and resume your work as a non−privileged user.

    +

    Logging out is done by entering the logout command, followed by Enter. You are successfully disconnected from the system when you see the login screen again.Don't power−off the computer after logging out. It is not meant to be shut off without application of the proper procedures for halting the system. Powering it off without going through the halting process might cause severe damage!

    +
    +
    +
    +
    +

    +3. Basic Commands

    +
    +

    +3.1. ls

    +

    When invoked without any arguments, ls lists the files in the current working directory. A directory that is not the current working directory can be specified and ls will list the files there. The user also may specify any list of files and directories. In this case, all files and all contents of specified directories will be listed. The name ls is derived from list segments which was used in earlier systems.

    +

    Files whose names start with "." are not listed, unless the -a flag is specified or the files are specified explicitly.

    +

    Without options, ls displays files in a bare format. This bare format however makes it difficult to establish the type, permissions, and size of the files. The most common options to reveal this information or change the list of files are:

    +<block_quote>
      +
    • -l long format, displaying Unix file types, permissions, number of hard links, owner, group, size, date, and filename

    • +
    • -F appends a character revealing the nature of a file, for example, * for an executable, or / for a directory. Regular files have no suffix.

    • +
    • -a lists all files in the given directory, including those whose names start with "." (which are hidden files in Unix). By default, these files are excluded from the list.

    • +
    • -R recursively lists subdirectories. The command ls -R / would therefore list all files.

    • +
    • -d shows information about a symbolic link or directory, rather than about the link's target or listing the contents of a directory.

    • +
    • -t sort the list of files by modification time.

    • +
    • -h print sizes in human readable format. (e.g., 1K, 234M, 2G, etc.)

    • +
    </block_quote>

    In some environments, providing the option --color (for GNU ls) or -G (FreeBSD ls) causes ls to highlight different types of files with different colors, instead of with characters as -F would. To determine what color to use for a file, GNU ls checks the Unix file type, the file permissions, and the file extension, while FreeBSD ls checks only the Unix file type and file permissions.:

    +
    $ ls
    +jeeves.rst psmith.html blandings.html
    +$ ls -l
    +drwxr--r--   1 plum  editors   4096  jeeves
    +-rw-r--r--   1 plum  editors  30405  psmith
    +-r-xr-xr-x   1 plum  plum      8460  blandings
    +

    Here "$" actually is the beginning of the prompt. This is typical in most Unix-based systems.

    +
    +
    +

    +3.2. date

    +

    The Unix date command displays the time and date. The super-user can use it to set the system clock.

    +

    With no options, the date command displays the current date and time, including the abbreviated day name, abbreviated month name, day of the month, the time separated by colons, the timezone name, and the year. For example:

    +
    $date
    +Tue Sep  8 12:01:45 IST 2009
    +

    On some systems to set the current date and time to September 8, 2004 01:22 you type:

    +
    $date --set="20040908 01:22"
    +

    In order to view the various options for the date command, type:

    +
    $man date
    +

    This will take you to the "Manual" page comprising of all the details on the date command. You can return to the terminal from the "man" page by pressing the Esc key in the keyboard and typing ":q" in that order.

    +
    +
    +

    +3.3. cd

    +

    This stands for "change directory". When one wants to go up to the parent directory, bypassing the tree of directories one has entered, “ cd ..” can be used.

    +

    One dot '.' represents the current directory while two dots '..' represent the parent directory.

    +

    “ cd -” will return you to the previous directory (a bit like an “undo”).

    +

    You can also use cd absolute path or cd relative path (see below):

    +

    Absolute paths:

    +<block_quote>

    An “ absolute path” is easily recognised from the leading forward slash, /. The / means that you start at the top level directory and continue down.

    </block_quote>

    For example to get to /boot/grub you would type:

    +
    $cd /boot/grub
    +

    This is an absolute path because you start at the top of the hierarchy and go downwards from there (it doesn't matter where in the filesystem you were when you typed the command).

    +

    Relative paths:

    +<block_quote>

    A “ relative path” doesn't have a preceding slash. Use a relative path when you start from a directory below the top level directory structure. This is dependent on where you are in the filesystem.

    +

    For example if you are in root's home directory and want to get to /root/music, you type:

    +
    $ cd music
    </block_quote>

    Please note that there is no / using the above cd command. Using a / would cause this to be an absolute path, working from the top of the hierarchy downward.

    +
    +
    +

    +3.4. who

    +

    The standard Unix command who displays a list of users who are currently logged into a computer.

    +

    The who command is related to the command w, which provides the same information but also displays additional data and statistics.:

    +
    $who
    +beeblebrox tty7         2009-09-08 10:50 (:0)
    +beeblebrox pts/0        2009-09-08 11:25 (:0.0)
    +dumbledore pts/1        2009-09-08 18:11 (potter.xyz.in)
    +beeblebrox pts/2        2009-09-08 18:53 (:0.0)
    +

    The command can be invoked with the arguments am i or am I (so it is invoked as who am i or * who am I*), showing information about the current terminal only (see the -m option below, of which this invocation is equivalent).

    +

    In order to find out the various options that can be appended to the who command, check the man page by typing out the following in the terminal:

    +
    $man who
    +

    This will take you to the "Manual" page containing details about the who command

    +
    +
    +

    +3.5. mkdir

    +

    This command is used to make a new directory. Normal usage is as straightforward as follows:

    +
    $mkdir name_of_directory
    +

    Where name_of_directory is the name of the directory one wants to create. When typed as above (ie. normal usage), the new directory would be created within the current directory. On Unix, multiple directories can be specified, and mkdir will try to create all of them.

    +
    +

    +3.5.1. Options

    +

    On Unix-like operating systems, mkdir takes options. Three of the most common options are:

    +<block_quote>
      +
    • -p: will also create all directories leading up to the given directory that do not exist already. If the given directory already exists, ignore the error.

    • +
    • -v: display each directory that mkdir creates. Most often used with -p.

    • +
    • -m: specify the octal permissions of directories created by mkdir.

    • +
    </block_quote>

    -p is most often used when using mkdir to build up complex directory hierarchies, in case a necessary directory is missing or already there. -m is commonly used to lock down temporary directories used by shell scripts.

    +
    +
    +

    +3.5.2. Examples

    +

    An example of -p in action is:

    +
    $mkdir -p /tmp/a/b/c
    +

    If /tmp/a exists but /tmp/a/b does not, mkdir will create /tmp/a/b before creating /tmp/a/b/c.

    +

    And an even more powerful command, creating a full tree at once (this however is a Shell extension, nothing mkdir does itself):

    +
    $mkdir -p tmpdir/{trunk/sources/{includes,docs},branches,tags}
    +

    This will create:

    +<definition_list><definition_list_item><term>tmpdir - branches</term><definition>
      +
    • tag

    • +
    • <definition_list><definition_list_item><term>trunk - sources - includes</term><definition>
      • docs

      </definition>
      </definition_list_item>
      </definition_list>
    • +
    </definition>
    </definition_list_item>
    </definition_list>
    +
    +
    +
    +
    +

    +4. Getting Help

    +
    +

    +4.1. apropos and whatis

    +

    This is a command to search the manual pages files in Unix and Unix-like operating systems.

    +
    $ apropos grep
    +egrep       egrep (1)       Search a file for a pattern using full regular expressions
    +fgrep       fgrep (1)       Search a file for a fixed-character string
    +fmlgrep     fmlgrep (1)     Search a file for a pattern
    +grep        grep (1)        Search a file for a pattern
    +gzgrep      gzgrep (1)      Search a possibly compressed file for a regular expression
    +nisgrep     nismatch (1)    Utilities for searching NIS+ tables
    +pgrep       pgrep (1)       Find or signal a process by name or other attribute
    +zgrep       zgrep (1)       Search a possibly compressed file for a regular expression
    +...
    +

    In this example, the user uses apropos to search for the string "grep", and apropos returns the indicated man pages that include the term "grep".

    +

    A short index of explanations for commands is available using the whatis command, like in the examples below:

    +
    $whatis ls
    +ls (1)           - list directory contents
    +

    This displays short information about a command, and the first section in the collection of man pages that contains an appropriate page.

    +

    If you don't know where to get started and which man page to read, apropos gives more information. Say that you do not know how to start a browser, then you could enter the following command:

    +
    $apropos browser
    +gmusicbrowser (1)    - Jukebox for large collections of audio files
    +infobrowser (1)      - read Info documents
    +libsmbclient (7)     - An extension library for browsers and that               can be used...
    +opera (1)            - a standards-compliant graphical Web browser
    +sensible-browser (1) - sensible editing, paging, and web browsing
    +smbtree (1)          - A text based smb network browser
    +tvtk_doc (1)         - A GUI based TVTK documentation search browser.
    +viewres (1)          - graphical class browser for Xt
    +w3m (1)              - a text based Web browser and pager
    +www-browser (1)      - a text based Web browser and pager
    +...
    +
    +
    +

    +4.2. man

    +

    Man pages (short for "manual pages") are the extensive documentation that comes preinstalled with almost all substantial Unix and Unix-like operating systems. The Unix command used to display them is man. Each page is a self-contained document.

    +

    To read a manual page for a Unix command, one can use:

    +
    $ man <command_name>
    +

    at a shell prompt; for example, "man ftp". In order to simplify navigation through the output, man generally uses the less terminal pager.

    +

    Pages are traditionally referred to using the notation "name(section)"; for example, ftp(1). The same page name may appear in more than one section of the manual, this can occur when the names of system calls, user commands, or macro packages coincide. Two examples are man(1) and man(7), or exit(2) and exit(3). The syntax for accessing the non-default manual section varies between different man implementations. On Linux and BSD, for example, the syntax for reading *printf(3) is:

    +
    $man 3 printf
    +

    Another example:

    +
    $man man
    +

    The previous example will take you to the "Manual" page entry about manual pages!

    +
    +

    +4.2.1. Layout

    +

    All man pages follow a common layout that is optimized for presentation on a simple ASCII text display, possibly without any form of highlighting or font control. Sections present may include:

    +<definition_list><definition_list_item><term>NAME</term><definition>

    The name of the command or function, followed by a one-line description of what it does.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SYNOPSIS</term><definition>

    In the case of a command, you get a formal description of how to run it and what command line options it takes. For program functions, a list of the parameters the function takes and which header file contains its definition. For experienced users, this may be all the documentation they need.

    </definition>
    </definition_list_item>
    <definition_list_item><term>DESCRIPTION</term><definition>

    A textual description of the functioning of the command or function.

    </definition>
    </definition_list_item>
    <definition_list_item><term>EXAMPLES</term><definition>

    Some examples of common usage.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SEE ALSO</term><definition>

    A list of related commands or functions.

    </definition>
    </definition_list_item>
    </definition_list>

    Other sections may be present, but these are not well standardized across man pages. Common examples include: OPTIONS, EXIT STATUS, ENVIRONMENT, KNOWN BUGS, FILES, AUTHOR, REPORTING BUGS, HISTORY and COPYRIGHT.

    +

    These days virtually every Unix command line application comes with its man page, and many Unix users perceive a lack of man pages as a sign of low quality; indeed, some projects, such as Debian, go out of their way to write man pages for programs lacking one. Few alternatives to man have enjoyed much popularity, with the possible exception of the GNU project's "info" system, an early and simple hypertext system.

    +

    However, the format of a single page for each application, the lack of classification within the sections and the relatively unsophisticated formatting facilities have motivated the development of alternative documentation systems, such as the previously mentioned "info" system.

    +

    Most Unix GUI applications (particularly those built using the GNOME and KDE development environments) now provide end-user documentation in HTML and include embedded HTML viewers such as yelp for reading the help within the application.

    +

    Usually the man pages are written in English. Translations into other languages can be also available on the system.

    +

    The default format of the man pages is troff, with either the macro package man (appearance oriented) or on some systems mdoc (semantic oriented). This makes it possible to typeset a man page to PostScript, PDF and various other formats for viewing or printing.

    +
    +
    +
    +

    +4.3. info

    +

    info is a software utility which forms a hypertextual, multipage documentation and help viewer working on a command line interface, useful when there is no GUI available.

    +

    The syntax is

    +
    $ info <command_name>
    +

    info processes info files, which are Texinfo formatted files, and presents the documentation as a tree, with simple commands to traverse the tree and to follow cross references. For instance

    +<block_quote>
      +
    • n goes to the next page.

    • +
    • p goes to the previous page.

    • +
    • u goes to the upper page.

    • +
    • l goes to the last(visited) node

    • +
    • To follow a cross reference, the cursor can be moved over a link (a word preceded by a <title_reference>*</title_reference>) and enter pressed.

    • +
    </block_quote>

    info was initially written for use with GNU/Linux and then ported to other Unix-like operating systems.

    +
    +
    +

    +4.4. --help

    +

    Most GNU commands support the --help, which gives a short explanation about how to use the command and a list of available options. Below is the output of this option with the cat command:

    +
    $ userprompt@host: cat --help
    +Usage: cat [OPTION] [FILE]...
    +Concatenate FILE(s), or standard input, to standard output.
    +
    +  -A, --show-all           equivalent to -vET
    +  -b, --number-nonblank    number nonempty output lines
    +  -e                       equivalent to -vE
    +  -E, --show-ends          display $ at end of each line
    +  -n, --number             number all output lines
    +  -s, --squeeze-blank      suppress repeated empty output lines
    +  -t                       equivalent to -vT
    +  -T, --show-tabs          display TAB characters as ^I
    +  -u                       (ignored)
    +  -v, --show-nonprinting   use ^ and M- notation, except for LFD and              TAB
    +  --help     display this help and exit
    +  --version  output version information and exit
    +
    +With no FILE, or when FILE is -, read standard input.
    +
    +Examples:
    +  cat f - g  Output f's contents, then standard input, then g's           contents.
    +  cat        Copy standard input to standard output.
    +
    +Report bugs to <bug-coreutils@gnu.org>.
    +
    +
    +
    +

    +5. Basic file handling

    +
    +

    +5.1. cp

    +

    cp is the command entered in a Unix shell to copy a file from one place to another, possibly on a different filesystem. The original file remains unchanged, and the new file may have the same or a different name.

    +
    +

    +5.1.1. Usage

    +

    To copy a file to another file:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ][ -- ] SourceFile TargetFile
    +

    To copy a file to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -r | -R ] [ -- ] SourceFile ...              TargetDirectory
    +

    To copy a directory to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -- ] { -r | -R }
    +SourceDirectory ... TargetDirectory
    +
    +
    +

    +5.1.2. Flags

    +

    -f (force) – specifies removal of the target file if it cannot be opened for write operations. The removal precedes any copying performed by the cp command.

    +

    -P – makes the cp command copy symbolic links. The default is to follow symbolic links, that is, to copy files to which symbolic links point.

    +

    -i (interactive) – prompts you with the name of a file to be overwritten. This occurs if the TargetDirectory or TargetFile parameter contains a file with the same name as a file specified in the SourceFile or SourceDirectory parameter. If you enter y or the locale's equivalent of y, the cp command continues. Any other answer prevents the cp command from overwriting the file.

    +

    -p (preserve) – duplicates the following characteristics of each SourceFile/SourceDirectory in the corresponding TargetFile and/or TargetDirectory:

    +<block_quote>
      +
    • The time of the last data modification and the time of the last access.

    • +
    • The user ID and group ID (only if it has permissions to do this)

    • +
    • The file permission bits and the SUID and SGID bits.

    • +
    </block_quote>

    -R (recursive) – copy directories (recursively copying all the contents)

    +
    +
    +

    +5.1.3. Examples

    +

    To make a copy of a file in the current directory, enter:

    +
    $ cp prog.c prog.bak
    +

    This copies prog.c to prog.bak. If the prog.bak file does not already exist, the cp command creates it. If it does exist, the cp command replaces it with a copy of the prog.c file.

    +

    To copy a file in your current directory into another directory, enter:

    +
    $ cp zaphod /home/books/hhgg
    +

    This copies the jones file to /home/books/hhgg/zaphod.

    +

    To copy a file to a new file and preserve the modification date, time, and access control list associated with the source file, enter:

    +
    $ cp -p martin_luther_king martin_luther_king.jr
    +

    This copies the martin_luther_king file to the martin_luther_king.jr file. Instead of creating the file with the current date and time stamp, the system gives the martin_luther_king.jr file the same date and time as the martin_luther_king file. The martin_luther_king.jr file also inherits the martin_luther_king file's access control protection.

    +

    To copy all the files in a directory to a new directory, enter:

    +
    $ cp /home/galactica/clients/* /home/hhgg/customers
    +

    This copies only the files in the clients directory to the customers directory.

    +

    To copy a directory, including all its files and subdirectories, to another directory, enter:

    +<block_quote>

    $ cp -R /home/hhgg/clients /home/hhgg/customers

    </block_quote>

    This copies the clients directory, including all its files, subdirectories, and the files in those subdirectories, to the customers/clients directory.

    +

    To copy a specific set of files of any extension to another directory, enter:

    +
    $ cp zaphod arthur ford /home/hhgg/clients
    +

    This copies the zaphod, arthur, and ford files in your current working directory to the /home/hhgg/clients directory.

    +

    To use pattern-matching characters to copy files, enter:

    +
    $ cp programs/*.py .
    +

    This copies the files in the programs directory that end with .py to the current directory, signified by the single "." (dot). You must type a space between the py and the final dot.

    +
    +
    +
    +

    +5.2. mv

    +

    mv (short for move) is a Unix command that moves one or more files or directories from one place to another. The original file is deleted, and the new file may have the same or a different name. If possible (i.e. when the original and new files are on the same file system), mv will rename the file instead. Write permission is required on all directories being modified.

    +
    +

    +5.2.1. Conflicting existing file

    +

    In all cases, when a file is moved to have the name of an existing file (in the same directory), the existing file is deleted. If the existing file is not writable but is in a directory that is writable, then the mv command asks for confirmation if possible (i.e. if run from a terminal) before proceeding, unless the -f (force) option is used.

    +
    +
    +

    +5.2.2. Differences with copy and delete

    +

    Note that, usually, when moving files within the same volume, moving (and/or renaming) is not the same as simply copying and then deleting the original. When moving a file, the link is simply removed from the old parent directory and added to the new parent directory. However, the file itself is untouched (i.e. it has the same inodes and resides at the same place on the disk). For example, you cannot copy a file you cannot read, but you can move (and/or rename) it (provided you have write permission to its old and new parent directories). Also, suppose there is a non-empty directory you do not have write permission to. You cannot delete this directory (since you cannot delete its contents); but you can move (and/or rename) it. Also, since moving between filenames on a single volume does not involve copying, it is faster and does not place strain of lots of reads and writes on the disk. Moving files across different volumes, however, does necessitate copying and deleting.

    +
    +
    +

    +5.2.3. Examples

    +
    $ mv myfile mynewfilename    renames a file
    +$ mv myfile otherfilename    renames a file and deletes the existing            file "myfile"
    +$ mv myfile /myfile          moves 'myfile' from the current            directory to the root directory
    +$ mv myfile dir/myfile       moves 'myfile' to 'dir/myfile' relative            to the current directory
    +$ mv myfile dir              same as the previous command (the          filename is implied to be the same)
    +$ mv myfile dir/myfile2      moves 'myfile' to dir and renames it to            'myfile2'
    +$ mv foo bar baz dir         moves multiple files to directory dir
    +$ mv --help                  shows a very concise help about the                syntax of the command
    +$ man mv                     prints an extensive user manual for                'mv' in the terminal
    +

    In all cases, the file or files being moved or renamed can be a directory.

    +

    Note that when the command is called with two arguments (as mv name1 name2 or mv name1 /dir/name2), it can have three different effects, depending on whether name2 does not exist, is an existing file, or is an existing directory. If the user intends to refer to an existing directory, /. (or in some Unix versions / is sufficient) may be appended to the name to force the system to check this. To move a file to a new directory, the directory must be created first.

    +
    +
    +
    +

    +5.3. rm

    +

    rm (short for "remove") is one of several basic Unix command lines that operates on files. It is used to delete files from a filesystem. The data is not actually destroyed. Only the index listing where the file is stored is destroyed, and the storage is made available for reuse. There are undelete utilities that will attempt to reconstruct the index and can bring the file back if the parts were not reused.

    +

    Here's example to remove a file named "foo" from a directory, here shown with the -i option:

    +
    $ rm -i foo
    +remove foo? y
    +
    +

    +5.3.1. Options

    +

    Common options that rm accepts include:

    +<block_quote>
      +
    • -r, which removes directories, removing the contents recursively beforehand (so as not to leave files without a directory to reside in) ("recursive")

    • +
    • -i, which asks for every deletion to be confirmed ("interactive")

    • +
    • -f, which ignores non-existent files and overrides any confirmation prompts ("force")

    • +
    • -v, which shows what is being removed as it happens ("verbose")

    • +
    </block_quote>

    rm is often aliased to "rm -i" so as to avoid accidental deletion of files. If a user still wishes to delete a large number of files without confirmation, they can manually cancel out the -i argument by adding the -f option (as the option specified later on the expanded command line "rm -i -f" takes precedence).

    +

    rm -rf (variously, rm -rf /, rm -rf <title_reference>*</title_reference>, and others) is frequently used in jokes and anecdotes about Unix disasters. The rm -rf variant of the command, if run by a superuser on the root directory, would cause the contents of every writable mounted filesystem on the computer to be deleted.

    +

    rm is often used in conjunction with xargs to supply a list of files to delete:

    +
    xargs rm < filelist
    +

    When rm is used on a symbolic link, it deletes the link, but does not affect the target of the link.

    +
    +
    +

    +5.3.2. Permissions

    +

    Usually, on most filesystems, deleting a file requires write permission on the parent directory (and execute permission, in order to enter the directory in the first place). (Note that, confusingly for beginners, permissions on the file itself are irrelevant. However, GNU rm asks for confirmation if a write-protected file is to be deleted, unless the -f option is used.)

    +

    To delete a directory (with rm -r), one must delete all of its contents recursively. This requires that one must have read and write and execute permission to that directory (if it's not empty) and all non-empty subdirectories recursively (if there are any). The read permissions are needed to list the contents of the directory in order to delete them. This sometimes leads to an odd situation where a non-empty directory cannot be deleted because one doesn't have write permission to it and so cannot delete its contents; but if the same directory were empty, one would be able to delete it.

    +

    If a file resides in a directory with the sticky bit set, then deleting the file requires one to be the owner of the file.

    +
    +
    +
    +
    +

    +6. Command Line Arguments

    +

    In computer command line interfaces, a command line argument is an argument sent to a program being called. In general, a program can take any number of command line arguments, which may be necessary for the program to run, or may even be ignored, depending on the function of that program.

    +

    For example, in Unix and Unix-like environments, an example of a command-line argument is:

    +
    rm file.s
    +

    "file.s" is a command line argument which tells the program rm to remove the file "file.s".

    +

    Programming languages such as C, C++ and Java allow a program to interpret the command line arguments by handling them as string parameters in the main function.

    +

    A command line option or simply option (also known as a command line parameter, flag, or a switch) is an indication by a user that a computer program should change its default output.

    +

    Long options are introduced via "--", and are typically whole words. For example, ls --long --classify --all. Arguments to long options are provided with "=", as ls --block-size=1024. Some Unix programs use long options with single dashes, for example MPlayer as in mplayer -nosound.

    +

    Linux also uses "--" to terminate option lists. For example, an attempt to delete a file called -file1 by using rm -file1 may produce an error, since rm may interpret -file1 as a command line switch. Using rm -- -file1 removes ambiguity.

    +
    +
    +

    +7. Basic Text Processing

    +
    +

    +7.1. head

    +

    head is a program on Unix and Unix-like systems used to display the first few lines of a text file or piped data. The command syntax is:

    +
    $ head [options] <file_name>
    +

    By default, head will print the first 10 lines of its input to the standard output. The number of lines printed may be changed with a command line option. The following example shows the first 20 lines of filename:

    +
    $ head -n 20 filename
    +

    This displays the first 5 lines of all files starting with foo:

    +
    $ head -n 5 foo*
    +

    Some versions omit the n and just let you say -5.

    +
    +

    +7.1.1. Flags

    +
    -c <x number of bytes> Copy first x number of bytes.
    +

    Other options: sed

    +

    Many early versions of Unix did not have this command, and so documentation and books had sed do this job:

    +
    sed 5q foo
    +

    This says to print every line (implicit), and quit after the fifth.

    +
    +
    +
    +

    +7.2. tail

    +

    tail is a program on Unix and Unix-like systems used to display the last few lines of a text file or piped data.

    +

    The command-syntax is:

    +
    $ tail [options] <file_name>
    +

    By default, tail will print the last 10 lines of its input to the standard output. With command line options the number of lines printed and the printing units (lines, blocks or bytes) may be changed. The following example shows the last 20 lines of filename:

    +
    $ tail -n 20 filename
    +

    This example shows the last 15 bytes of all files starting with foo:

    +
    $ tail -c 15 foo*
    +

    This example shows all lines of filename from the second line onwards:

    +
    $ tail -n +2 filename
    +

    Using an older syntax (still used in Sun Solaris as the -n option is not supported), the last 20 lines and the last 50 bytes of filename can be shown with the following command:

    +
    $ tail -20 filename
    +$ tail -50c filename
    +

    However this syntax is now obsolete and does not conform with the POSIX 1003.1-2001 standard. Even if still supported in current versions, when used with other options (like -f, see below), these switches could not work at all.

    +
    +

    +7.2.1. File monitoring

    +

    tail has a special command line option -f (follow) that allows a file to be monitored. Instead of displaying the last few lines and exiting, tail displays the lines and then monitors the file. As new lines are added to the file by another process, tail updates the display. This is particularly useful for monitoring log files. The following command will display the last 10 lines of messages and append new lines to the display as new lines are added to messages:

    +
    $ tail -f /var/adm/messages
    +

    To interrupt tail while it is monitoring, break-in with Ctrl+C. This command can be run "in the background" with &, see job control.

    +

    If you have a command's result to monitor, you can use the watch command.

    +
    +
    +
    +

    +7.3. cut

    +

    In computing, cut is a Unix command line utility which is used to extract sections from each line of input — usually from a file.

    +

    Extraction of line segments can typically be done by bytes (-b), characters (-c), or fields (-f) separated by a delimiter (-d — the tab character by default). A range must be provided in each case which consists of one of N, N-M, N- (N to the end of the line), or -M (beginning of the line to M), where N and M are counted from 1 (there is no zeroth value). Since version 6, an error is thrown if you include a zeroth value. Prior to this the value was ignored and assumed to be 1.

    +

    Assuming a file named file containing the lines:

    +
    foo:bar:baz:qux:quux
    +one:two:three:four:five:six:seven
    +alpha:beta:gamma:delta:epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +

    To output the fourth through tenth characters of each line:

    +
    $ cut -c 4-10 file
    +

    This gives the output:

    +
    :bar:ba
    +:two:th
    +ha:beta
    +

    To output the fifth field through the end of the line of each line using the colon character as the field delimiter:

    +
    $ cut -d : -f 5- file
    +

    This gives the output:

    +
    quux
    +five:six:seven
    +epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +
    +
    +

    +7.4. paste

    +

    paste is a Unix command line utility which is used to join files horizontally (parallel merging) by outputting lines consisting of the sequentially corresponding lines of each file specified, separated by tabs, to the standard output. It is effectively the horizontal equivalent to the utility cat command which operates on the vertical plane of two or more files.

    +

    To paste several columns of data together into the file www from files who, where, and when:

    +
    $ paste who where when > www
    +

    If the files contain:

    +++++
    +

    This creates the file named www containing:

    +
    Batman            GothamCity       January 3
    +Trillian          Andromeda        February 4
    +Jeeves            London           March 19
    +
    +
    +
    +

    +8. Shell Meta Characters

    +

    Unix recognizes certain special characters, called "meta characters," as command directives. The shell meta characters are recognized anywhere they appear in the command line, even if they are not surrounded by blank space. For that reason, it is safest to only use the characters A-Z, a-z, 0-9, and the period, dash, and underscore characters when naming files and directories on Unix. If your file or directory has a shell meta character in the name, you will find it difficult to use the name in a shell command.

    +

    The shell meta characters include:

    +

    / < > ! $ % ^ & * | { } [ ] " ' ` ~ ;

    +

    Different shells may differ in the meta characters recognized.

    +

    As an example,

    +
    $ ls file.*
    +

    run on a directory containing the files file, file.c, file.lst, and myfile would list the files file.c and file.lst. However,:

    +
    $ ls file.?
    +

    run on the same directory would only list file.c because the ? only matches one character, no more, no less. This can save you a great deal of typing time. For example, if there is a file called california_cornish_hens_with_wild_rice and no other files whose names begin with 'c', you could view the file without typing the whole name by typing this:

    +
    $ more c*
    +

    because the c* matches that long file name.

    +

    Filenames containing metacharacters can pose many problems and should never be intentionally created. If you do find that you've created a file with metacharacters, and you would like to remove it, you have three options. You may use wildcards to match metacharacter, use the to directly enter the filename, or put the command in double quotes (except in the case of double quotes within the file name, these must be captured with one of the first two methods). For example, deleting a file named <title_reference>"``*`|more</title_reference>"` can be accomplished with:

    +
    $ rm ??more
    +

    or:

    +
    $ rm $\backslash$*$\backslash$|more
    +

    or:

    +
    $ rm ''*|more''
    +
    +
    +

    +9. Looking At Files

    +
    +

    +9.1. cat

    +

    The cat command is a standard Unix program used to concatenate and display files. The name is from "catenate", a synonym of concatenate.

    +

    The Single Unix Specification specifies the behavior that the contents of each of the files given in sequence as arguments will be written to the standard output in the same sequence, and mandates one option, -u, where each byte is printed as it is read.

    +

    If the filename is specified as -, then cat will read from standard input at that point in the sequence. If no files are specified, cat will read from standard input entered.

    +
    +

    +9.1.1. Jargon File Definition

    +

    The Jargon File version 4.4.7 lists this as the definition of cat:

    +
    1. To spew an entire file to the screen or some other output sink without
    +     pause (syn. blast).
    +
    +2. By extension, to dump large amounts of data at an unprepared target or
    +     with no intention of browsing it carefully. Usage: considered silly.
    +     Rare outside Unix sites. See also dd, BLT.
    +
    +     Among Unix fans, *cat(1)* is considered an excellent example of
    +     user-interface design, because it delivers the file contents without
    +     such verbosity as spacing or headers between the files, and because
    +     it does not require the files to consist of lines of text, but works
    +     with any sort of data.
    +
    +     Among Unix critics, *cat(1)* is considered the canonical example of
    +     bad user-interface design, because of its woefully unobvious name.
    +     It is far more often used to blast a single file to standard output
    +     than to concatenate two or more files. The name cat for the former
    +     operation is just as unintuitive as, say, LISP's cdr.
    +
    +     Of such oppositions are holy wars made...
    +
    +
    +

    +9.1.2. Useless Use of 'cat'

    +

    UUOC (from comp.unix.shell on Usenet) stands for “Useless Use of cat”. As it is observed on comp.unix.shell, “The purpose of cat is to concatenate (or 'catenate') files. If it's only one file, concatenating it with nothing at all is a waste of time, and costs you a process.”

    +

    Nevertheless one sees people doing:

    +
    $ cat file | some_command and its args ...
    +

    instead of the equivalent and cheaper:

    +
    <file some_command and its args ...
    +

    or (equivalently and more classically):

    +
    some_command and its args ... <file
    +

    Since 1995, occasional awards for UUOC have been given out. The activity of fixing instances of UUOC is sometimes called 'demoggification'.

    +

    Amongst many, it is still considered safer to use cat for such cases given that the < and > keys are next to each other in many popular keyboard mappings. While the risk might be low, the impact of using > instead of < can be high and prohibitive.

    +
    +
    +

    +9.1.3. zcat

    +

    zcat is a Unix program similar to cat, that decompresses individual files and concatenates them to standard output. Traditionally zcat operated on files compressed by compress but today it is usually able to operate on gzip or even bzip2 archives. On such systems, it is equivalent to gunzip -c

    +
    +
    +
    +

    +9.2. more

    +

    In computing, more is a command to view (but not modify) the contents of a text file one screen at a time (terminal pager). It is available on Unix and Unix-like systems, DOS, OS/2 and Microsoft Windows. Programs of this sort are called pagers.

    +
    +

    +9.2.1. Usage

    +

    The command-syntax is:

    +
    $ more [options] [file_name]
    +

    If no file name is provided, more looks for input from stdin.

    +

    Once more has obtained input, it displays as much as can fit on the current screen and waits for user input to advance, with the exception that a form feed (^L) will also cause more to wait at that line, regardless of the amount of text on the screen. In the lower-left corner of the screen is displayed the text "--More--" and a percentage, representing the percent of the file that more has paged through. (This percentage includes the text displayed on the current screen.) When more reaches the end of a file (100%) it exits. The most common methods of navigating through a file are Enter, which advances the output by one line, and Space, which advances the output by one screen.

    +

    There are also other commands that can be used while navigating through the document; consult more's man page for more details.

    +

    Options are typically entered before the file name, but can also be entered in the environment variable $MORE. Options entered in the actual command line will override those entered in the $MORE environment variable. Available options may vary between Unix systems.

    +
    +
    +
    +

    +9.3. less

    +

    less is a terminal pager program on Unix, Windows and Unix-like systems used to view (but not change) the contents of a text file one screen at a time. It is similar to more, but has the extended capability of allowing both forward and backward navigation through the file. Unlike most Unix text editors/viewers, less does not need to read the entire file before starting, resulting in faster load times with large files.

    +
    +

    +9.3.1. Usage

    +

    less can be invoked with options to change its behaviour, for example, the number of lines to display on the screen. A few options vary depending on the operating system. While less is displaying the file, various commands can be used to navigate through the file. These commands are based on those used by both more and vi. It is also possible to search for character patterns in the file.

    +

    By default, less displays the contents of the file to the standard output (one screen at a time). If the file name argument is omitted, it displays the contents from standard input (usually the output of another command through a pipe). If the output is redirected to anything other than a terminal, for example a pipe to another command, less behaves like cat.

    +

    The command-syntax is:

    +
    $ less [options] file_name
    +
    +
    +

    +9.3.2. Frequently Used Options

    +<block_quote>
      +
    • -g: Highlights just the current match of any searched string.

    • +
    • -I: Case-insensitive searches.

    • +
    • -M: Shows more detailed prompt, including file position.

    • +
    • -N: Shows line numbers (useful for source code viewing).

    • +
    • -S: Disables line wrap ("chop long lines"). Long lines can be seen by side scrolling.

    • +
    • -?: Shows help.

    • +
    </block_quote>
    +
    +
    +

    +9.3.3. Frequently Used Commands

    +<block_quote>
      +
    • [Arrows]/[Page Up]/[Page Down]/[Home]/[End]: Navigation.

    • +
    • [Space bar]: Next page.

    • +
    • b: Previous page.

    • +
    • ng: Jump to line number n. Default is the start of the file.

    • +
    • nG: Jump to line number n. Default is the end of the file.

    • +
    • /pattern: Search for pattern. Regular expressions can be used.

    • +
    • '^ or g: Go to start of file.

    • +
    • '$ or G: Go to end of file.

    • +
    • s: Save current content (got from another program like grep) in a file.

    • +
    • =: File information.

    • +
    • h: Help.

    • +
    • q: Quit.

    • +
    </block_quote>
    +
    +
    +

    +9.3.4. Examples

    +
    $ less -M readme.txt                     #Read "readme.txt."
    +$ less +F /var/log/mail.log              #Follow mode for log
    +$ file * | less                          #Easier file analysis.
    +$ grep -i void *.c | less -I -p void     #Case insensitive search                                                         for "void" in all .c files
    +
    +
    +
    +
    +

    +10. Directory Structure

    +

    In the File Hierarchy Standard (FHS) all files and directories appear under the root directory "/", even if they are stored on different physical devices. Note however that some of these directories may or may not be present on a Unix system depending on whether certain subsystems, such as the X Window System, are installed.

    +

    The majority of these directories exist in all UNIX operating systems and are generally used in much the same way; however, the descriptions here are those used specifically for the FHS, and are not considered authoritative for platforms other than Linux.

    ++++
    +
    +

    +10.1. man hier

    +

    This is the manual page on the UNIX filesystem. The syntax for this is:

    +
    $ man hier
    +
    +
    +

    +10.2. ls -l

    +

    Shows you huge amounts of information (permissions, owners, size, and when last modified) for folders and files. The syntax is

    +
    $ ls -l
    +

    This can be done after entering the required directory.

    +
    +
    +
    +

    +11. Permissions and Ownership

    +
    +

    +11.1. chmod

    +

    The chmod command (abbreviated from 'change mode') is a shell command and C language function in Unix and Unix-like environments. When executed, it can change file system modes of files and directories. The modes include permissions and special modes.A chmod command first appeared in AT&T Unix version 1, and is still used today on Unix-like machines.

    +
    +

    +11.1.1. Usage

    +

    The chmod command options are specified like this:

    +
    $ chmod [options] mode[,mode] file1 [file2 ...]
    +

    To view what the permissions currently are, type:

    +
    $ ls -l file
    +
    +
    +

    +11.1.2. Command line options

    +

    The chmod command has a number of command line options that affect its behavior. The most common options are:

    +<block_quote>
      +
    • -R: Changes the modes of directories and files recursively

    • +
    • -v: Verbose mode; lists all files as they are being processed

    • +
    </block_quote>
    +
    +11.1.2.1. Symbolic modes
    +

    To the chmod utility, all permissions and special modes are represented by its mode parameter. One way to adjust the mode of files or directories is to specify a symbolic mode. The symbolic mode is composed of three components, which are combined to form a single string of text:

    +
    $ chmod [references][operator][modes] file1 ...
    +

    The references (or classes) are used to distinguish the users to whom the permissions apply. If no references are specified it defaults to “all” but modifies only the permissions allowed by the umask. The references are represented by one or more of the following letters:

    +++++
    +

    The chmod program uses an operator to specify how the modes of a file should be adjusted. The following operators are accepted:

    ++++
    +

    The modes indicate which permissions are to be granted or taken away from the specified classes. There are three basic modes which correspond to the basic permissions:

    +++++
    +

    The combination of these three components produces a string that is understood by the chmod command. Multiple changes can be specified by separating multiple symbolic modes with commas.

    +
    +
    +
    +11.1.2.2. Symbolic examples
    +

    Add the 'read' and 'write' permissions to the 'user' and 'group' classes of a directory:

    +
    $ chmod ug+rw mydir
    +$ ls -ld mydir
    +drw-rw----   2 starwars  yoda  96 Dec 8 12:53 mydir
    +

    For a file, remove write permissions for all classes:

    +
    $ chmod a-w myfile
    +$ ls -l myfile
    +-r-xr-xr-x   2 starwars  yoda 96 Dec 8 12:53 myfile
    +

    Set the permissions for the u*ser and the *g*roup to read and execute only (no write permission) on *mydir.

    +
    $ chmod ug=rx mydir
    +$ ls -ld mydir
    +dr-xr-x---   2 starwars  yoda 96 Dec 8 12:53 mydir
    +
    +
    +
    +11.1.2.3. Octal numbers
    +

    The chmod command also accepts three and four-digit octal numbers representing modes. Using a three-digit octal number to set the modes of a file named myfile :

    +
    $ chmod 664 myfile
    +$ ls -l myfile
    +-rw-rw-r--  1   57 Jul  3 10:13  myfile
    +

    Since the setuid, setgid and sticky bits are not set, this is equivalent to:

    +
    $ chmod 0664 myfile
    +
    +
    +
    +11.1.2.4. Special modes
    +

    The chmod command is also capable of changing the additional permissions or special modes of a file or directory. The symbolic modes use s to represent the setuid and setgid modes, and t to represent the sticky mode. The modes are only applied to the appropriate classes, regardless of whether or not other classes are specified.

    +

    Most operating systems support the specification of special modes using octal modes, but some do not. On these systems, only the symbolic modes can be used.

    +
    +
    +
    +
    +
    +

    +12. Redirection and Piping

    +

    In computing, redirection is a function common to most command-line interpreters, including the various Unix shells that can redirect standard streams to user-specified locations.

    +

    Programs do redirection with the dup2(2) system call, or its less-flexible but higher-level stdio analogues, freopen(3) and popen(3).

    +
    +

    +12.1. Redirecting standard input and standard output

    +

    Redirection is usually implemented by placing certain characters between commands. Typically, the syntax of these characters is as follows:

    +
    $ command1 > file1
    +

    executes command1, placing the output in file1. Note that this will truncate any existing data in file1. To append output to the end of the file, use the >> operator.:

    +
    $ command1 < file1
    +

    executes command1, using file1 as the source of input (as opposed to the keyboard).:

    +
    $ command1 < infile > outfile
    +

    combines the two capabilities: command1 reads from infile and writes to outfile

    +
    +
    +

    +12.2. Piping

    +

    Programs can be run together such that one program reads the output from another with no need for an explicit intermediate file: +A pipeline of three programs run on a text terminal:

    +
    $ command1 | command2
    +

    executes command1, using its output as the input for command2 (commonly called piping, since the "|" character is known as a "pipe").

    +

    This is equivalent to using two redirects and a temporary file:

    +
    $ command1 > tempfile
    +$ command2 < tempfile
    +$ rm tempfile
    +

    A good example for command piping is combining echo with another command to achieve something interactive in a non-interactive shell, e.g.:

    +
    $ echo -e "user\npass" | ftp localhost
    +

    This runs the ftp client with input user, press return, then pass.

    +
    +
    +

    +12.3. Redirecting to and from the standard file handles

    +

    In Unix shells derived from the original Bourne shell, the first two actions can be further modified by placing a number (the file descriptor) immediately before the character; this will affect which stream is used for the redirection. The Unix standard I/O streams are:

    +++++
    +

    For example:

    +
    $ command1 2> file1
    +

    executes command1, directing the standard error stream to file1.

    +

    In shells derived from csh (the C shell), the syntax instead appends the & character to the redirect characters, thus achieving a similar result.

    +

    Another useful capability is to redirect one standard file handle to another. The most popular variation is to merge standard error into standard output so error messages can be processed together with (or alternately to) the usual output. Example:

    +
    $ find / -name .profile > results 2>&1
    +

    will try to find all files named .profile. Executed without redirection, it will output hits to stdout and errors (e.g. for lack of privilege to traverse protected directories) to stderr. If standard output is directed to file results, error messages appear on the console. To see both hits and error messages in file results, merge stderr (handle 2) into stdout (handle 1) using 2>&1 .

    +

    It's possible use 2>&1 before ">" but it doesn't work. In fact, when the interpreter reads 2>&1, it doesn't know yet where standard output is redirected and then standard error isn't merged.

    +

    If the merged output is to be piped into another program, the file merge sequence 2>&1 must precede the pipe symbol, thus:

    +
    $ find / -name .profile 2>&1 | less
    +

    A simplified form of the command:

    +
    $ command > file 2>&1
    +

    is:

    +
    $ command &>file
    +

    or:

    +
    $command >&file
    +
    +
    +

    +12.4. Chained pipelines

    +

    The redirection and piping tokens can be chained together to create complex commands. For example:

    +
    $ ls | grep '\.sh' | sort > shlist
    +

    lists the contents of the current directory, where this output is filtered to only contain lines which contain .sh, sort this resultant output lexicographically, and place the final output in shlist. This type of construction is used very commonly in shell scripts and batch files.

    +
    +
    +

    +12.5. Redirect to multiple outputs

    +

    The standard command tee can redirect output from a command to several destinations.

    +
    $ ls -lrt | tee xyz
    +

    This directs the file list output to both standard output as well as to the file xyz.

    +
    +
    +
    +

    +13. More Text Processing

    +
    +

    +13.1. grep

    +

    grep is a command line text search utility originally written for Unix. The name is taken from the first letters in global / regular expression / print, a series of instructions for the ed text editor. The grep command searches files or standard input globally for lines matching a given regular expression, and prints them to the program's standard output.

    +
    +

    +13.1.1. Usage

    +

    This is an example of a common grep usage:

    +
    $ grep apple fruitlist.txt
    +

    In this case, grep prints all lines containing 'apple' from the file fruitlist.txt, regardless of word boundaries; therefore lines containing 'pineapple' or 'apples' are also printed. The grep command is case sensitive by default, so this example's output does not include lines containing 'Apple' (with a capital A) unless they also contain 'apple'.

    +

    Like most Unix commands, grep accepts command line arguments to change this and many other behaviors. For example:

    +
    $ grep -i apple fruitlist.txt
    +

    This prints all lines containing 'apple' regardless of capitalization. The '-i' argument tells grep to be case insensitive, or to ignore case.

    +

    To print all lines containing 'apple' as a word ('pineapple' and 'apples' will not match):

    +
    $ grep -w apple fruitlist.txt
    +

    Regular expressions can be used to match more complicated queries.

    +
    +
    +13.1.1.1. Variations
    +

    There are countless implementations and derivatives of grep available for many operating systems. Early variants of grep included egrep and fgrep. The former applies an extended regular expression syntax that was added to Unix after Ken Thompson's original regular expression implementation. The latter searches for any of a list of 'fixed' strings using the Aho-Corasick algorithm. These variants are embodied in most modern grep implementations as command-line switches (and standardized as -E and -F in POSIX). In such combined implementations, grep may also behave differently depending on the name by which it is invoked, allowing fgrep, egrep, and grep to be links to the same program.

    +

    pcregrep is an implementation of grep that uses Perl regular expression syntax.

    +

    Other commands contain the word 'grep' to indicate that they search (usually for regular expression matches). The pgrep utility, for instance, displays the processes whose names match a given regular expression.

    +
    +
    +
    +
    +

    +13.2. tr

    +

    tr (abbreviated from translate or transliterate) is a command in Unix-like operating systems.

    +

    When executed, the program reads from the standard input and writes to the standard output. It takes as parameters two sets of characters, and replaces occurrences of the characters in the first set with the corresponding elements from the other set. For example,

    +
    $ tr 'abcd' 'jkmn'
    +

    maps 'a' to 'j', 'b' to 'k', 'c' to 'm', and 'd' to 'n'.

    +

    Sets of characters may be abbreviated by using character ranges. The previous example could be written:

    +
    $ tr 'a-d' 'jkmn'
    +

    In POSIX compliant versions of tr the set represented by a character range depends on the locale's collating order, so it is safer to avoid character ranges in scripts that might be executed in a locale different from that in which they were written. Ranges can often be replaced with POSIX character sets such as [:alpha:].

    +

    The -c flag complements the first set of characters.

    +
    $ tr -cd '[:alnum:]'
    +

    therefore removes all non-alphanumeric characters.

    +

    The -s flag causes tr to compress sequences of identical adjacent characters in its output to a single token. For example,

    +
    $ tr -s '\n' '\n'
    +

    replaces sequences of one or more newline characters with a single newline.

    +

    The -d flag causes tr to delete all tokens of the specified set of characters from its input. In this case, only a single character set argument is used. The following command removes carriage return characters, thereby converting a file in DOS/Windows format to one in Unix format.

    +
    $ tr -d '\r'
    +

    Most versions of tr, including GNU tr and classic Unix tr, operate on single byte characters and are not Unicode compliant. An exception is the Heirloom Toolchest implementation, which provides basic Unicode support.

    +

    Ruby and Perl also have an internal tr operator, which operates analogously. Tcl's string map command is more general in that it maps strings to strings while tr maps characters to characters.

    +
    +
    +
    +

    +14. Elementary Regex

    +

    In computing, regular expressions provide a concise and flexible means for identifying strings of text of interest, such as particular characters, words, or patterns of characters. A regular expression (often shortened to regex or regexp) is written in a formal language that can be interpreted by a regular expression processor, a program that either serves as a parser generator or examines text and identifies parts that match the provided specification.

    +

    Regular expressions are used by many text editors, utilities, and programming languages to search and manipulate text based on patterns. For example, Perl, Ruby and Tcl have a powerful regular expression engine built directly into their syntax. Several utilities provided by Unix distributions—including the editor ed and the filter grep — were the first to popularize the concept of regular expressions.

    +

    Traditional Unix regular expression syntax followed common conventions but often differed from tool to tool. The IEEE POSIX Basic Regular Expressions (BRE) standard (released alongside an alternative flavor called Extended Regular Expressions or ERE) was designed mostly for backward compatibility with the traditional (Simple Regular Expression) syntax but provided a common standard which has since been adopted as the default syntax of many Unix regular expression tools, though there is often some variation or additional features. Many such tools also provide support for ERE syntax with command line arguments.

    +

    In the BRE syntax, most characters are treated as literals — they match only themselves (i.e., a matches "a"). The exceptions, listed below, are called metacharacters or metasequences.

    ++++
    +
    +

    +14.1. Lazy quantification

    +

    The standard quantifiers in regular expressions are greedy, meaning they match as much as they can, only giving back as necessary to match the remainder of the regex. For example, someone new to regexes wishing to find the first instance of an item between < and > symbols in this example:

    +
    Another whale explosion occurred on <January 26>, <2004>.
    +

    ...would likely come up with the pattern <.*>, or similar. However, this pattern will actually return "<January 26>, <2004>" instead of the "<January 26>" which might be expected, because the <title_reference>*</title_reference> quantifier is greedy — it will consume as many characters as possible from the input, and "January 26>, <2004" has more characters than "January 26".

    +

    Though this problem can be avoided in a number of ways (e.g., by specifying the text that is not to be matched: <[^>]*>), modern regular expression tools allow a quantifier to be specified as lazy (also known as non-greedy, reluctant, minimal, or ungreedy) by putting a question mark after the quantifier (e.g., <.*?>), or by using a modifier which reverses the greediness of quantifiers (though changing the meaning of the standard quantifiers can be confusing). By using a lazy quantifier, the expression tries the minimal match first. Though in the previous example lazy matching is used to select one of many matching results, in some cases it can also be used to improve performance when greedy matching would require more backtracking.

    +
    +
    +
    +

    +15. One Liners

    +

    A one-liner is textual input to the command-line of an operating system shell that performs some function in just one line of input.

    +

    The one liner can be

    +<block_quote>
      +
    1. An expression written in the language of the shell.

    2. +
    3. The invocation of an interpreter together with program source for the interpreter to run.

    4. +
    5. The invocation of a compiler together with source to compile and +instructions for executing the compiled program.

    6. +
    </block_quote>

    Certain dynamic scripting languages such as AWK, sed, and perl have traditionally been adept at expressing one-liners. Specialist shell interpreters such as these Unix shells or the Windows PowerShell, allow for the construction of powerful one-liners.

    +

    The use of the phrase one-liner has been widened to also include program-source for any language that does something useful in one line.

    +

    The word One-liner has two references in the index of the book The AWK Programming Language (the book is often referred to by the abbreviation TAPL). It explains the programming language AWK, which is part of the Unix operating system. The authors explain the birth of the One-liner paradigm with their daily work on early Unix machines:

    +
    “The 1977 version had only a few built-in variables and predefined functions. It was designed for writing short programs [...] Our model was that an invocation would be one or two lines long, typed in and used immediately. Defaults were chosen to match this style [...] We, being the authors, knew how the language was supposed to be used, and so we only wrote one-liners.”
    +

    Notice that this original definition of a One-liner implies immediate execution of the program without any compilation. So, in a strict sense, only source code for interpreted languages qualifies as a One-liner. But this strict understanding of a One-liner was broadened in 1985 when the IOCCC introduced the category of Best One Liner for C, which is a compiled language.

    +

    The TAPL book contains 20 examples of One-liners (A Handful of Useful awk One-Liners) at the end of the book's first chapter.

    +

    Here are the first few of them:

    +<block_quote>
      +
    1. +

      Print the total number of input lines:

      +

      END { print NR }

      +
    2. +
    3. +

      Print the tenth input line:

      +

      NR == 10

      +
    4. +
    5. +

      Print the last field of every input line:

      +

      { print $NF }

      +
    6. +
    </block_quote>

    One-liners are also used to show off the differential expressive power of programming languages. Frequently, one-liners are used to demonstrate programming ability. Contests are often held to see who can create the most exceptional one-liner.

    +

    The following example is a C program (a winning entry in the "Best one-liner" category of the IOCCC, here split to two lines for presentation).:

    +
    main(int c,char**v){return!m(v[1],v[2]);}m(char*s,char*t){return
    +*t-42?*s?63==*t|*s==*t&&m(s+1,t+1):!*t:m(s,t+1)||*s&&m(s+1,t);}
    +

    This one-liner program is a glob pattern matcher. It understands the glob characters '*' meaning 'zero or more characters' and '?' meaning exactly one character, just like most Unix shells.

    +

    Run it with two args, the string and the glob pattern. The exit status is 0 (shell true) when the pattern matches, 1 otherwise. The glob pattern must match the whole string, so you may want to use * at the beginning and end of the pattern if you are looking for something in the middle. Examples:

    +
    $ prog foo 'f??'; echo $?
    +
    +$ prog 'best short program' '??st*o**p?*'; echo $?
    +

    Here is a one line shell script to show directories:

    +
    $ ls -R | grep ":$" | sed -e 's/:$//' -e 's/[^-][^\/]*\//--/g' -e 's/^/   /' -e 's/-/|/'
    +
    +
    +
    + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/ch9Using_Linux_Tools.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/ch9Using_Linux_Tools.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1139 @@ + + + +Using Linux tools + + + + + + + +
    +
    + +
    +
    +

    +
    +
    + +
    +

    +1. Introducing Linux

    +

    (Attribution : A significant chunk of the content under this section is based on data from Wikipedia and the Linux Documentation Project)

    +

    Linux (usually pronounced ˈlɪnəks') is a generic term referring to Unix-like computer operating systems based on the Linux kernel, where a kernel is the intermediate layer between the hardware and the applications. The kernel is, on an abstract level, the core of (most) operating systems, that manages the various system resources. The development of the Linux OS is considered the basis for Free and Open Source Software (FOSS) collaboration since typically the underlying source code can be used, modified freely, and redistributed by anyone under the terms of the GNU (a recursive acronym for "GNU's Not Unix!") Global Public License (GPL) and other free software licences. This freedom to access and reuse various components of a system, is one of the primary reasons for the popularity of Linux.

    +

    Linux is installed on a variety of computer hardware, that include mobile phones, embedded devices and supercomputers, but is infamous for its use in servers.

    +

    The name "Linux" comes from the Linux kernel, originally written in 1991 by Linus Torvalds. The rest of the system usually comprises components such as the Apache HTTP Server, the X Window System, the GNOME and KDE desktop environments, and utilities and libraries from the GNU Project (announced in 1983 by Richard Stallman). Commonly-used applications with desktop Linux systems include the Mozilla Firefox web-browser and the OpenOffice.org office application suite. The GNU contribution is the basis for the Free Software Foundation's preferred name GNU/Linux. The kernel's mascot is a penguin named "Tux". Mozilla Firefox and OpenOffice.org are open-source projects which can be run on most Operating Systems, including proprietary ones.

    +
    +

    +1.1. Historical Background

    +
    +

    +1.1.1. Events leading to the creation

    +
      +
    • The Unix operating system was developed in the 1960s and released for public use in 1970. Its accessibility and portability caused it to be widely adopted, copied and modified by academic institutions and businesses. Its design became influential to authors of other systems. Other free operating systems include the Berkeley Software Distribution (BSD), developed at the University of California at Berkeley, and MINIX which was released by Andrew S. Tanenbaum. The development and adoption of BSD and MINIX were limited due to various reasons, and this lack of a widely-adopted and free kernel triggered Linus Torvalds into starting his project.

    • +
    • In 1983, Richard Stallman started the GNU project with the goal of creating a free UNIX-like operating system. As part of this work, he wrote the GNU General Public License (GPL). By the early 1990s there was almost enough available software to create a full operating system. However, the GNU kernel, called Hurd, failed to attract enough attention from developers leaving GNU incomplete.

    • +
    +
    +
    +

    +1.1.2. The Creation of Linux

    +

    In 1991, Linus Torvalds began a project at the University of Helsinki that later became the Linux kernel. It was initially a terminal (command-line) emulator, which Torvalds used to access the large UNIX servers of the university. He wrote the program targeting just the hardware he was using and independent of an operating system because he wanted to use the functions of his computer with an 80386 processor. Development was done on Minix using the GNU C compiler. This application is still the main choice for compiling Linux today (although the code can be built with other compilers, such as the Intel C Compiler).

    +

    Torvalds continues to direct the development of the kernel. Stallman heads the Free Software Foundation, which in turn supports the GNU components. Finally, individuals and corporations develop third-party non-GNU components, which constitute a vast body of work and including kernel modules, and user applications and libraries. Linux vendors and communities combine and distribute the kernel, GNU components, and non-GNU components, with additional package management software in the form of Linux distributions.

    +
    +
    +
    +

    +1.2. Design and Implications

    +

    A Linux-based system is a modular Unix-like operating system, deriving much of its basic design from principles established in Unix earlier. Such a system uses a monolithic kernel, called the Linux kernel, which handles process control, networking, and peripheral and file system access. Device drivers are integrated directly with the kernel. Separate projects that interface with the kernel provide much of the system's higher-level functionality. The GNU userland is an important part of most Linux-based systems, providing the most common implementation of the C library, a popular shell, and many of the common Unix tools which carry out many basic operating system tasks. The graphical user interface (or GUI) used by most Linux systems is based on the "X Window System".

    +
    +

    +1.2.1. User Interface

    +

    Users can control a Linux-based system through a command line interface (or CLI), a graphical user interface (or GUI), or through controls attached to the associated hardware (this is common for embedded systems). For desktop systems, the default mode is usually the GUI. On desktop machines, "KDE", "GNOME" and "Xfce" are the most popular user interfaces,though a variety of additional user interfaces exist. Most popular user interfaces run on top of the "X Window System" (or X), which enables a graphical application running on one machine to be displayed and controlled from another in a network.

    +

    A Linux system also provides a CLI of some sort through a shell, which is the traditional way of interacting with a Unix system. A Linux distribution specialized for servers may use the CLI as its only interface. A “headless system” (system run without even a monitor) can be controlled by the command line via a remote-control protocol such as SSH or telnet. The CLI is particularly suited for automation of repetitive or delayed tasks, and provides very simple inter-process communication. A graphical terminal emulator program is often used to access the CLI from a Linux desktop.

    +
    +
    +

    +1.2.2. Development

    +

    The primary difference between Linux and many other popular contemporary operating systems is that the Linux kernel and other components are free and open source software. Linux is not the only such operating system, although it is by far the most widely used. Some free and open source software licenses are based on the principle of "copyleft", a kind of reciprocity: any work derived from a copyleft piece of software must also be copyleft itself. The most common free software license, the GNU GPL, is a form of copyleft, and is used for the Linux kernel and many of the components from the GNU project.

    +

    Linux based distributions are intended by developers for interoperability with other operating systems and established computing standards. Linux systems adhere to POSIX, SUS, ISO and ANSI standards where possible, although to date only one Linux distribution has been POSIX.1 certified, Linux-FT.Free software projects, although developed in a collaborative fashion, are often produced independently of each other. The fact that the software licenses explicitly permit redistribution, however, provides a basis for larger scale projects that collect the software produced by stand-alone projects and make it available all at once in the form of a Linux distribution.

    +

    A Linux distribution, commonly called a "distro", is a project that manages a remote collection of system software and application software packages available for download and installation through a network connection. This allows the user to adapt the operating system to his/her specific needs. Distributions are maintained by individuals, loose-knit teams, volunteer organizations, and commercial entities. A distribution can be installed using a CD that contains distribution-specific software for initial system installation and configuration. A package manager such as Synaptic or YAST allows later package upgrades and installations. A distribution is responsible for the default configuration of the installed Linux kernel, general system security, and more generally integration of the different software packages into a coherent whole.

    +
    +
    +

    +1.2.3. Community

    +

    A distribution is largely driven by its developer and user communities. Some vendors develop and fund their distributions on a volunteer basis. Examples include Debian and the Debian-based, Ubuntu. Others maintain a community version of their commercial distributions, as Red Hat does with Fedora.

    +

    In many cities and regions, local associations known as Linux Users Groups (LUGs) seek to promote their preferred distribution and by extension free software. They hold meetings and provide free demonstrations, training, technical support, and operating system installation to new users. Many Internet communities also provide support to Linux users and developers. Most distributions and free software / open source projects have IRC (Internet Relay Chat) chatrooms or newsgroups. Online forums are another means for support. Linux distributions host mailing lists; commonly there will be a specific topic such as usage or development for a given list. All these can be found simply by running an appropriate search on Google.

    +

    Although Linux distributions are generally available without charge, several large corporations sell, support, and contribute to the development of the components of the system and of free software. These include Dell, IBM, HP, Oracle, Sun Microsystems, Novell, Nokia. A number of corporations, notably Red Hat, have built their entire business around Linux distributions.

    +
    +
    +

    +1.2.4. Can I make a profit out of running a business involving Linux?

    +

    The answer is, "Yes!". The free software licenses, on which the various software packages of a distribution built on the Linux kernel are based, explicitly accommodate and encourage commercialization; the relationship between a Linux distribution as a whole and individual vendors may be seen as symbiotic. One common business model of commercial suppliers is charging for support, especially for business users. A number of companies also offer a specialized business version of their distribution, which adds proprietary support packages and tools to administer higher numbers of installations or to simplify administrative tasks. Another business model is to give away the software in order to sell hardware. Examples of corporations that are extensively (and sometimes exclusively) open-source and Linux-powered , with successful revenue generation models involving these, are Google, SUN, Mozilla, etc.

    +
    +
    +

    +1.2.5. Programming on Linux

    +

    Most Linux distributions support dozens of programming languages. The most common collection of utilities for building both Linux applications and operating system programs is found within the GNU toolchain, which includes the GNU Compiler Collection (GCC) and the GNU build system. Amongst others, GCC provides compilers for Ada, C, C++, Java, and Fortran. The Linux kernel itself is written to be compiled with GCC. Proprietary compilers for Linux include the Intel C++ Compiler, Sun Studio, and IBM XL C/C++ Compiler.

    +

    Most distributions also include support for PHP, Perl, Ruby, Python and other dynamic languages. Examples of languages that are less common, but still supported, are C# via the Mono project, sponsored by Novell, and Scheme. A number of Java Virtual Machines and development kits run on Linux, including the original Sun Microsystems JVM (HotSpot), and IBM's J2SE RE, as well as many open-source projects like Kaffe.

    +

    The two main frameworks for developing graphical applications are those of GNOME and KDE. These projects are based on the GTK+ and Qt widget toolkits, respectively, which can also be used independently of the larger framework. Both support a wide variety of languages. There are a number of Integrated Development Environments (IDEs) available including Anjuta, Code::Blocks, Eclipse, KDevelop, Lazarus, MonoDevelop, NetBeans, and Omnis Studio while the long-established editors Vim and Emacs remain popular.

    +
    +
    +
    +

    +1.3. Reasons for Using Linux

    +
    • Linux is free:

    +

    As in "free beer". Linux can be downloaded in its entirety from the Internet completely for free. No registration fees, no costs per user, free updates, and freely available source code in case you want to change the behavior of your system. +Most of all, Linux is free as in "free speech": +The license commonly used is the GNU Public License (GPL). The license says that anybody who may want to do so, has the right to change Linux and eventually to redistribute a changed version, on the one condition that the code is still available after redistribution. In practice, you are free to grab a kernel image and sell the new code, as long as your customers can still have a copy of that code.

    +
    • Linux is portable to any hardware platform:

    +

    A vendor, who wants to sell a new type of computer and who does not know what kind of OS his/her new machine will run, can take a Linux kernel and make it work on his/her hardware, because documentation related to this activity is freely available.

    +
    • Linux was made to keep on running:

    +

    As with UNIX, a Linux system expects to run without rebooting all the time. That is why a lot of tasks are being executed at night or scheduled automatically for other times, resulting in higher availability during busier periods and a more balanced use of the hardware. This property allows for Linux to be applicable to environments where people do not have the time or the possibility to control their systems constantly.

    +
    • Linux is secure and versatile:

    +

    The security model used in Linux is based on the UNIX idea of security, which is known to be robust and of proven quality. But Linux is not only safe from attacks from the Internet: it will adapt equally to other situations, utilizing the same high standards for security.

    +
    • Linux is scalable:

    +

    From a Palmtop with 2 MB of memory to a petabyte storage cluster with hundreds of nodes: add or remove the appropriate packages and Linux fits all. One does not need a supercomputer anymore,because you can use Linux to do big things using the building blocks provided with the system. If one wants to do little things, such as making an operating system for an embedded processor or just recycling your old 486, Linux will do that as well.

    +
    • The Linux OS and Linux applications have very short debug−times:

    +

    Because Linux has been developed and tested by thousands of people, both errors and people to fix them are found very quickly. It often happens that there are only a couple of hours between discovery and fixing of a bug.

    +
    +
    +
    +

    +2. Getting Started

    +
    +

    +2.1. Logging in, activating the user interface and logging out

    +

    In order to work on a Linux system directly, one needs to provide a user name and password. You always need to authenticate to the system. Most PC−based Linux systems have two basic modes for a system to run in: either quick and clean in text console mode,which includes with mouse, multitasking and multi−user features, or in graphical console mode, which looks better but eats more system resources.

    +
    +

    +2.1.1. Graphical Mode

    +

    This is the default nowadays on most desktop computers. You know you will be connecting to the system using graphical mode when you are first asked for your user name, and then to type your password.

    +

    To log in, make sure the mouse pointer is in the login window, provide your user name and password to the system and click OK or press Enter. +It is generally considered a bad idea to connect (graphically) using the root user name, the system adminstrator's account, since the use of graphics includes running a lot of extra programs, in root's case with a lot of extra permissions. To keep all risks as low as possible, use a normal user account to connect graphically. But there are enough risks to keep this in mind as a general advice, for all use of the root account: only log in as root when extra privileges are required.

    +

    After entering your user name/password combination, it can take a little while before the graphical environment is started, depending on the CPU speed of your computer, on the software you use and on your personal settings.

    +

    To continue, you will need to open a terminal window or xterm for short (X being the name for the underlying software supporting the graphical environment). This program can be found in the Applications−>Utilities->System Tools or Internet menu, depending on what window manager you are using. There might be icons that you can use as a shortcut to get an xterm window as well, and clicking the right mouse button on the desktop background will usually present you with a menu containing a terminal window application.

    +

    While browsing the menus, you will notice that a lot of things can be done without entering commands via the keyboard. For most users, the good old point−n−click method of dealing with the computer will do. But for those who want to enter the "heart" of the system, a tool stronger than a mouse will be required to handle the various tasks. This tool is the shell, and when in graphical mode, we activate our shell by opening a terminal window.

    +

    A terminal window should always show a command prompt when you open one. This terminal shows a standard prompt, which displays the user's login name, and the current working directory, represented by the twiddle (~)

    +

    Another common form for a prompt is this one: +[

    +
    +

    user@host
    +

    dir]

    +

    In the above example, user will be your login name, hosts the name of the machine you are working on, and dir an indication of your current location in the file system. Prompts can display all kinds of information, but they are not part of the commands you are giving to your system. To disconnect from the system in graphical mode, you need to close all terminal windows and other applications. After that, hit the logout icon or find Log Out in the menu. Closing everything is not really necessary, and the system can do this for you, but session management might put all currently open applications back on your screen when you connect again, which takes longer and is not always the desired effect. However, this behavior is configurable.

    +

    When you see the login screen again, asking to enter user name and password, logout was successful.

    +
    +
    +

    +2.1.2. Text Mode

    +

    One is in text mode when the whole screen is black, showing (in most cases white) characters. A text mode login screen typically shows some information about the machine you are working on, the name of the machine and a prompt waiting for you to log in.

    +

    The login is different from a graphical login, in that you have to hit the Enter key after providing your user name, because there are no buttons on the screen that you can click with the mouse. Then you should type your password, followed by another Enter. You will not see any indication that you are entering something, not even an asterisk, and you won't see the cursor move. But this is normal on Linux and is done for security +reasons.

    +

    When the system has accepted you as a valid user, you may get some more information, called the message of the day, which can be anything. Additionally, it is popular on UNIX systems to display a fortune cookie, which contains some general wise or unwise (this is up to you) thoughts. After that, you will be given a shell, indicated with the same prompt that you would get in graphical mode.

    +

    Also in text mode: log in as root only to do setup and configuration that absolutely requires administrator privileges, such as adding users, installing software packages, and performing network and other system configuration. Once you are finished, immediately leave the special account and resume your work as a non−privileged user.

    +

    Logging out is done by entering the logout command, followed by Enter. You are successfully disconnected from the system when you see the login screen again.Don't power−off the computer after logging out. It is not meant to be shut off without application of the proper procedures for halting the system. Powering it off without going through the halting process might cause severe damage!

    +
    +
    +
    +
    +

    +3. Basic Commands

    +
    +

    +3.1. ls

    +

    When invoked without any arguments, ls lists the files in the current working directory. A directory that is not the current working directory can be specified and ls will list the files there. The user also may specify any list of files and directories. In this case, all files and all contents of specified directories will be listed. The name ls is derived from list segments which was used in earlier systems.

    +

    Files whose names start with "." are not listed, unless the -a flag is specified or the files are specified explicitly.

    +

    Without options, ls displays files in a bare format. This bare format however makes it difficult to establish the type, permissions, and size of the files. The most common options to reveal this information or change the list of files are:

    +<block_quote>
      +
    • -l long format, displaying Unix file types, permissions, number of hard links, owner, group, size, date, and filename

    • +
    • -F appends a character revealing the nature of a file, for example, * for an executable, or / for a directory. Regular files have no suffix.

    • +
    • -a lists all files in the given directory, including those whose names start with "." (which are hidden files in Unix). By default, these files are excluded from the list.

    • +
    • -R recursively lists subdirectories. The command ls -R / would therefore list all files.

    • +
    • -d shows information about a symbolic link or directory, rather than about the link's target or listing the contents of a directory.

    • +
    • -t sort the list of files by modification time.

    • +
    • -h print sizes in human readable format. (e.g., 1K, 234M, 2G, etc.)

    • +
    </block_quote>

    In some environments, providing the option --color (for GNU ls) or -G (FreeBSD ls) causes ls to highlight different types of files with different colors, instead of with characters as -F would. To determine what color to use for a file, GNU ls checks the Unix file type, the file permissions, and the file extension, while FreeBSD ls checks only the Unix file type and file permissions.:

    +
    $ ls
    +jeeves.rst psmith.html blandings.html
    +$ ls -l
    +drwxr--r--   1 plum  editors   4096  jeeves
    +-rw-r--r--   1 plum  editors  30405  psmith
    +-r-xr-xr-x   1 plum  plum      8460  blandings
    +

    Here "$" actually is the beginning of the prompt. This is typical in most Unix-based systems.

    +
    +
    +

    +3.2. date

    +

    The Unix date command displays the time and date. The super-user can use it to set the system clock.

    +

    With no options, the date command displays the current date and time, including the abbreviated day name, abbreviated month name, day of the month, the time separated by colons, the timezone name, and the year. For example:

    +
    $date
    +Tue Sep  8 12:01:45 IST 2009
    +

    On some systems to set the current date and time to September 8, 2004 01:22 you type:

    +
    $date --set="20040908 01:22"
    +

    In order to view the various options for the date command, type:

    +
    $man date
    +

    This will take you to the "Manual" page comprising of all the details on the date command. You can return to the terminal from the "man" page by pressing the Esc key in the keyboard and typing ":q" in that order.

    +
    +
    +

    +3.3. cd

    +

    This stands for "change directory". When one wants to go up to the parent directory, bypassing the tree of directories one has entered, “ cd ..” can be used.

    +

    One dot '.' represents the current directory while two dots '..' represent the parent directory.

    +

    “ cd -” will return you to the previous directory (a bit like an “undo”).

    +

    You can also use cd absolute path or cd relative path (see below):

    +

    Absolute paths:

    +<block_quote>

    An “ absolute path” is easily recognised from the leading forward slash, /. The / means that you start at the top level directory and continue down.

    </block_quote>

    For example to get to /boot/grub you would type:

    +
    $cd /boot/grub
    +

    This is an absolute path because you start at the top of the hierarchy and go downwards from there (it doesn't matter where in the filesystem you were when you typed the command).

    +

    Relative paths:

    +<block_quote>

    A “ relative path” doesn't have a preceding slash. Use a relative path when you start from a directory below the top level directory structure. This is dependent on where you are in the filesystem.

    +

    For example if you are in root's home directory and want to get to /root/music, you type:

    +
    $ cd music
    </block_quote>

    Please note that there is no / using the above cd command. Using a / would cause this to be an absolute path, working from the top of the hierarchy downward.

    +
    +
    +

    +3.4. who

    +

    The standard Unix command who displays a list of users who are currently logged into a computer.

    +

    The who command is related to the command w, which provides the same information but also displays additional data and statistics.:

    +
    $who
    +beeblebrox tty7         2009-09-08 10:50 (:0)
    +beeblebrox pts/0        2009-09-08 11:25 (:0.0)
    +dumbledore pts/1        2009-09-08 18:11 (potter.xyz.in)
    +beeblebrox pts/2        2009-09-08 18:53 (:0.0)
    +

    The command can be invoked with the arguments am i or am I (so it is invoked as who am i or * who am I*), showing information about the current terminal only (see the -m option below, of which this invocation is equivalent).

    +

    In order to find out the various options that can be appended to the who command, check the man page by typing out the following in the terminal:

    +
    $man who
    +

    This will take you to the "Manual" page containing details about the who command

    +
    +
    +

    +3.5. mkdir

    +

    This command is used to make a new directory. Normal usage is as straightforward as follows:

    +
    $mkdir name_of_directory
    +

    Where name_of_directory is the name of the directory one wants to create. When typed as above (ie. normal usage), the new directory would be created within the current directory. On Unix, multiple directories can be specified, and mkdir will try to create all of them.

    +
    +

    +3.5.1. Options

    +

    On Unix-like operating systems, mkdir takes options. Three of the most common options are:

    +<block_quote>
      +
    • -p: will also create all directories leading up to the given directory that do not exist already. If the given directory already exists, ignore the error.

    • +
    • -v: display each directory that mkdir creates. Most often used with -p.

    • +
    • -m: specify the octal permissions of directories created by mkdir.

    • +
    </block_quote>

    -p is most often used when using mkdir to build up complex directory hierarchies, in case a necessary directory is missing or already there. -m is commonly used to lock down temporary directories used by shell scripts.

    +
    +
    +

    +3.5.2. Examples

    +

    An example of -p in action is:

    +
    $mkdir -p /tmp/a/b/c
    +

    If /tmp/a exists but /tmp/a/b does not, mkdir will create /tmp/a/b before creating /tmp/a/b/c.

    +

    And an even more powerful command, creating a full tree at once (this however is a Shell extension, nothing mkdir does itself):

    +
    $mkdir -p tmpdir/{trunk/sources/{includes,docs},branches,tags}
    +

    This will create:

    +<definition_list><definition_list_item><term>tmpdir - branches</term><definition>
      +
    • tag

    • +
    • <definition_list><definition_list_item><term>trunk - sources - includes</term><definition>
      • docs

      </definition>
      </definition_list_item>
      </definition_list>
    • +
    </definition>
    </definition_list_item>
    </definition_list>
    +
    +
    +
    +
    +

    +4. Getting Help

    +
    +

    +4.1. apropos and whatis

    +

    This is a command to search the manual pages files in Unix and Unix-like operating systems.

    +
    $ apropos grep
    +egrep       egrep (1)       Search a file for a pattern using full regular expressions
    +fgrep       fgrep (1)       Search a file for a fixed-character string
    +fmlgrep     fmlgrep (1)     Search a file for a pattern
    +grep        grep (1)        Search a file for a pattern
    +gzgrep      gzgrep (1)      Search a possibly compressed file for a regular expression
    +nisgrep     nismatch (1)    Utilities for searching NIS+ tables
    +pgrep       pgrep (1)       Find or signal a process by name or other attribute
    +zgrep       zgrep (1)       Search a possibly compressed file for a regular expression
    +...
    +

    In this example, the user uses apropos to search for the string "grep", and apropos returns the indicated man pages that include the term "grep".

    +

    A short index of explanations for commands is available using the whatis command, like in the examples below:

    +
    $whatis ls
    +ls (1)           - list directory contents
    +

    This displays short information about a command, and the first section in the collection of man pages that contains an appropriate page.

    +

    If you don't know where to get started and which man page to read, apropos gives more information. Say that you do not know how to start a browser, then you could enter the following command:

    +
    $apropos browser
    +gmusicbrowser (1)    - Jukebox for large collections of audio files
    +infobrowser (1)      - read Info documents
    +libsmbclient (7)     - An extension library for browsers and that               can be used...
    +opera (1)            - a standards-compliant graphical Web browser
    +sensible-browser (1) - sensible editing, paging, and web browsing
    +smbtree (1)          - A text based smb network browser
    +tvtk_doc (1)         - A GUI based TVTK documentation search browser.
    +viewres (1)          - graphical class browser for Xt
    +w3m (1)              - a text based Web browser and pager
    +www-browser (1)      - a text based Web browser and pager
    +...
    +
    +
    +

    +4.2. man

    +

    Man pages (short for "manual pages") are the extensive documentation that comes preinstalled with almost all substantial Unix and Unix-like operating systems. The Unix command used to display them is man. Each page is a self-contained document.

    +

    To read a manual page for a Unix command, one can use:

    +
    $ man <command_name>
    +

    at a shell prompt; for example, "man ftp". In order to simplify navigation through the output, man generally uses the less terminal pager.

    +

    Pages are traditionally referred to using the notation "name(section)"; for example, ftp(1). The same page name may appear in more than one section of the manual, this can occur when the names of system calls, user commands, or macro packages coincide. Two examples are man(1) and man(7), or exit(2) and exit(3). The syntax for accessing the non-default manual section varies between different man implementations. On Linux and BSD, for example, the syntax for reading *printf(3) is:

    +
    $man 3 printf
    +

    Another example:

    +
    $man man
    +

    The previous example will take you to the "Manual" page entry about manual pages!

    +
    +

    +4.2.1. Layout

    +

    All man pages follow a common layout that is optimized for presentation on a simple ASCII text display, possibly without any form of highlighting or font control. Sections present may include:

    +<definition_list><definition_list_item><term>NAME</term><definition>

    The name of the command or function, followed by a one-line description of what it does.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SYNOPSIS</term><definition>

    In the case of a command, you get a formal description of how to run it and what command line options it takes. For program functions, a list of the parameters the function takes and which header file contains its definition. For experienced users, this may be all the documentation they need.

    </definition>
    </definition_list_item>
    <definition_list_item><term>DESCRIPTION</term><definition>

    A textual description of the functioning of the command or function.

    </definition>
    </definition_list_item>
    <definition_list_item><term>EXAMPLES</term><definition>

    Some examples of common usage.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SEE ALSO</term><definition>

    A list of related commands or functions.

    </definition>
    </definition_list_item>
    </definition_list>

    Other sections may be present, but these are not well standardized across man pages. Common examples include: OPTIONS, EXIT STATUS, ENVIRONMENT, KNOWN BUGS, FILES, AUTHOR, REPORTING BUGS, HISTORY and COPYRIGHT.

    +

    These days virtually every Unix command line application comes with its man page, and many Unix users perceive a lack of man pages as a sign of low quality; indeed, some projects, such as Debian, go out of their way to write man pages for programs lacking one. Few alternatives to man have enjoyed much popularity, with the possible exception of the GNU project's "info" system, an early and simple hypertext system.

    +

    However, the format of a single page for each application, the lack of classification within the sections and the relatively unsophisticated formatting facilities have motivated the development of alternative documentation systems, such as the previously mentioned "info" system.

    +

    Most Unix GUI applications (particularly those built using the GNOME and KDE development environments) now provide end-user documentation in HTML and include embedded HTML viewers such as yelp for reading the help within the application.

    +

    Usually the man pages are written in English. Translations into other languages can be also available on the system.

    +

    The default format of the man pages is troff, with either the macro package man (appearance oriented) or on some systems mdoc (semantic oriented). This makes it possible to typeset a man page to PostScript, PDF and various other formats for viewing or printing.

    +
    +
    +
    +

    +4.3. info

    +

    info is a software utility which forms a hypertextual, multipage documentation and help viewer working on a command line interface, useful when there is no GUI available.

    +

    The syntax is

    +
    $ info <command_name>
    +

    info processes info files, which are Texinfo formatted files, and presents the documentation as a tree, with simple commands to traverse the tree and to follow cross references. For instance

    +<block_quote>
      +
    • n goes to the next page.

    • +
    • p goes to the previous page.

    • +
    • u goes to the upper page.

    • +
    • l goes to the last(visited) node

    • +
    • To follow a cross reference, the cursor can be moved over a link (a word preceded by a <title_reference>*</title_reference>) and enter pressed.

    • +
    </block_quote>

    info was initially written for use with GNU/Linux and then ported to other Unix-like operating systems.

    +
    +
    +

    +4.4. --help

    +

    Most GNU commands support the --help, which gives a short explanation about how to use the command and a list of available options. Below is the output of this option with the cat command:

    +
    $ userprompt@host: cat --help
    +Usage: cat [OPTION] [FILE]...
    +Concatenate FILE(s), or standard input, to standard output.
    +
    +  -A, --show-all           equivalent to -vET
    +  -b, --number-nonblank    number nonempty output lines
    +  -e                       equivalent to -vE
    +  -E, --show-ends          display $ at end of each line
    +  -n, --number             number all output lines
    +  -s, --squeeze-blank      suppress repeated empty output lines
    +  -t                       equivalent to -vT
    +  -T, --show-tabs          display TAB characters as ^I
    +  -u                       (ignored)
    +  -v, --show-nonprinting   use ^ and M- notation, except for LFD and              TAB
    +  --help     display this help and exit
    +  --version  output version information and exit
    +
    +With no FILE, or when FILE is -, read standard input.
    +
    +Examples:
    +  cat f - g  Output f's contents, then standard input, then g's           contents.
    +  cat        Copy standard input to standard output.
    +
    +Report bugs to <bug-coreutils@gnu.org>.
    +
    +
    +
    +

    +5. Basic file handling

    +
    +

    +5.1. cp

    +

    cp is the command entered in a Unix shell to copy a file from one place to another, possibly on a different filesystem. The original file remains unchanged, and the new file may have the same or a different name.

    +
    +

    +5.1.1. Usage

    +

    To copy a file to another file:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ][ -- ] SourceFile TargetFile
    +

    To copy a file to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -r | -R ] [ -- ] SourceFile ...              TargetDirectory
    +

    To copy a directory to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -- ] { -r | -R }
    +SourceDirectory ... TargetDirectory
    +
    +
    +

    +5.1.2. Flags

    +

    -f (force) – specifies removal of the target file if it cannot be opened for write operations. The removal precedes any copying performed by the cp command.

    +

    -P – makes the cp command copy symbolic links. The default is to follow symbolic links, that is, to copy files to which symbolic links point.

    +

    -i (interactive) – prompts you with the name of a file to be overwritten. This occurs if the TargetDirectory or TargetFile parameter contains a file with the same name as a file specified in the SourceFile or SourceDirectory parameter. If you enter y or the locale's equivalent of y, the cp command continues. Any other answer prevents the cp command from overwriting the file.

    +

    -p (preserve) – duplicates the following characteristics of each SourceFile/SourceDirectory in the corresponding TargetFile and/or TargetDirectory:

    +<block_quote>
      +
    • The time of the last data modification and the time of the last access.

    • +
    • The user ID and group ID (only if it has permissions to do this)

    • +
    • The file permission bits and the SUID and SGID bits.

    • +
    </block_quote>

    -R (recursive) – copy directories (recursively copying all the contents)

    +
    +
    +

    +5.1.3. Examples

    +

    To make a copy of a file in the current directory, enter:

    +
    $ cp prog.c prog.bak
    +

    This copies prog.c to prog.bak. If the prog.bak file does not already exist, the cp command creates it. If it does exist, the cp command replaces it with a copy of the prog.c file.

    +

    To copy a file in your current directory into another directory, enter:

    +
    $ cp zaphod /home/books/hhgg
    +

    This copies the jones file to /home/books/hhgg/zaphod.

    +

    To copy a file to a new file and preserve the modification date, time, and access control list associated with the source file, enter:

    +
    $ cp -p martin_luther_king martin_luther_king.jr
    +

    This copies the martin_luther_king file to the martin_luther_king.jr file. Instead of creating the file with the current date and time stamp, the system gives the martin_luther_king.jr file the same date and time as the martin_luther_king file. The martin_luther_king.jr file also inherits the martin_luther_king file's access control protection.

    +

    To copy all the files in a directory to a new directory, enter:

    +
    $ cp /home/galactica/clients/* /home/hhgg/customers
    +

    This copies only the files in the clients directory to the customers directory.

    +

    To copy a directory, including all its files and subdirectories, to another directory, enter:

    +<block_quote>

    $ cp -R /home/hhgg/clients /home/hhgg/customers

    </block_quote>

    This copies the clients directory, including all its files, subdirectories, and the files in those subdirectories, to the customers/clients directory.

    +

    To copy a specific set of files of any extension to another directory, enter:

    +
    $ cp zaphod arthur ford /home/hhgg/clients
    +

    This copies the zaphod, arthur, and ford files in your current working directory to the /home/hhgg/clients directory.

    +

    To use pattern-matching characters to copy files, enter:

    +
    $ cp programs/*.py .
    +

    This copies the files in the programs directory that end with .py to the current directory, signified by the single "." (dot). You must type a space between the py and the final dot.

    +
    +
    +
    +

    +5.2. mv

    +

    mv (short for move) is a Unix command that moves one or more files or directories from one place to another. The original file is deleted, and the new file may have the same or a different name. If possible (i.e. when the original and new files are on the same file system), mv will rename the file instead. Write permission is required on all directories being modified.

    +
    +

    +5.2.1. Conflicting existing file

    +

    In all cases, when a file is moved to have the name of an existing file (in the same directory), the existing file is deleted. If the existing file is not writable but is in a directory that is writable, then the mv command asks for confirmation if possible (i.e. if run from a terminal) before proceeding, unless the -f (force) option is used.

    +
    +
    +

    +5.2.2. Differences with copy and delete

    +

    Note that, usually, when moving files within the same volume, moving (and/or renaming) is not the same as simply copying and then deleting the original. When moving a file, the link is simply removed from the old parent directory and added to the new parent directory. However, the file itself is untouched (i.e. it has the same inodes and resides at the same place on the disk). For example, you cannot copy a file you cannot read, but you can move (and/or rename) it (provided you have write permission to its old and new parent directories). Also, suppose there is a non-empty directory you do not have write permission to. You cannot delete this directory (since you cannot delete its contents); but you can move (and/or rename) it. Also, since moving between filenames on a single volume does not involve copying, it is faster and does not place strain of lots of reads and writes on the disk. Moving files across different volumes, however, does necessitate copying and deleting.

    +
    +
    +

    +5.2.3. Examples

    +
    $ mv myfile mynewfilename    renames a file
    +$ mv myfile otherfilename    renames a file and deletes the existing            file "myfile"
    +$ mv myfile /myfile          moves 'myfile' from the current            directory to the root directory
    +$ mv myfile dir/myfile       moves 'myfile' to 'dir/myfile' relative            to the current directory
    +$ mv myfile dir              same as the previous command (the          filename is implied to be the same)
    +$ mv myfile dir/myfile2      moves 'myfile' to dir and renames it to            'myfile2'
    +$ mv foo bar baz dir         moves multiple files to directory dir
    +$ mv --help                  shows a very concise help about the                syntax of the command
    +$ man mv                     prints an extensive user manual for                'mv' in the terminal
    +

    In all cases, the file or files being moved or renamed can be a directory.

    +

    Note that when the command is called with two arguments (as mv name1 name2 or mv name1 /dir/name2), it can have three different effects, depending on whether name2 does not exist, is an existing file, or is an existing directory. If the user intends to refer to an existing directory, /. (or in some Unix versions / is sufficient) may be appended to the name to force the system to check this. To move a file to a new directory, the directory must be created first.

    +
    +
    +
    +

    +5.3. rm

    +

    rm (short for "remove") is one of several basic Unix command lines that operates on files. It is used to delete files from a filesystem. The data is not actually destroyed. Only the index listing where the file is stored is destroyed, and the storage is made available for reuse. There are undelete utilities that will attempt to reconstruct the index and can bring the file back if the parts were not reused.

    +

    Here's example to remove a file named "foo" from a directory, here shown with the -i option:

    +
    $ rm -i foo
    +remove foo? y
    +
    +

    +5.3.1. Options

    +

    Common options that rm accepts include:

    +<block_quote>
      +
    • -r, which removes directories, removing the contents recursively beforehand (so as not to leave files without a directory to reside in) ("recursive")

    • +
    • -i, which asks for every deletion to be confirmed ("interactive")

    • +
    • -f, which ignores non-existent files and overrides any confirmation prompts ("force")

    • +
    • -v, which shows what is being removed as it happens ("verbose")

    • +
    </block_quote>

    rm is often aliased to "rm -i" so as to avoid accidental deletion of files. If a user still wishes to delete a large number of files without confirmation, they can manually cancel out the -i argument by adding the -f option (as the option specified later on the expanded command line "rm -i -f" takes precedence).

    +

    rm -rf (variously, rm -rf /, rm -rf <title_reference>*</title_reference>, and others) is frequently used in jokes and anecdotes about Unix disasters. The rm -rf variant of the command, if run by a superuser on the root directory, would cause the contents of every writable mounted filesystem on the computer to be deleted.

    +

    rm is often used in conjunction with xargs to supply a list of files to delete:

    +
    xargs rm < filelist
    +

    When rm is used on a symbolic link, it deletes the link, but does not affect the target of the link.

    +
    +
    +

    +5.3.2. Permissions

    +

    Usually, on most filesystems, deleting a file requires write permission on the parent directory (and execute permission, in order to enter the directory in the first place). (Note that, confusingly for beginners, permissions on the file itself are irrelevant. However, GNU rm asks for confirmation if a write-protected file is to be deleted, unless the -f option is used.)

    +

    To delete a directory (with rm -r), one must delete all of its contents recursively. This requires that one must have read and write and execute permission to that directory (if it's not empty) and all non-empty subdirectories recursively (if there are any). The read permissions are needed to list the contents of the directory in order to delete them. This sometimes leads to an odd situation where a non-empty directory cannot be deleted because one doesn't have write permission to it and so cannot delete its contents; but if the same directory were empty, one would be able to delete it.

    +

    If a file resides in a directory with the sticky bit set, then deleting the file requires one to be the owner of the file.

    +
    +
    +
    +
    +

    +6. Command Line Arguments

    +

    In computer command line interfaces, a command line argument is an argument sent to a program being called. In general, a program can take any number of command line arguments, which may be necessary for the program to run, or may even be ignored, depending on the function of that program.

    +

    For example, in Unix and Unix-like environments, an example of a command-line argument is:

    +
    rm file.s
    +

    "file.s" is a command line argument which tells the program rm to remove the file "file.s".

    +

    Programming languages such as C, C++ and Java allow a program to interpret the command line arguments by handling them as string parameters in the main function.

    +

    A command line option or simply option (also known as a command line parameter, flag, or a switch) is an indication by a user that a computer program should change its default output.

    +

    Long options are introduced via "--", and are typically whole words. For example, ls --long --classify --all. Arguments to long options are provided with "=", as ls --block-size=1024. Some Unix programs use long options with single dashes, for example MPlayer as in mplayer -nosound.

    +

    Linux also uses "--" to terminate option lists. For example, an attempt to delete a file called -file1 by using rm -file1 may produce an error, since rm may interpret -file1 as a command line switch. Using rm -- -file1 removes ambiguity.

    +
    +
    +

    +7. Basic Text Processing

    +
    +

    +7.1. head

    +

    head is a program on Unix and Unix-like systems used to display the first few lines of a text file or piped data. The command syntax is:

    +
    $ head [options] <file_name>
    +

    By default, head will print the first 10 lines of its input to the standard output. The number of lines printed may be changed with a command line option. The following example shows the first 20 lines of filename:

    +
    $ head -n 20 filename
    +

    This displays the first 5 lines of all files starting with foo:

    +
    $ head -n 5 foo*
    +

    Some versions omit the n and just let you say -5.

    +
    +

    +7.1.1. Flags

    +
    -c <x number of bytes> Copy first x number of bytes.
    +

    Other options: sed

    +

    Many early versions of Unix did not have this command, and so documentation and books had sed do this job:

    +
    sed 5q foo
    +

    This says to print every line (implicit), and quit after the fifth.

    +
    +
    +
    +

    +7.2. tail

    +

    tail is a program on Unix and Unix-like systems used to display the last few lines of a text file or piped data.

    +

    The command-syntax is:

    +
    $ tail [options] <file_name>
    +

    By default, tail will print the last 10 lines of its input to the standard output. With command line options the number of lines printed and the printing units (lines, blocks or bytes) may be changed. The following example shows the last 20 lines of filename:

    +
    $ tail -n 20 filename
    +

    This example shows the last 15 bytes of all files starting with foo:

    +
    $ tail -c 15 foo*
    +

    This example shows all lines of filename from the second line onwards:

    +
    $ tail -n +2 filename
    +

    Using an older syntax (still used in Sun Solaris as the -n option is not supported), the last 20 lines and the last 50 bytes of filename can be shown with the following command:

    +
    $ tail -20 filename
    +$ tail -50c filename
    +

    However this syntax is now obsolete and does not conform with the POSIX 1003.1-2001 standard. Even if still supported in current versions, when used with other options (like -f, see below), these switches could not work at all.

    +
    +

    +7.2.1. File monitoring

    +

    tail has a special command line option -f (follow) that allows a file to be monitored. Instead of displaying the last few lines and exiting, tail displays the lines and then monitors the file. As new lines are added to the file by another process, tail updates the display. This is particularly useful for monitoring log files. The following command will display the last 10 lines of messages and append new lines to the display as new lines are added to messages:

    +
    $ tail -f /var/adm/messages
    +

    To interrupt tail while it is monitoring, break-in with Ctrl+C. This command can be run "in the background" with &, see job control.

    +

    If you have a command's result to monitor, you can use the watch command.

    +
    +
    +
    +

    +7.3. cut

    +

    In computing, cut is a Unix command line utility which is used to extract sections from each line of input — usually from a file.

    +

    Extraction of line segments can typically be done by bytes (-b), characters (-c), or fields (-f) separated by a delimiter (-d — the tab character by default). A range must be provided in each case which consists of one of N, N-M, N- (N to the end of the line), or -M (beginning of the line to M), where N and M are counted from 1 (there is no zeroth value). Since version 6, an error is thrown if you include a zeroth value. Prior to this the value was ignored and assumed to be 1.

    +

    Assuming a file named file containing the lines:

    +
    foo:bar:baz:qux:quux
    +one:two:three:four:five:six:seven
    +alpha:beta:gamma:delta:epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +

    To output the fourth through tenth characters of each line:

    +
    $ cut -c 4-10 file
    +

    This gives the output:

    +
    :bar:ba
    +:two:th
    +ha:beta
    +

    To output the fifth field through the end of the line of each line using the colon character as the field delimiter:

    +
    $ cut -d : -f 5- file
    +

    This gives the output:

    +
    quux
    +five:six:seven
    +epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +
    +
    +

    +7.4. paste

    +

    paste is a Unix command line utility which is used to join files horizontally (parallel merging) by outputting lines consisting of the sequentially corresponding lines of each file specified, separated by tabs, to the standard output. It is effectively the horizontal equivalent to the utility cat command which operates on the vertical plane of two or more files.

    +

    To paste several columns of data together into the file www from files who, where, and when:

    +
    $ paste who where when > www
    +

    If the files contain:

    +++++
    +

    This creates the file named www containing:

    +
    Batman            GothamCity       January 3
    +Trillian          Andromeda        February 4
    +Jeeves            London           March 19
    +
    +
    +
    +

    +8. Shell Meta Characters

    +

    Unix recognizes certain special characters, called "meta characters," as command directives. The shell meta characters are recognized anywhere they appear in the command line, even if they are not surrounded by blank space. For that reason, it is safest to only use the characters A-Z, a-z, 0-9, and the period, dash, and underscore characters when naming files and directories on Unix. If your file or directory has a shell meta character in the name, you will find it difficult to use the name in a shell command.

    +

    The shell meta characters include:

    +

    / < > ! $ % ^ & * | { } [ ] " ' ` ~ ;

    +

    Different shells may differ in the meta characters recognized.

    +

    As an example,

    +
    $ ls file.*
    +

    run on a directory containing the files file, file.c, file.lst, and myfile would list the files file.c and file.lst. However,:

    +
    $ ls file.?
    +

    run on the same directory would only list file.c because the ? only matches one character, no more, no less. This can save you a great deal of typing time. For example, if there is a file called california_cornish_hens_with_wild_rice and no other files whose names begin with 'c', you could view the file without typing the whole name by typing this:

    +
    $ more c*
    +

    because the c* matches that long file name.

    +

    Filenames containing metacharacters can pose many problems and should never be intentionally created. If you do find that you've created a file with metacharacters, and you would like to remove it, you have three options. You may use wildcards to match metacharacter, use the to directly enter the filename, or put the command in double quotes (except in the case of double quotes within the file name, these must be captured with one of the first two methods). For example, deleting a file named <title_reference>"``*`|more</title_reference>"` can be accomplished with:

    +
    $ rm ??more
    +

    or:

    +
    $ rm $\backslash$*$\backslash$|more
    +

    or:

    +
    $ rm ''*|more''
    +
    +
    +

    +9. Looking At Files

    +
    +

    +9.1. cat

    +

    The cat command is a standard Unix program used to concatenate and display files. The name is from "catenate", a synonym of concatenate.

    +

    The Single Unix Specification specifies the behavior that the contents of each of the files given in sequence as arguments will be written to the standard output in the same sequence, and mandates one option, -u, where each byte is printed as it is read.

    +

    If the filename is specified as -, then cat will read from standard input at that point in the sequence. If no files are specified, cat will read from standard input entered.

    +
    +

    +9.1.1. Jargon File Definition

    +

    The Jargon File version 4.4.7 lists this as the definition of cat:

    +
    1. To spew an entire file to the screen or some other output sink without
    +     pause (syn. blast).
    +
    +2. By extension, to dump large amounts of data at an unprepared target or
    +     with no intention of browsing it carefully. Usage: considered silly.
    +     Rare outside Unix sites. See also dd, BLT.
    +
    +     Among Unix fans, *cat(1)* is considered an excellent example of
    +     user-interface design, because it delivers the file contents without
    +     such verbosity as spacing or headers between the files, and because
    +     it does not require the files to consist of lines of text, but works
    +     with any sort of data.
    +
    +     Among Unix critics, *cat(1)* is considered the canonical example of
    +     bad user-interface design, because of its woefully unobvious name.
    +     It is far more often used to blast a single file to standard output
    +     than to concatenate two or more files. The name cat for the former
    +     operation is just as unintuitive as, say, LISP's cdr.
    +
    +     Of such oppositions are holy wars made...
    +
    +
    +

    +9.1.2. Useless Use of 'cat'

    +

    UUOC (from comp.unix.shell on Usenet) stands for “Useless Use of cat”. As it is observed on comp.unix.shell, “The purpose of cat is to concatenate (or 'catenate') files. If it's only one file, concatenating it with nothing at all is a waste of time, and costs you a process.”

    +

    Nevertheless one sees people doing:

    +
    $ cat file | some_command and its args ...
    +

    instead of the equivalent and cheaper:

    +
    <file some_command and its args ...
    +

    or (equivalently and more classically):

    +
    some_command and its args ... <file
    +

    Since 1995, occasional awards for UUOC have been given out. The activity of fixing instances of UUOC is sometimes called 'demoggification'.

    +

    Amongst many, it is still considered safer to use cat for such cases given that the < and > keys are next to each other in many popular keyboard mappings. While the risk might be low, the impact of using > instead of < can be high and prohibitive.

    +
    +
    +

    +9.1.3. zcat

    +

    zcat is a Unix program similar to cat, that decompresses individual files and concatenates them to standard output. Traditionally zcat operated on files compressed by compress but today it is usually able to operate on gzip or even bzip2 archives. On such systems, it is equivalent to gunzip -c

    +
    +
    +
    +

    +9.2. more

    +

    In computing, more is a command to view (but not modify) the contents of a text file one screen at a time (terminal pager). It is available on Unix and Unix-like systems, DOS, OS/2 and Microsoft Windows. Programs of this sort are called pagers.

    +
    +

    +9.2.1. Usage

    +

    The command-syntax is:

    +
    $ more [options] [file_name]
    +

    If no file name is provided, more looks for input from stdin.

    +

    Once more has obtained input, it displays as much as can fit on the current screen and waits for user input to advance, with the exception that a form feed (^L) will also cause more to wait at that line, regardless of the amount of text on the screen. In the lower-left corner of the screen is displayed the text "--More--" and a percentage, representing the percent of the file that more has paged through. (This percentage includes the text displayed on the current screen.) When more reaches the end of a file (100%) it exits. The most common methods of navigating through a file are Enter, which advances the output by one line, and Space, which advances the output by one screen.

    +

    There are also other commands that can be used while navigating through the document; consult more's man page for more details.

    +

    Options are typically entered before the file name, but can also be entered in the environment variable $MORE. Options entered in the actual command line will override those entered in the $MORE environment variable. Available options may vary between Unix systems.

    +
    +
    +
    +

    +9.3. less

    +

    less is a terminal pager program on Unix, Windows and Unix-like systems used to view (but not change) the contents of a text file one screen at a time. It is similar to more, but has the extended capability of allowing both forward and backward navigation through the file. Unlike most Unix text editors/viewers, less does not need to read the entire file before starting, resulting in faster load times with large files.

    +
    +

    +9.3.1. Usage

    +

    less can be invoked with options to change its behaviour, for example, the number of lines to display on the screen. A few options vary depending on the operating system. While less is displaying the file, various commands can be used to navigate through the file. These commands are based on those used by both more and vi. It is also possible to search for character patterns in the file.

    +

    By default, less displays the contents of the file to the standard output (one screen at a time). If the file name argument is omitted, it displays the contents from standard input (usually the output of another command through a pipe). If the output is redirected to anything other than a terminal, for example a pipe to another command, less behaves like cat.

    +

    The command-syntax is:

    +
    $ less [options] file_name
    +
    +
    +

    +9.3.2. Frequently Used Options

    +<block_quote>
      +
    • -g: Highlights just the current match of any searched string.

    • +
    • -I: Case-insensitive searches.

    • +
    • -M: Shows more detailed prompt, including file position.

    • +
    • -N: Shows line numbers (useful for source code viewing).

    • +
    • -S: Disables line wrap ("chop long lines"). Long lines can be seen by side scrolling.

    • +
    • -?: Shows help.

    • +
    </block_quote>
    +
    +
    +

    +9.3.3. Frequently Used Commands

    +<block_quote>
      +
    • [Arrows]/[Page Up]/[Page Down]/[Home]/[End]: Navigation.

    • +
    • [Space bar]: Next page.

    • +
    • b: Previous page.

    • +
    • ng: Jump to line number n. Default is the start of the file.

    • +
    • nG: Jump to line number n. Default is the end of the file.

    • +
    • /pattern: Search for pattern. Regular expressions can be used.

    • +
    • '^ or g: Go to start of file.

    • +
    • '$ or G: Go to end of file.

    • +
    • s: Save current content (got from another program like grep) in a file.

    • +
    • =: File information.

    • +
    • h: Help.

    • +
    • q: Quit.

    • +
    </block_quote>
    +
    +
    +

    +9.3.4. Examples

    +
    $ less -M readme.txt                     #Read "readme.txt."
    +$ less +F /var/log/mail.log              #Follow mode for log
    +$ file * | less                          #Easier file analysis.
    +$ grep -i void *.c | less -I -p void     #Case insensitive search                                                         for "void" in all .c files
    +
    +
    +
    +
    +

    +10. Directory Structure

    +

    In the File Hierarchy Standard (FHS) all files and directories appear under the root directory "/", even if they are stored on different physical devices. Note however that some of these directories may or may not be present on a Unix system depending on whether certain subsystems, such as the X Window System, are installed.

    +

    The majority of these directories exist in all UNIX operating systems and are generally used in much the same way; however, the descriptions here are those used specifically for the FHS, and are not considered authoritative for platforms other than Linux.

    ++++
    +
    +

    +10.1. man hier

    +

    This is the manual page on the UNIX filesystem. The syntax for this is:

    +
    $ man hier
    +
    +
    +

    +10.2. ls -l

    +

    Shows you huge amounts of information (permissions, owners, size, and when last modified) for folders and files. The syntax is

    +
    $ ls -l
    +

    This can be done after entering the required directory.

    +
    +
    +
    +

    +11. Permissions and Ownership

    +
    +

    +11.1. chmod

    +

    The chmod command (abbreviated from 'change mode') is a shell command and C language function in Unix and Unix-like environments. When executed, it can change file system modes of files and directories. The modes include permissions and special modes.A chmod command first appeared in AT&T Unix version 1, and is still used today on Unix-like machines.

    +
    +

    +11.1.1. Usage

    +

    The chmod command options are specified like this:

    +
    $ chmod [options] mode[,mode] file1 [file2 ...]
    +

    To view what the permissions currently are, type:

    +
    $ ls -l file
    +
    +
    +

    +11.1.2. Command line options

    +

    The chmod command has a number of command line options that affect its behavior. The most common options are:

    +<block_quote>
      +
    • -R: Changes the modes of directories and files recursively

    • +
    • -v: Verbose mode; lists all files as they are being processed

    • +
    </block_quote>
    +
    +11.1.2.1. Symbolic modes
    +

    To the chmod utility, all permissions and special modes are represented by its mode parameter. One way to adjust the mode of files or directories is to specify a symbolic mode. The symbolic mode is composed of three components, which are combined to form a single string of text:

    +
    $ chmod [references][operator][modes] file1 ...
    +

    The references (or classes) are used to distinguish the users to whom the permissions apply. If no references are specified it defaults to “all” but modifies only the permissions allowed by the umask. The references are represented by one or more of the following letters:

    +++++
    +

    The chmod program uses an operator to specify how the modes of a file should be adjusted. The following operators are accepted:

    ++++
    +

    The modes indicate which permissions are to be granted or taken away from the specified classes. There are three basic modes which correspond to the basic permissions:

    +++++
    +

    The combination of these three components produces a string that is understood by the chmod command. Multiple changes can be specified by separating multiple symbolic modes with commas.

    +
    +
    +
    +11.1.2.2. Symbolic examples
    +

    Add the 'read' and 'write' permissions to the 'user' and 'group' classes of a directory:

    +
    $ chmod ug+rw mydir
    +$ ls -ld mydir
    +drw-rw----   2 starwars  yoda  96 Dec 8 12:53 mydir
    +

    For a file, remove write permissions for all classes:

    +
    $ chmod a-w myfile
    +$ ls -l myfile
    +-r-xr-xr-x   2 starwars  yoda 96 Dec 8 12:53 myfile
    +

    Set the permissions for the u*ser and the *g*roup to read and execute only (no write permission) on *mydir.

    +
    $ chmod ug=rx mydir
    +$ ls -ld mydir
    +dr-xr-x---   2 starwars  yoda 96 Dec 8 12:53 mydir
    +
    +
    +
    +11.1.2.3. Octal numbers
    +

    The chmod command also accepts three and four-digit octal numbers representing modes. Using a three-digit octal number to set the modes of a file named myfile :

    +
    $ chmod 664 myfile
    +$ ls -l myfile
    +-rw-rw-r--  1   57 Jul  3 10:13  myfile
    +

    Since the setuid, setgid and sticky bits are not set, this is equivalent to:

    +
    $ chmod 0664 myfile
    +
    +
    +
    +11.1.2.4. Special modes
    +

    The chmod command is also capable of changing the additional permissions or special modes of a file or directory. The symbolic modes use s to represent the setuid and setgid modes, and t to represent the sticky mode. The modes are only applied to the appropriate classes, regardless of whether or not other classes are specified.

    +

    Most operating systems support the specification of special modes using octal modes, but some do not. On these systems, only the symbolic modes can be used.

    +
    +
    +
    +
    +
    +

    +12. Redirection and Piping

    +

    In computing, redirection is a function common to most command-line interpreters, including the various Unix shells that can redirect standard streams to user-specified locations.

    +

    Programs do redirection with the dup2(2) system call, or its less-flexible but higher-level stdio analogues, freopen(3) and popen(3).

    +
    +

    +12.1. Redirecting standard input and standard output

    +

    Redirection is usually implemented by placing certain characters between commands. Typically, the syntax of these characters is as follows:

    +
    $ command1 > file1
    +

    executes command1, placing the output in file1. Note that this will truncate any existing data in file1. To append output to the end of the file, use the >> operator.:

    +
    $ command1 < file1
    +

    executes command1, using file1 as the source of input (as opposed to the keyboard).:

    +
    $ command1 < infile > outfile
    +

    combines the two capabilities: command1 reads from infile and writes to outfile

    +
    +
    +

    +12.2. Piping

    +

    Programs can be run together such that one program reads the output from another with no need for an explicit intermediate file: +A pipeline of three programs run on a text terminal:

    +
    $ command1 | command2
    +

    executes command1, using its output as the input for command2 (commonly called piping, since the "|" character is known as a "pipe").

    +

    This is equivalent to using two redirects and a temporary file:

    +
    $ command1 > tempfile
    +$ command2 < tempfile
    +$ rm tempfile
    +

    A good example for command piping is combining echo with another command to achieve something interactive in a non-interactive shell, e.g.:

    +
    $ echo -e "user\npass" | ftp localhost
    +

    This runs the ftp client with input user, press return, then pass.

    +
    +
    +

    +12.3. Redirecting to and from the standard file handles

    +

    In Unix shells derived from the original Bourne shell, the first two actions can be further modified by placing a number (the file descriptor) immediately before the character; this will affect which stream is used for the redirection. The Unix standard I/O streams are:

    +++++
    +

    For example:

    +
    $ command1 2> file1
    +

    executes command1, directing the standard error stream to file1.

    +

    In shells derived from csh (the C shell), the syntax instead appends the & character to the redirect characters, thus achieving a similar result.

    +

    Another useful capability is to redirect one standard file handle to another. The most popular variation is to merge standard error into standard output so error messages can be processed together with (or alternately to) the usual output. Example:

    +
    $ find / -name .profile > results 2>&1
    +

    will try to find all files named .profile. Executed without redirection, it will output hits to stdout and errors (e.g. for lack of privilege to traverse protected directories) to stderr. If standard output is directed to file results, error messages appear on the console. To see both hits and error messages in file results, merge stderr (handle 2) into stdout (handle 1) using 2>&1 .

    +

    It's possible use 2>&1 before ">" but it doesn't work. In fact, when the interpreter reads 2>&1, it doesn't know yet where standard output is redirected and then standard error isn't merged.

    +

    If the merged output is to be piped into another program, the file merge sequence 2>&1 must precede the pipe symbol, thus:

    +
    $ find / -name .profile 2>&1 | less
    +

    A simplified form of the command:

    +
    $ command > file 2>&1
    +

    is:

    +
    $ command &>file
    +

    or:

    +
    $command >&file
    +
    +
    +

    +12.4. Chained pipelines

    +

    The redirection and piping tokens can be chained together to create complex commands. For example:

    +
    $ ls | grep '\.sh' | sort > shlist
    +

    lists the contents of the current directory, where this output is filtered to only contain lines which contain .sh, sort this resultant output lexicographically, and place the final output in shlist. This type of construction is used very commonly in shell scripts and batch files.

    +
    +
    +

    +12.5. Redirect to multiple outputs

    +

    The standard command tee can redirect output from a command to several destinations.

    +
    $ ls -lrt | tee xyz
    +

    This directs the file list output to both standard output as well as to the file xyz.

    +
    +
    +
    +

    +13. More Text Processing

    +
    +

    +13.1. grep

    +

    grep is a command line text search utility originally written for Unix. The name is taken from the first letters in global / regular expression / print, a series of instructions for the ed text editor. The grep command searches files or standard input globally for lines matching a given regular expression, and prints them to the program's standard output.

    +
    +

    +13.1.1. Usage

    +

    This is an example of a common grep usage:

    +
    $ grep apple fruitlist.txt
    +

    In this case, grep prints all lines containing 'apple' from the file fruitlist.txt, regardless of word boundaries; therefore lines containing 'pineapple' or 'apples' are also printed. The grep command is case sensitive by default, so this example's output does not include lines containing 'Apple' (with a capital A) unless they also contain 'apple'.

    +

    Like most Unix commands, grep accepts command line arguments to change this and many other behaviors. For example:

    +
    $ grep -i apple fruitlist.txt
    +

    This prints all lines containing 'apple' regardless of capitalization. The '-i' argument tells grep to be case insensitive, or to ignore case.

    +

    To print all lines containing 'apple' as a word ('pineapple' and 'apples' will not match):

    +
    $ grep -w apple fruitlist.txt
    +

    Regular expressions can be used to match more complicated queries.

    +
    +
    +13.1.1.1. Variations
    +

    There are countless implementations and derivatives of grep available for many operating systems. Early variants of grep included egrep and fgrep. The former applies an extended regular expression syntax that was added to Unix after Ken Thompson's original regular expression implementation. The latter searches for any of a list of 'fixed' strings using the Aho-Corasick algorithm. These variants are embodied in most modern grep implementations as command-line switches (and standardized as -E and -F in POSIX). In such combined implementations, grep may also behave differently depending on the name by which it is invoked, allowing fgrep, egrep, and grep to be links to the same program.

    +

    pcregrep is an implementation of grep that uses Perl regular expression syntax.

    +

    Other commands contain the word 'grep' to indicate that they search (usually for regular expression matches). The pgrep utility, for instance, displays the processes whose names match a given regular expression.

    +
    +
    +
    +
    +

    +13.2. tr

    +

    tr (abbreviated from translate or transliterate) is a command in Unix-like operating systems.

    +

    When executed, the program reads from the standard input and writes to the standard output. It takes as parameters two sets of characters, and replaces occurrences of the characters in the first set with the corresponding elements from the other set. For example,

    +
    $ tr 'abcd' 'jkmn'
    +

    maps 'a' to 'j', 'b' to 'k', 'c' to 'm', and 'd' to 'n'.

    +

    Sets of characters may be abbreviated by using character ranges. The previous example could be written:

    +
    $ tr 'a-d' 'jkmn'
    +

    In POSIX compliant versions of tr the set represented by a character range depends on the locale's collating order, so it is safer to avoid character ranges in scripts that might be executed in a locale different from that in which they were written. Ranges can often be replaced with POSIX character sets such as [:alpha:].

    +

    The -c flag complements the first set of characters.

    +
    $ tr -cd '[:alnum:]'
    +

    therefore removes all non-alphanumeric characters.

    +

    The -s flag causes tr to compress sequences of identical adjacent characters in its output to a single token. For example,

    +
    $ tr -s '\n' '\n'
    +

    replaces sequences of one or more newline characters with a single newline.

    +

    The -d flag causes tr to delete all tokens of the specified set of characters from its input. In this case, only a single character set argument is used. The following command removes carriage return characters, thereby converting a file in DOS/Windows format to one in Unix format.

    +
    $ tr -d '\r'
    +

    Most versions of tr, including GNU tr and classic Unix tr, operate on single byte characters and are not Unicode compliant. An exception is the Heirloom Toolchest implementation, which provides basic Unicode support.

    +

    Ruby and Perl also have an internal tr operator, which operates analogously. Tcl's string map command is more general in that it maps strings to strings while tr maps characters to characters.

    +
    +
    +
    +

    +14. Elementary Regex

    +

    In computing, regular expressions provide a concise and flexible means for identifying strings of text of interest, such as particular characters, words, or patterns of characters. A regular expression (often shortened to regex or regexp) is written in a formal language that can be interpreted by a regular expression processor, a program that either serves as a parser generator or examines text and identifies parts that match the provided specification.

    +

    Regular expressions are used by many text editors, utilities, and programming languages to search and manipulate text based on patterns. For example, Perl, Ruby and Tcl have a powerful regular expression engine built directly into their syntax. Several utilities provided by Unix distributions—including the editor ed and the filter grep — were the first to popularize the concept of regular expressions.

    +

    Traditional Unix regular expression syntax followed common conventions but often differed from tool to tool. The IEEE POSIX Basic Regular Expressions (BRE) standard (released alongside an alternative flavor called Extended Regular Expressions or ERE) was designed mostly for backward compatibility with the traditional (Simple Regular Expression) syntax but provided a common standard which has since been adopted as the default syntax of many Unix regular expression tools, though there is often some variation or additional features. Many such tools also provide support for ERE syntax with command line arguments.

    +

    In the BRE syntax, most characters are treated as literals — they match only themselves (i.e., a matches "a"). The exceptions, listed below, are called metacharacters or metasequences.

    ++++
    +
    +

    +14.1. Lazy quantification

    +

    The standard quantifiers in regular expressions are greedy, meaning they match as much as they can, only giving back as necessary to match the remainder of the regex. For example, someone new to regexes wishing to find the first instance of an item between < and > symbols in this example:

    +
    Another whale explosion occurred on <January 26>, <2004>.
    +

    ...would likely come up with the pattern <.*>, or similar. However, this pattern will actually return "<January 26>, <2004>" instead of the "<January 26>" which might be expected, because the <title_reference>*</title_reference> quantifier is greedy — it will consume as many characters as possible from the input, and "January 26>, <2004" has more characters than "January 26".

    +

    Though this problem can be avoided in a number of ways (e.g., by specifying the text that is not to be matched: <[^>]*>), modern regular expression tools allow a quantifier to be specified as lazy (also known as non-greedy, reluctant, minimal, or ungreedy) by putting a question mark after the quantifier (e.g., <.*?>), or by using a modifier which reverses the greediness of quantifiers (though changing the meaning of the standard quantifiers can be confusing). By using a lazy quantifier, the expression tries the minimal match first. Though in the previous example lazy matching is used to select one of many matching results, in some cases it can also be used to improve performance when greedy matching would require more backtracking.

    +
    +
    +
    +

    +15. One Liners

    +

    A one-liner is textual input to the command-line of an operating system shell that performs some function in just one line of input.

    +

    The one liner can be

    +<block_quote>
      +
    1. An expression written in the language of the shell.

    2. +
    3. The invocation of an interpreter together with program source for the interpreter to run.

    4. +
    5. The invocation of a compiler together with source to compile and +instructions for executing the compiled program.

    6. +
    </block_quote>

    Certain dynamic scripting languages such as AWK, sed, and perl have traditionally been adept at expressing one-liners. Specialist shell interpreters such as these Unix shells or the Windows PowerShell, allow for the construction of powerful one-liners.

    +

    The use of the phrase one-liner has been widened to also include program-source for any language that does something useful in one line.

    +

    The word One-liner has two references in the index of the book The AWK Programming Language (the book is often referred to by the abbreviation TAPL). It explains the programming language AWK, which is part of the Unix operating system. The authors explain the birth of the One-liner paradigm with their daily work on early Unix machines:

    +
    “The 1977 version had only a few built-in variables and predefined functions. It was designed for writing short programs [...] Our model was that an invocation would be one or two lines long, typed in and used immediately. Defaults were chosen to match this style [...] We, being the authors, knew how the language was supposed to be used, and so we only wrote one-liners.”
    +

    Notice that this original definition of a One-liner implies immediate execution of the program without any compilation. So, in a strict sense, only source code for interpreted languages qualifies as a One-liner. But this strict understanding of a One-liner was broadened in 1985 when the IOCCC introduced the category of Best One Liner for C, which is a compiled language.

    +

    The TAPL book contains 20 examples of One-liners (A Handful of Useful awk One-Liners) at the end of the book's first chapter.

    +

    Here are the first few of them:

    +<block_quote>
      +
    1. +

      Print the total number of input lines:

      +

      END { print NR }

      +
    2. +
    3. +

      Print the tenth input line:

      +

      NR == 10

      +
    4. +
    5. +

      Print the last field of every input line:

      +

      { print $NF }

      +
    6. +
    </block_quote>

    One-liners are also used to show off the differential expressive power of programming languages. Frequently, one-liners are used to demonstrate programming ability. Contests are often held to see who can create the most exceptional one-liner.

    +

    The following example is a C program (a winning entry in the "Best one-liner" category of the IOCCC, here split to two lines for presentation).:

    +
    main(int c,char**v){return!m(v[1],v[2]);}m(char*s,char*t){return
    +*t-42?*s?63==*t|*s==*t&&m(s+1,t+1):!*t:m(s,t+1)||*s&&m(s+1,t);}
    +

    This one-liner program is a glob pattern matcher. It understands the glob characters '*' meaning 'zero or more characters' and '?' meaning exactly one character, just like most Unix shells.

    +

    Run it with two args, the string and the glob pattern. The exit status is 0 (shell true) when the pattern matches, 1 otherwise. The glob pattern must match the whole string, so you may want to use * at the beginning and end of the pattern if you are looking for something in the middle. Examples:

    +
    $ prog foo 'f??'; echo $?
    +
    +$ prog 'best short program' '??st*o**p?*'; echo $?
    +

    Here is a one line shell script to show directories:

    +
    $ ls -R | grep ":$" | sed -e 's/:$//' -e 's/[^-][^\/]*\//--/g' -e 's/^/   /' -e 's/-/|/'
    +
    +
    +
    + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/chn10session4.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/chn10session4.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1139 @@ + + + +Session 4  + + + + + + + +
    +
    + +
    +
    +

    +
    +
    + +
    +

    +1. Introducing Linux

    +

    (Attribution : A significant chunk of the content under this section is based on data from Wikipedia and the Linux Documentation Project)

    +

    Linux (usually pronounced ˈlɪnəks') is a generic term referring to Unix-like computer operating systems based on the Linux kernel, where a kernel is the intermediate layer between the hardware and the applications. The kernel is, on an abstract level, the core of (most) operating systems, that manages the various system resources. The development of the Linux OS is considered the basis for Free and Open Source Software (FOSS) collaboration since typically the underlying source code can be used, modified freely, and redistributed by anyone under the terms of the GNU (a recursive acronym for "GNU's Not Unix!") Global Public License (GPL) and other free software licences. This freedom to access and reuse various components of a system, is one of the primary reasons for the popularity of Linux.

    +

    Linux is installed on a variety of computer hardware, that include mobile phones, embedded devices and supercomputers, but is infamous for its use in servers.

    +

    The name "Linux" comes from the Linux kernel, originally written in 1991 by Linus Torvalds. The rest of the system usually comprises components such as the Apache HTTP Server, the X Window System, the GNOME and KDE desktop environments, and utilities and libraries from the GNU Project (announced in 1983 by Richard Stallman). Commonly-used applications with desktop Linux systems include the Mozilla Firefox web-browser and the OpenOffice.org office application suite. The GNU contribution is the basis for the Free Software Foundation's preferred name GNU/Linux. The kernel's mascot is a penguin named "Tux". Mozilla Firefox and OpenOffice.org are open-source projects which can be run on most Operating Systems, including proprietary ones.

    +
    +

    +1.1. Historical Background

    +
    +

    +1.1.1. Events leading to the creation

    +
      +
    • The Unix operating system was developed in the 1960s and released for public use in 1970. Its accessibility and portability caused it to be widely adopted, copied and modified by academic institutions and businesses. Its design became influential to authors of other systems. Other free operating systems include the Berkeley Software Distribution (BSD), developed at the University of California at Berkeley, and MINIX which was released by Andrew S. Tanenbaum. The development and adoption of BSD and MINIX were limited due to various reasons, and this lack of a widely-adopted and free kernel triggered Linus Torvalds into starting his project.

    • +
    • In 1983, Richard Stallman started the GNU project with the goal of creating a free UNIX-like operating system. As part of this work, he wrote the GNU General Public License (GPL). By the early 1990s there was almost enough available software to create a full operating system. However, the GNU kernel, called Hurd, failed to attract enough attention from developers leaving GNU incomplete.

    • +
    +
    +
    +

    +1.1.2. The Creation of Linux

    +

    In 1991, Linus Torvalds began a project at the University of Helsinki that later became the Linux kernel. It was initially a terminal (command-line) emulator, which Torvalds used to access the large UNIX servers of the university. He wrote the program targeting just the hardware he was using and independent of an operating system because he wanted to use the functions of his computer with an 80386 processor. Development was done on Minix using the GNU C compiler. This application is still the main choice for compiling Linux today (although the code can be built with other compilers, such as the Intel C Compiler).

    +

    Torvalds continues to direct the development of the kernel. Stallman heads the Free Software Foundation, which in turn supports the GNU components. Finally, individuals and corporations develop third-party non-GNU components, which constitute a vast body of work and including kernel modules, and user applications and libraries. Linux vendors and communities combine and distribute the kernel, GNU components, and non-GNU components, with additional package management software in the form of Linux distributions.

    +
    +
    +
    +

    +1.2. Design and Implications

    +

    A Linux-based system is a modular Unix-like operating system, deriving much of its basic design from principles established in Unix earlier. Such a system uses a monolithic kernel, called the Linux kernel, which handles process control, networking, and peripheral and file system access. Device drivers are integrated directly with the kernel. Separate projects that interface with the kernel provide much of the system's higher-level functionality. The GNU userland is an important part of most Linux-based systems, providing the most common implementation of the C library, a popular shell, and many of the common Unix tools which carry out many basic operating system tasks. The graphical user interface (or GUI) used by most Linux systems is based on the "X Window System".

    +
    +

    +1.2.1. User Interface

    +

    Users can control a Linux-based system through a command line interface (or CLI), a graphical user interface (or GUI), or through controls attached to the associated hardware (this is common for embedded systems). For desktop systems, the default mode is usually the GUI. On desktop machines, "KDE", "GNOME" and "Xfce" are the most popular user interfaces,though a variety of additional user interfaces exist. Most popular user interfaces run on top of the "X Window System" (or X), which enables a graphical application running on one machine to be displayed and controlled from another in a network.

    +

    A Linux system also provides a CLI of some sort through a shell, which is the traditional way of interacting with a Unix system. A Linux distribution specialized for servers may use the CLI as its only interface. A “headless system” (system run without even a monitor) can be controlled by the command line via a remote-control protocol such as SSH or telnet. The CLI is particularly suited for automation of repetitive or delayed tasks, and provides very simple inter-process communication. A graphical terminal emulator program is often used to access the CLI from a Linux desktop.

    +
    +
    +

    +1.2.2. Development

    +

    The primary difference between Linux and many other popular contemporary operating systems is that the Linux kernel and other components are free and open source software. Linux is not the only such operating system, although it is by far the most widely used. Some free and open source software licenses are based on the principle of "copyleft", a kind of reciprocity: any work derived from a copyleft piece of software must also be copyleft itself. The most common free software license, the GNU GPL, is a form of copyleft, and is used for the Linux kernel and many of the components from the GNU project.

    +

    Linux based distributions are intended by developers for interoperability with other operating systems and established computing standards. Linux systems adhere to POSIX, SUS, ISO and ANSI standards where possible, although to date only one Linux distribution has been POSIX.1 certified, Linux-FT.Free software projects, although developed in a collaborative fashion, are often produced independently of each other. The fact that the software licenses explicitly permit redistribution, however, provides a basis for larger scale projects that collect the software produced by stand-alone projects and make it available all at once in the form of a Linux distribution.

    +

    A Linux distribution, commonly called a "distro", is a project that manages a remote collection of system software and application software packages available for download and installation through a network connection. This allows the user to adapt the operating system to his/her specific needs. Distributions are maintained by individuals, loose-knit teams, volunteer organizations, and commercial entities. A distribution can be installed using a CD that contains distribution-specific software for initial system installation and configuration. A package manager such as Synaptic or YAST allows later package upgrades and installations. A distribution is responsible for the default configuration of the installed Linux kernel, general system security, and more generally integration of the different software packages into a coherent whole.

    +
    +
    +

    +1.2.3. Community

    +

    A distribution is largely driven by its developer and user communities. Some vendors develop and fund their distributions on a volunteer basis. Examples include Debian and the Debian-based, Ubuntu. Others maintain a community version of their commercial distributions, as Red Hat does with Fedora.

    +

    In many cities and regions, local associations known as Linux Users Groups (LUGs) seek to promote their preferred distribution and by extension free software. They hold meetings and provide free demonstrations, training, technical support, and operating system installation to new users. Many Internet communities also provide support to Linux users and developers. Most distributions and free software / open source projects have IRC (Internet Relay Chat) chatrooms or newsgroups. Online forums are another means for support. Linux distributions host mailing lists; commonly there will be a specific topic such as usage or development for a given list. All these can be found simply by running an appropriate search on Google.

    +

    Although Linux distributions are generally available without charge, several large corporations sell, support, and contribute to the development of the components of the system and of free software. These include Dell, IBM, HP, Oracle, Sun Microsystems, Novell, Nokia. A number of corporations, notably Red Hat, have built their entire business around Linux distributions.

    +
    +
    +

    +1.2.4. Can I make a profit out of running a business involving Linux?

    +

    The answer is, "Yes!". The free software licenses, on which the various software packages of a distribution built on the Linux kernel are based, explicitly accommodate and encourage commercialization; the relationship between a Linux distribution as a whole and individual vendors may be seen as symbiotic. One common business model of commercial suppliers is charging for support, especially for business users. A number of companies also offer a specialized business version of their distribution, which adds proprietary support packages and tools to administer higher numbers of installations or to simplify administrative tasks. Another business model is to give away the software in order to sell hardware. Examples of corporations that are extensively (and sometimes exclusively) open-source and Linux-powered , with successful revenue generation models involving these, are Google, SUN, Mozilla, etc.

    +
    +
    +

    +1.2.5. Programming on Linux

    +

    Most Linux distributions support dozens of programming languages. The most common collection of utilities for building both Linux applications and operating system programs is found within the GNU toolchain, which includes the GNU Compiler Collection (GCC) and the GNU build system. Amongst others, GCC provides compilers for Ada, C, C++, Java, and Fortran. The Linux kernel itself is written to be compiled with GCC. Proprietary compilers for Linux include the Intel C++ Compiler, Sun Studio, and IBM XL C/C++ Compiler.

    +

    Most distributions also include support for PHP, Perl, Ruby, Python and other dynamic languages. Examples of languages that are less common, but still supported, are C# via the Mono project, sponsored by Novell, and Scheme. A number of Java Virtual Machines and development kits run on Linux, including the original Sun Microsystems JVM (HotSpot), and IBM's J2SE RE, as well as many open-source projects like Kaffe.

    +

    The two main frameworks for developing graphical applications are those of GNOME and KDE. These projects are based on the GTK+ and Qt widget toolkits, respectively, which can also be used independently of the larger framework. Both support a wide variety of languages. There are a number of Integrated Development Environments (IDEs) available including Anjuta, Code::Blocks, Eclipse, KDevelop, Lazarus, MonoDevelop, NetBeans, and Omnis Studio while the long-established editors Vim and Emacs remain popular.

    +
    +
    +
    +

    +1.3. Reasons for Using Linux

    +
    • Linux is free:

    +

    As in "free beer". Linux can be downloaded in its entirety from the Internet completely for free. No registration fees, no costs per user, free updates, and freely available source code in case you want to change the behavior of your system. +Most of all, Linux is free as in "free speech": +The license commonly used is the GNU Public License (GPL). The license says that anybody who may want to do so, has the right to change Linux and eventually to redistribute a changed version, on the one condition that the code is still available after redistribution. In practice, you are free to grab a kernel image and sell the new code, as long as your customers can still have a copy of that code.

    +
    • Linux is portable to any hardware platform:

    +

    A vendor, who wants to sell a new type of computer and who does not know what kind of OS his/her new machine will run, can take a Linux kernel and make it work on his/her hardware, because documentation related to this activity is freely available.

    +
    • Linux was made to keep on running:

    +

    As with UNIX, a Linux system expects to run without rebooting all the time. That is why a lot of tasks are being executed at night or scheduled automatically for other times, resulting in higher availability during busier periods and a more balanced use of the hardware. This property allows for Linux to be applicable to environments where people do not have the time or the possibility to control their systems constantly.

    +
    • Linux is secure and versatile:

    +

    The security model used in Linux is based on the UNIX idea of security, which is known to be robust and of proven quality. But Linux is not only safe from attacks from the Internet: it will adapt equally to other situations, utilizing the same high standards for security.

    +
    • Linux is scalable:

    +

    From a Palmtop with 2 MB of memory to a petabyte storage cluster with hundreds of nodes: add or remove the appropriate packages and Linux fits all. One does not need a supercomputer anymore,because you can use Linux to do big things using the building blocks provided with the system. If one wants to do little things, such as making an operating system for an embedded processor or just recycling your old 486, Linux will do that as well.

    +
    • The Linux OS and Linux applications have very short debug−times:

    +

    Because Linux has been developed and tested by thousands of people, both errors and people to fix them are found very quickly. It often happens that there are only a couple of hours between discovery and fixing of a bug.

    +
    +
    +
    +

    +2. Getting Started

    +
    +

    +2.1. Logging in, activating the user interface and logging out

    +

    In order to work on a Linux system directly, one needs to provide a user name and password. You always need to authenticate to the system. Most PC−based Linux systems have two basic modes for a system to run in: either quick and clean in text console mode,which includes with mouse, multitasking and multi−user features, or in graphical console mode, which looks better but eats more system resources.

    +
    +

    +2.1.1. Graphical Mode

    +

    This is the default nowadays on most desktop computers. You know you will be connecting to the system using graphical mode when you are first asked for your user name, and then to type your password.

    +

    To log in, make sure the mouse pointer is in the login window, provide your user name and password to the system and click OK or press Enter. +It is generally considered a bad idea to connect (graphically) using the root user name, the system adminstrator's account, since the use of graphics includes running a lot of extra programs, in root's case with a lot of extra permissions. To keep all risks as low as possible, use a normal user account to connect graphically. But there are enough risks to keep this in mind as a general advice, for all use of the root account: only log in as root when extra privileges are required.

    +

    After entering your user name/password combination, it can take a little while before the graphical environment is started, depending on the CPU speed of your computer, on the software you use and on your personal settings.

    +

    To continue, you will need to open a terminal window or xterm for short (X being the name for the underlying software supporting the graphical environment). This program can be found in the Applications−>Utilities->System Tools or Internet menu, depending on what window manager you are using. There might be icons that you can use as a shortcut to get an xterm window as well, and clicking the right mouse button on the desktop background will usually present you with a menu containing a terminal window application.

    +

    While browsing the menus, you will notice that a lot of things can be done without entering commands via the keyboard. For most users, the good old point−n−click method of dealing with the computer will do. But for those who want to enter the "heart" of the system, a tool stronger than a mouse will be required to handle the various tasks. This tool is the shell, and when in graphical mode, we activate our shell by opening a terminal window.

    +

    A terminal window should always show a command prompt when you open one. This terminal shows a standard prompt, which displays the user's login name, and the current working directory, represented by the twiddle (~)

    +

    Another common form for a prompt is this one: +[

    +
    +

    user@host
    +

    dir]

    +

    In the above example, user will be your login name, hosts the name of the machine you are working on, and dir an indication of your current location in the file system. Prompts can display all kinds of information, but they are not part of the commands you are giving to your system. To disconnect from the system in graphical mode, you need to close all terminal windows and other applications. After that, hit the logout icon or find Log Out in the menu. Closing everything is not really necessary, and the system can do this for you, but session management might put all currently open applications back on your screen when you connect again, which takes longer and is not always the desired effect. However, this behavior is configurable.

    +

    When you see the login screen again, asking to enter user name and password, logout was successful.

    +
    +
    +

    +2.1.2. Text Mode

    +

    One is in text mode when the whole screen is black, showing (in most cases white) characters. A text mode login screen typically shows some information about the machine you are working on, the name of the machine and a prompt waiting for you to log in.

    +

    The login is different from a graphical login, in that you have to hit the Enter key after providing your user name, because there are no buttons on the screen that you can click with the mouse. Then you should type your password, followed by another Enter. You will not see any indication that you are entering something, not even an asterisk, and you won't see the cursor move. But this is normal on Linux and is done for security +reasons.

    +

    When the system has accepted you as a valid user, you may get some more information, called the message of the day, which can be anything. Additionally, it is popular on UNIX systems to display a fortune cookie, which contains some general wise or unwise (this is up to you) thoughts. After that, you will be given a shell, indicated with the same prompt that you would get in graphical mode.

    +

    Also in text mode: log in as root only to do setup and configuration that absolutely requires administrator privileges, such as adding users, installing software packages, and performing network and other system configuration. Once you are finished, immediately leave the special account and resume your work as a non−privileged user.

    +

    Logging out is done by entering the logout command, followed by Enter. You are successfully disconnected from the system when you see the login screen again.Don't power−off the computer after logging out. It is not meant to be shut off without application of the proper procedures for halting the system. Powering it off without going through the halting process might cause severe damage!

    +
    +
    +
    +
    +

    +3. Basic Commands

    +
    +

    +3.1. ls

    +

    When invoked without any arguments, ls lists the files in the current working directory. A directory that is not the current working directory can be specified and ls will list the files there. The user also may specify any list of files and directories. In this case, all files and all contents of specified directories will be listed. The name ls is derived from list segments which was used in earlier systems.

    +

    Files whose names start with "." are not listed, unless the -a flag is specified or the files are specified explicitly.

    +

    Without options, ls displays files in a bare format. This bare format however makes it difficult to establish the type, permissions, and size of the files. The most common options to reveal this information or change the list of files are:

    +<block_quote>
      +
    • -l long format, displaying Unix file types, permissions, number of hard links, owner, group, size, date, and filename

    • +
    • -F appends a character revealing the nature of a file, for example, * for an executable, or / for a directory. Regular files have no suffix.

    • +
    • -a lists all files in the given directory, including those whose names start with "." (which are hidden files in Unix). By default, these files are excluded from the list.

    • +
    • -R recursively lists subdirectories. The command ls -R / would therefore list all files.

    • +
    • -d shows information about a symbolic link or directory, rather than about the link's target or listing the contents of a directory.

    • +
    • -t sort the list of files by modification time.

    • +
    • -h print sizes in human readable format. (e.g., 1K, 234M, 2G, etc.)

    • +
    </block_quote>

    In some environments, providing the option --color (for GNU ls) or -G (FreeBSD ls) causes ls to highlight different types of files with different colors, instead of with characters as -F would. To determine what color to use for a file, GNU ls checks the Unix file type, the file permissions, and the file extension, while FreeBSD ls checks only the Unix file type and file permissions.:

    +
    $ ls
    +jeeves.rst psmith.html blandings.html
    +$ ls -l
    +drwxr--r--   1 plum  editors   4096  jeeves
    +-rw-r--r--   1 plum  editors  30405  psmith
    +-r-xr-xr-x   1 plum  plum      8460  blandings
    +

    Here "$" actually is the beginning of the prompt. This is typical in most Unix-based systems.

    +
    +
    +

    +3.2. date

    +

    The Unix date command displays the time and date. The super-user can use it to set the system clock.

    +

    With no options, the date command displays the current date and time, including the abbreviated day name, abbreviated month name, day of the month, the time separated by colons, the timezone name, and the year. For example:

    +
    $date
    +Tue Sep  8 12:01:45 IST 2009
    +

    On some systems to set the current date and time to September 8, 2004 01:22 you type:

    +
    $date --set="20040908 01:22"
    +

    In order to view the various options for the date command, type:

    +
    $man date
    +

    This will take you to the "Manual" page comprising of all the details on the date command. You can return to the terminal from the "man" page by pressing the Esc key in the keyboard and typing ":q" in that order.

    +
    +
    +

    +3.3. cd

    +

    This stands for "change directory". When one wants to go up to the parent directory, bypassing the tree of directories one has entered, “ cd ..” can be used.

    +

    One dot '.' represents the current directory while two dots '..' represent the parent directory.

    +

    “ cd -” will return you to the previous directory (a bit like an “undo”).

    +

    You can also use cd absolute path or cd relative path (see below):

    +

    Absolute paths:

    +<block_quote>

    An “ absolute path” is easily recognised from the leading forward slash, /. The / means that you start at the top level directory and continue down.

    </block_quote>

    For example to get to /boot/grub you would type:

    +
    $cd /boot/grub
    +

    This is an absolute path because you start at the top of the hierarchy and go downwards from there (it doesn't matter where in the filesystem you were when you typed the command).

    +

    Relative paths:

    +<block_quote>

    A “ relative path” doesn't have a preceding slash. Use a relative path when you start from a directory below the top level directory structure. This is dependent on where you are in the filesystem.

    +

    For example if you are in root's home directory and want to get to /root/music, you type:

    +
    $ cd music
    </block_quote>

    Please note that there is no / using the above cd command. Using a / would cause this to be an absolute path, working from the top of the hierarchy downward.

    +
    +
    +

    +3.4. who

    +

    The standard Unix command who displays a list of users who are currently logged into a computer.

    +

    The who command is related to the command w, which provides the same information but also displays additional data and statistics.:

    +
    $who
    +beeblebrox tty7         2009-09-08 10:50 (:0)
    +beeblebrox pts/0        2009-09-08 11:25 (:0.0)
    +dumbledore pts/1        2009-09-08 18:11 (potter.xyz.in)
    +beeblebrox pts/2        2009-09-08 18:53 (:0.0)
    +

    The command can be invoked with the arguments am i or am I (so it is invoked as who am i or * who am I*), showing information about the current terminal only (see the -m option below, of which this invocation is equivalent).

    +

    In order to find out the various options that can be appended to the who command, check the man page by typing out the following in the terminal:

    +
    $man who
    +

    This will take you to the "Manual" page containing details about the who command

    +
    +
    +

    +3.5. mkdir

    +

    This command is used to make a new directory. Normal usage is as straightforward as follows:

    +
    $mkdir name_of_directory
    +

    Where name_of_directory is the name of the directory one wants to create. When typed as above (ie. normal usage), the new directory would be created within the current directory. On Unix, multiple directories can be specified, and mkdir will try to create all of them.

    +
    +

    +3.5.1. Options

    +

    On Unix-like operating systems, mkdir takes options. Three of the most common options are:

    +<block_quote>
      +
    • -p: will also create all directories leading up to the given directory that do not exist already. If the given directory already exists, ignore the error.

    • +
    • -v: display each directory that mkdir creates. Most often used with -p.

    • +
    • -m: specify the octal permissions of directories created by mkdir.

    • +
    </block_quote>

    -p is most often used when using mkdir to build up complex directory hierarchies, in case a necessary directory is missing or already there. -m is commonly used to lock down temporary directories used by shell scripts.

    +
    +
    +

    +3.5.2. Examples

    +

    An example of -p in action is:

    +
    $mkdir -p /tmp/a/b/c
    +

    If /tmp/a exists but /tmp/a/b does not, mkdir will create /tmp/a/b before creating /tmp/a/b/c.

    +

    And an even more powerful command, creating a full tree at once (this however is a Shell extension, nothing mkdir does itself):

    +
    $mkdir -p tmpdir/{trunk/sources/{includes,docs},branches,tags}
    +

    This will create:

    +<definition_list><definition_list_item><term>tmpdir - branches</term><definition>
      +
    • tag

    • +
    • <definition_list><definition_list_item><term>trunk - sources - includes</term><definition>
      • docs

      </definition>
      </definition_list_item>
      </definition_list>
    • +
    </definition>
    </definition_list_item>
    </definition_list>
    +
    +
    +
    +
    +

    +4. Getting Help

    +
    +

    +4.1. apropos and whatis

    +

    This is a command to search the manual pages files in Unix and Unix-like operating systems.

    +
    $ apropos grep
    +egrep       egrep (1)       Search a file for a pattern using full regular expressions
    +fgrep       fgrep (1)       Search a file for a fixed-character string
    +fmlgrep     fmlgrep (1)     Search a file for a pattern
    +grep        grep (1)        Search a file for a pattern
    +gzgrep      gzgrep (1)      Search a possibly compressed file for a regular expression
    +nisgrep     nismatch (1)    Utilities for searching NIS+ tables
    +pgrep       pgrep (1)       Find or signal a process by name or other attribute
    +zgrep       zgrep (1)       Search a possibly compressed file for a regular expression
    +...
    +

    In this example, the user uses apropos to search for the string "grep", and apropos returns the indicated man pages that include the term "grep".

    +

    A short index of explanations for commands is available using the whatis command, like in the examples below:

    +
    $whatis ls
    +ls (1)           - list directory contents
    +

    This displays short information about a command, and the first section in the collection of man pages that contains an appropriate page.

    +

    If you don't know where to get started and which man page to read, apropos gives more information. Say that you do not know how to start a browser, then you could enter the following command:

    +
    $apropos browser
    +gmusicbrowser (1)    - Jukebox for large collections of audio files
    +infobrowser (1)      - read Info documents
    +libsmbclient (7)     - An extension library for browsers and that               can be used...
    +opera (1)            - a standards-compliant graphical Web browser
    +sensible-browser (1) - sensible editing, paging, and web browsing
    +smbtree (1)          - A text based smb network browser
    +tvtk_doc (1)         - A GUI based TVTK documentation search browser.
    +viewres (1)          - graphical class browser for Xt
    +w3m (1)              - a text based Web browser and pager
    +www-browser (1)      - a text based Web browser and pager
    +...
    +
    +
    +

    +4.2. man

    +

    Man pages (short for "manual pages") are the extensive documentation that comes preinstalled with almost all substantial Unix and Unix-like operating systems. The Unix command used to display them is man. Each page is a self-contained document.

    +

    To read a manual page for a Unix command, one can use:

    +
    $ man <command_name>
    +

    at a shell prompt; for example, "man ftp". In order to simplify navigation through the output, man generally uses the less terminal pager.

    +

    Pages are traditionally referred to using the notation "name(section)"; for example, ftp(1). The same page name may appear in more than one section of the manual, this can occur when the names of system calls, user commands, or macro packages coincide. Two examples are man(1) and man(7), or exit(2) and exit(3). The syntax for accessing the non-default manual section varies between different man implementations. On Linux and BSD, for example, the syntax for reading *printf(3) is:

    +
    $man 3 printf
    +

    Another example:

    +
    $man man
    +

    The previous example will take you to the "Manual" page entry about manual pages!

    +
    +

    +4.2.1. Layout

    +

    All man pages follow a common layout that is optimized for presentation on a simple ASCII text display, possibly without any form of highlighting or font control. Sections present may include:

    +<definition_list><definition_list_item><term>NAME</term><definition>

    The name of the command or function, followed by a one-line description of what it does.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SYNOPSIS</term><definition>

    In the case of a command, you get a formal description of how to run it and what command line options it takes. For program functions, a list of the parameters the function takes and which header file contains its definition. For experienced users, this may be all the documentation they need.

    </definition>
    </definition_list_item>
    <definition_list_item><term>DESCRIPTION</term><definition>

    A textual description of the functioning of the command or function.

    </definition>
    </definition_list_item>
    <definition_list_item><term>EXAMPLES</term><definition>

    Some examples of common usage.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SEE ALSO</term><definition>

    A list of related commands or functions.

    </definition>
    </definition_list_item>
    </definition_list>

    Other sections may be present, but these are not well standardized across man pages. Common examples include: OPTIONS, EXIT STATUS, ENVIRONMENT, KNOWN BUGS, FILES, AUTHOR, REPORTING BUGS, HISTORY and COPYRIGHT.

    +

    These days virtually every Unix command line application comes with its man page, and many Unix users perceive a lack of man pages as a sign of low quality; indeed, some projects, such as Debian, go out of their way to write man pages for programs lacking one. Few alternatives to man have enjoyed much popularity, with the possible exception of the GNU project's "info" system, an early and simple hypertext system.

    +

    However, the format of a single page for each application, the lack of classification within the sections and the relatively unsophisticated formatting facilities have motivated the development of alternative documentation systems, such as the previously mentioned "info" system.

    +

    Most Unix GUI applications (particularly those built using the GNOME and KDE development environments) now provide end-user documentation in HTML and include embedded HTML viewers such as yelp for reading the help within the application.

    +

    Usually the man pages are written in English. Translations into other languages can be also available on the system.

    +

    The default format of the man pages is troff, with either the macro package man (appearance oriented) or on some systems mdoc (semantic oriented). This makes it possible to typeset a man page to PostScript, PDF and various other formats for viewing or printing.

    +
    +
    +
    +

    +4.3. info

    +

    info is a software utility which forms a hypertextual, multipage documentation and help viewer working on a command line interface, useful when there is no GUI available.

    +

    The syntax is

    +
    $ info <command_name>
    +

    info processes info files, which are Texinfo formatted files, and presents the documentation as a tree, with simple commands to traverse the tree and to follow cross references. For instance

    +<block_quote>
      +
    • n goes to the next page.

    • +
    • p goes to the previous page.

    • +
    • u goes to the upper page.

    • +
    • l goes to the last(visited) node

    • +
    • To follow a cross reference, the cursor can be moved over a link (a word preceded by a <title_reference>*</title_reference>) and enter pressed.

    • +
    </block_quote>

    info was initially written for use with GNU/Linux and then ported to other Unix-like operating systems.

    +
    +
    +

    +4.4. --help

    +

    Most GNU commands support the --help, which gives a short explanation about how to use the command and a list of available options. Below is the output of this option with the cat command:

    +
    $ userprompt@host: cat --help
    +Usage: cat [OPTION] [FILE]...
    +Concatenate FILE(s), or standard input, to standard output.
    +
    +  -A, --show-all           equivalent to -vET
    +  -b, --number-nonblank    number nonempty output lines
    +  -e                       equivalent to -vE
    +  -E, --show-ends          display $ at end of each line
    +  -n, --number             number all output lines
    +  -s, --squeeze-blank      suppress repeated empty output lines
    +  -t                       equivalent to -vT
    +  -T, --show-tabs          display TAB characters as ^I
    +  -u                       (ignored)
    +  -v, --show-nonprinting   use ^ and M- notation, except for LFD and              TAB
    +  --help     display this help and exit
    +  --version  output version information and exit
    +
    +With no FILE, or when FILE is -, read standard input.
    +
    +Examples:
    +  cat f - g  Output f's contents, then standard input, then g's           contents.
    +  cat        Copy standard input to standard output.
    +
    +Report bugs to <bug-coreutils@gnu.org>.
    +
    +
    +
    +

    +5. Basic file handling

    +
    +

    +5.1. cp

    +

    cp is the command entered in a Unix shell to copy a file from one place to another, possibly on a different filesystem. The original file remains unchanged, and the new file may have the same or a different name.

    +
    +

    +5.1.1. Usage

    +

    To copy a file to another file:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ][ -- ] SourceFile TargetFile
    +

    To copy a file to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -r | -R ] [ -- ] SourceFile ...              TargetDirectory
    +

    To copy a directory to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -- ] { -r | -R }
    +SourceDirectory ... TargetDirectory
    +
    +
    +

    +5.1.2. Flags

    +

    -f (force) – specifies removal of the target file if it cannot be opened for write operations. The removal precedes any copying performed by the cp command.

    +

    -P – makes the cp command copy symbolic links. The default is to follow symbolic links, that is, to copy files to which symbolic links point.

    +

    -i (interactive) – prompts you with the name of a file to be overwritten. This occurs if the TargetDirectory or TargetFile parameter contains a file with the same name as a file specified in the SourceFile or SourceDirectory parameter. If you enter y or the locale's equivalent of y, the cp command continues. Any other answer prevents the cp command from overwriting the file.

    +

    -p (preserve) – duplicates the following characteristics of each SourceFile/SourceDirectory in the corresponding TargetFile and/or TargetDirectory:

    +<block_quote>
      +
    • The time of the last data modification and the time of the last access.

    • +
    • The user ID and group ID (only if it has permissions to do this)

    • +
    • The file permission bits and the SUID and SGID bits.

    • +
    </block_quote>

    -R (recursive) – copy directories (recursively copying all the contents)

    +
    +
    +

    +5.1.3. Examples

    +

    To make a copy of a file in the current directory, enter:

    +
    $ cp prog.c prog.bak
    +

    This copies prog.c to prog.bak. If the prog.bak file does not already exist, the cp command creates it. If it does exist, the cp command replaces it with a copy of the prog.c file.

    +

    To copy a file in your current directory into another directory, enter:

    +
    $ cp zaphod /home/books/hhgg
    +

    This copies the jones file to /home/books/hhgg/zaphod.

    +

    To copy a file to a new file and preserve the modification date, time, and access control list associated with the source file, enter:

    +
    $ cp -p martin_luther_king martin_luther_king.jr
    +

    This copies the martin_luther_king file to the martin_luther_king.jr file. Instead of creating the file with the current date and time stamp, the system gives the martin_luther_king.jr file the same date and time as the martin_luther_king file. The martin_luther_king.jr file also inherits the martin_luther_king file's access control protection.

    +

    To copy all the files in a directory to a new directory, enter:

    +
    $ cp /home/galactica/clients/* /home/hhgg/customers
    +

    This copies only the files in the clients directory to the customers directory.

    +

    To copy a directory, including all its files and subdirectories, to another directory, enter:

    +<block_quote>

    $ cp -R /home/hhgg/clients /home/hhgg/customers

    </block_quote>

    This copies the clients directory, including all its files, subdirectories, and the files in those subdirectories, to the customers/clients directory.

    +

    To copy a specific set of files of any extension to another directory, enter:

    +
    $ cp zaphod arthur ford /home/hhgg/clients
    +

    This copies the zaphod, arthur, and ford files in your current working directory to the /home/hhgg/clients directory.

    +

    To use pattern-matching characters to copy files, enter:

    +
    $ cp programs/*.py .
    +

    This copies the files in the programs directory that end with .py to the current directory, signified by the single "." (dot). You must type a space between the py and the final dot.

    +
    +
    +
    +

    +5.2. mv

    +

    mv (short for move) is a Unix command that moves one or more files or directories from one place to another. The original file is deleted, and the new file may have the same or a different name. If possible (i.e. when the original and new files are on the same file system), mv will rename the file instead. Write permission is required on all directories being modified.

    +
    +

    +5.2.1. Conflicting existing file

    +

    In all cases, when a file is moved to have the name of an existing file (in the same directory), the existing file is deleted. If the existing file is not writable but is in a directory that is writable, then the mv command asks for confirmation if possible (i.e. if run from a terminal) before proceeding, unless the -f (force) option is used.

    +
    +
    +

    +5.2.2. Differences with copy and delete

    +

    Note that, usually, when moving files within the same volume, moving (and/or renaming) is not the same as simply copying and then deleting the original. When moving a file, the link is simply removed from the old parent directory and added to the new parent directory. However, the file itself is untouched (i.e. it has the same inodes and resides at the same place on the disk). For example, you cannot copy a file you cannot read, but you can move (and/or rename) it (provided you have write permission to its old and new parent directories). Also, suppose there is a non-empty directory you do not have write permission to. You cannot delete this directory (since you cannot delete its contents); but you can move (and/or rename) it. Also, since moving between filenames on a single volume does not involve copying, it is faster and does not place strain of lots of reads and writes on the disk. Moving files across different volumes, however, does necessitate copying and deleting.

    +
    +
    +

    +5.2.3. Examples

    +
    $ mv myfile mynewfilename    renames a file
    +$ mv myfile otherfilename    renames a file and deletes the existing            file "myfile"
    +$ mv myfile /myfile          moves 'myfile' from the current            directory to the root directory
    +$ mv myfile dir/myfile       moves 'myfile' to 'dir/myfile' relative            to the current directory
    +$ mv myfile dir              same as the previous command (the          filename is implied to be the same)
    +$ mv myfile dir/myfile2      moves 'myfile' to dir and renames it to            'myfile2'
    +$ mv foo bar baz dir         moves multiple files to directory dir
    +$ mv --help                  shows a very concise help about the                syntax of the command
    +$ man mv                     prints an extensive user manual for                'mv' in the terminal
    +

    In all cases, the file or files being moved or renamed can be a directory.

    +

    Note that when the command is called with two arguments (as mv name1 name2 or mv name1 /dir/name2), it can have three different effects, depending on whether name2 does not exist, is an existing file, or is an existing directory. If the user intends to refer to an existing directory, /. (or in some Unix versions / is sufficient) may be appended to the name to force the system to check this. To move a file to a new directory, the directory must be created first.

    +
    +
    +
    +

    +5.3. rm

    +

    rm (short for "remove") is one of several basic Unix command lines that operates on files. It is used to delete files from a filesystem. The data is not actually destroyed. Only the index listing where the file is stored is destroyed, and the storage is made available for reuse. There are undelete utilities that will attempt to reconstruct the index and can bring the file back if the parts were not reused.

    +

    Here's example to remove a file named "foo" from a directory, here shown with the -i option:

    +
    $ rm -i foo
    +remove foo? y
    +
    +

    +5.3.1. Options

    +

    Common options that rm accepts include:

    +<block_quote>
      +
    • -r, which removes directories, removing the contents recursively beforehand (so as not to leave files without a directory to reside in) ("recursive")

    • +
    • -i, which asks for every deletion to be confirmed ("interactive")

    • +
    • -f, which ignores non-existent files and overrides any confirmation prompts ("force")

    • +
    • -v, which shows what is being removed as it happens ("verbose")

    • +
    </block_quote>

    rm is often aliased to "rm -i" so as to avoid accidental deletion of files. If a user still wishes to delete a large number of files without confirmation, they can manually cancel out the -i argument by adding the -f option (as the option specified later on the expanded command line "rm -i -f" takes precedence).

    +

    rm -rf (variously, rm -rf /, rm -rf <title_reference>*</title_reference>, and others) is frequently used in jokes and anecdotes about Unix disasters. The rm -rf variant of the command, if run by a superuser on the root directory, would cause the contents of every writable mounted filesystem on the computer to be deleted.

    +

    rm is often used in conjunction with xargs to supply a list of files to delete:

    +
    xargs rm < filelist
    +

    When rm is used on a symbolic link, it deletes the link, but does not affect the target of the link.

    +
    +
    +

    +5.3.2. Permissions

    +

    Usually, on most filesystems, deleting a file requires write permission on the parent directory (and execute permission, in order to enter the directory in the first place). (Note that, confusingly for beginners, permissions on the file itself are irrelevant. However, GNU rm asks for confirmation if a write-protected file is to be deleted, unless the -f option is used.)

    +

    To delete a directory (with rm -r), one must delete all of its contents recursively. This requires that one must have read and write and execute permission to that directory (if it's not empty) and all non-empty subdirectories recursively (if there are any). The read permissions are needed to list the contents of the directory in order to delete them. This sometimes leads to an odd situation where a non-empty directory cannot be deleted because one doesn't have write permission to it and so cannot delete its contents; but if the same directory were empty, one would be able to delete it.

    +

    If a file resides in a directory with the sticky bit set, then deleting the file requires one to be the owner of the file.

    +
    +
    +
    +
    +

    +6. Command Line Arguments

    +

    In computer command line interfaces, a command line argument is an argument sent to a program being called. In general, a program can take any number of command line arguments, which may be necessary for the program to run, or may even be ignored, depending on the function of that program.

    +

    For example, in Unix and Unix-like environments, an example of a command-line argument is:

    +
    rm file.s
    +

    "file.s" is a command line argument which tells the program rm to remove the file "file.s".

    +

    Programming languages such as C, C++ and Java allow a program to interpret the command line arguments by handling them as string parameters in the main function.

    +

    A command line option or simply option (also known as a command line parameter, flag, or a switch) is an indication by a user that a computer program should change its default output.

    +

    Long options are introduced via "--", and are typically whole words. For example, ls --long --classify --all. Arguments to long options are provided with "=", as ls --block-size=1024. Some Unix programs use long options with single dashes, for example MPlayer as in mplayer -nosound.

    +

    Linux also uses "--" to terminate option lists. For example, an attempt to delete a file called -file1 by using rm -file1 may produce an error, since rm may interpret -file1 as a command line switch. Using rm -- -file1 removes ambiguity.

    +
    +
    +

    +7. Basic Text Processing

    +
    +

    +7.1. head

    +

    head is a program on Unix and Unix-like systems used to display the first few lines of a text file or piped data. The command syntax is:

    +
    $ head [options] <file_name>
    +

    By default, head will print the first 10 lines of its input to the standard output. The number of lines printed may be changed with a command line option. The following example shows the first 20 lines of filename:

    +
    $ head -n 20 filename
    +

    This displays the first 5 lines of all files starting with foo:

    +
    $ head -n 5 foo*
    +

    Some versions omit the n and just let you say -5.

    +
    +

    +7.1.1. Flags

    +
    -c <x number of bytes> Copy first x number of bytes.
    +

    Other options: sed

    +

    Many early versions of Unix did not have this command, and so documentation and books had sed do this job:

    +
    sed 5q foo
    +

    This says to print every line (implicit), and quit after the fifth.

    +
    +
    +
    +

    +7.2. tail

    +

    tail is a program on Unix and Unix-like systems used to display the last few lines of a text file or piped data.

    +

    The command-syntax is:

    +
    $ tail [options] <file_name>
    +

    By default, tail will print the last 10 lines of its input to the standard output. With command line options the number of lines printed and the printing units (lines, blocks or bytes) may be changed. The following example shows the last 20 lines of filename:

    +
    $ tail -n 20 filename
    +

    This example shows the last 15 bytes of all files starting with foo:

    +
    $ tail -c 15 foo*
    +

    This example shows all lines of filename from the second line onwards:

    +
    $ tail -n +2 filename
    +

    Using an older syntax (still used in Sun Solaris as the -n option is not supported), the last 20 lines and the last 50 bytes of filename can be shown with the following command:

    +
    $ tail -20 filename
    +$ tail -50c filename
    +

    However this syntax is now obsolete and does not conform with the POSIX 1003.1-2001 standard. Even if still supported in current versions, when used with other options (like -f, see below), these switches could not work at all.

    +
    +

    +7.2.1. File monitoring

    +

    tail has a special command line option -f (follow) that allows a file to be monitored. Instead of displaying the last few lines and exiting, tail displays the lines and then monitors the file. As new lines are added to the file by another process, tail updates the display. This is particularly useful for monitoring log files. The following command will display the last 10 lines of messages and append new lines to the display as new lines are added to messages:

    +
    $ tail -f /var/adm/messages
    +

    To interrupt tail while it is monitoring, break-in with Ctrl+C. This command can be run "in the background" with &, see job control.

    +

    If you have a command's result to monitor, you can use the watch command.

    +
    +
    +
    +

    +7.3. cut

    +

    In computing, cut is a Unix command line utility which is used to extract sections from each line of input — usually from a file.

    +

    Extraction of line segments can typically be done by bytes (-b), characters (-c), or fields (-f) separated by a delimiter (-d — the tab character by default). A range must be provided in each case which consists of one of N, N-M, N- (N to the end of the line), or -M (beginning of the line to M), where N and M are counted from 1 (there is no zeroth value). Since version 6, an error is thrown if you include a zeroth value. Prior to this the value was ignored and assumed to be 1.

    +

    Assuming a file named file containing the lines:

    +
    foo:bar:baz:qux:quux
    +one:two:three:four:five:six:seven
    +alpha:beta:gamma:delta:epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +

    To output the fourth through tenth characters of each line:

    +
    $ cut -c 4-10 file
    +

    This gives the output:

    +
    :bar:ba
    +:two:th
    +ha:beta
    +

    To output the fifth field through the end of the line of each line using the colon character as the field delimiter:

    +
    $ cut -d : -f 5- file
    +

    This gives the output:

    +
    quux
    +five:six:seven
    +epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +
    +
    +

    +7.4. paste

    +

    paste is a Unix command line utility which is used to join files horizontally (parallel merging) by outputting lines consisting of the sequentially corresponding lines of each file specified, separated by tabs, to the standard output. It is effectively the horizontal equivalent to the utility cat command which operates on the vertical plane of two or more files.

    +

    To paste several columns of data together into the file www from files who, where, and when:

    +
    $ paste who where when > www
    +

    If the files contain:

    +++++
    +

    This creates the file named www containing:

    +
    Batman            GothamCity       January 3
    +Trillian          Andromeda        February 4
    +Jeeves            London           March 19
    +
    +
    +
    +

    +8. Shell Meta Characters

    +

    Unix recognizes certain special characters, called "meta characters," as command directives. The shell meta characters are recognized anywhere they appear in the command line, even if they are not surrounded by blank space. For that reason, it is safest to only use the characters A-Z, a-z, 0-9, and the period, dash, and underscore characters when naming files and directories on Unix. If your file or directory has a shell meta character in the name, you will find it difficult to use the name in a shell command.

    +

    The shell meta characters include:

    +

    / < > ! $ % ^ & * | { } [ ] " ' ` ~ ;

    +

    Different shells may differ in the meta characters recognized.

    +

    As an example,

    +
    $ ls file.*
    +

    run on a directory containing the files file, file.c, file.lst, and myfile would list the files file.c and file.lst. However,:

    +
    $ ls file.?
    +

    run on the same directory would only list file.c because the ? only matches one character, no more, no less. This can save you a great deal of typing time. For example, if there is a file called california_cornish_hens_with_wild_rice and no other files whose names begin with 'c', you could view the file without typing the whole name by typing this:

    +
    $ more c*
    +

    because the c* matches that long file name.

    +

    Filenames containing metacharacters can pose many problems and should never be intentionally created. If you do find that you've created a file with metacharacters, and you would like to remove it, you have three options. You may use wildcards to match metacharacter, use the to directly enter the filename, or put the command in double quotes (except in the case of double quotes within the file name, these must be captured with one of the first two methods). For example, deleting a file named <title_reference>"``*`|more</title_reference>"` can be accomplished with:

    +
    $ rm ??more
    +

    or:

    +
    $ rm $\backslash$*$\backslash$|more
    +

    or:

    +
    $ rm ''*|more''
    +
    +
    +

    +9. Looking At Files

    +
    +

    +9.1. cat

    +

    The cat command is a standard Unix program used to concatenate and display files. The name is from "catenate", a synonym of concatenate.

    +

    The Single Unix Specification specifies the behavior that the contents of each of the files given in sequence as arguments will be written to the standard output in the same sequence, and mandates one option, -u, where each byte is printed as it is read.

    +

    If the filename is specified as -, then cat will read from standard input at that point in the sequence. If no files are specified, cat will read from standard input entered.

    +
    +

    +9.1.1. Jargon File Definition

    +

    The Jargon File version 4.4.7 lists this as the definition of cat:

    +
    1. To spew an entire file to the screen or some other output sink without
    +     pause (syn. blast).
    +
    +2. By extension, to dump large amounts of data at an unprepared target or
    +     with no intention of browsing it carefully. Usage: considered silly.
    +     Rare outside Unix sites. See also dd, BLT.
    +
    +     Among Unix fans, *cat(1)* is considered an excellent example of
    +     user-interface design, because it delivers the file contents without
    +     such verbosity as spacing or headers between the files, and because
    +     it does not require the files to consist of lines of text, but works
    +     with any sort of data.
    +
    +     Among Unix critics, *cat(1)* is considered the canonical example of
    +     bad user-interface design, because of its woefully unobvious name.
    +     It is far more often used to blast a single file to standard output
    +     than to concatenate two or more files. The name cat for the former
    +     operation is just as unintuitive as, say, LISP's cdr.
    +
    +     Of such oppositions are holy wars made...
    +
    +
    +

    +9.1.2. Useless Use of 'cat'

    +

    UUOC (from comp.unix.shell on Usenet) stands for “Useless Use of cat”. As it is observed on comp.unix.shell, “The purpose of cat is to concatenate (or 'catenate') files. If it's only one file, concatenating it with nothing at all is a waste of time, and costs you a process.”

    +

    Nevertheless one sees people doing:

    +
    $ cat file | some_command and its args ...
    +

    instead of the equivalent and cheaper:

    +
    <file some_command and its args ...
    +

    or (equivalently and more classically):

    +
    some_command and its args ... <file
    +

    Since 1995, occasional awards for UUOC have been given out. The activity of fixing instances of UUOC is sometimes called 'demoggification'.

    +

    Amongst many, it is still considered safer to use cat for such cases given that the < and > keys are next to each other in many popular keyboard mappings. While the risk might be low, the impact of using > instead of < can be high and prohibitive.

    +
    +
    +

    +9.1.3. zcat

    +

    zcat is a Unix program similar to cat, that decompresses individual files and concatenates them to standard output. Traditionally zcat operated on files compressed by compress but today it is usually able to operate on gzip or even bzip2 archives. On such systems, it is equivalent to gunzip -c

    +
    +
    +
    +

    +9.2. more

    +

    In computing, more is a command to view (but not modify) the contents of a text file one screen at a time (terminal pager). It is available on Unix and Unix-like systems, DOS, OS/2 and Microsoft Windows. Programs of this sort are called pagers.

    +
    +

    +9.2.1. Usage

    +

    The command-syntax is:

    +
    $ more [options] [file_name]
    +

    If no file name is provided, more looks for input from stdin.

    +

    Once more has obtained input, it displays as much as can fit on the current screen and waits for user input to advance, with the exception that a form feed (^L) will also cause more to wait at that line, regardless of the amount of text on the screen. In the lower-left corner of the screen is displayed the text "--More--" and a percentage, representing the percent of the file that more has paged through. (This percentage includes the text displayed on the current screen.) When more reaches the end of a file (100%) it exits. The most common methods of navigating through a file are Enter, which advances the output by one line, and Space, which advances the output by one screen.

    +

    There are also other commands that can be used while navigating through the document; consult more's man page for more details.

    +

    Options are typically entered before the file name, but can also be entered in the environment variable $MORE. Options entered in the actual command line will override those entered in the $MORE environment variable. Available options may vary between Unix systems.

    +
    +
    +
    +

    +9.3. less

    +

    less is a terminal pager program on Unix, Windows and Unix-like systems used to view (but not change) the contents of a text file one screen at a time. It is similar to more, but has the extended capability of allowing both forward and backward navigation through the file. Unlike most Unix text editors/viewers, less does not need to read the entire file before starting, resulting in faster load times with large files.

    +
    +

    +9.3.1. Usage

    +

    less can be invoked with options to change its behaviour, for example, the number of lines to display on the screen. A few options vary depending on the operating system. While less is displaying the file, various commands can be used to navigate through the file. These commands are based on those used by both more and vi. It is also possible to search for character patterns in the file.

    +

    By default, less displays the contents of the file to the standard output (one screen at a time). If the file name argument is omitted, it displays the contents from standard input (usually the output of another command through a pipe). If the output is redirected to anything other than a terminal, for example a pipe to another command, less behaves like cat.

    +

    The command-syntax is:

    +
    $ less [options] file_name
    +
    +
    +

    +9.3.2. Frequently Used Options

    +<block_quote>
      +
    • -g: Highlights just the current match of any searched string.

    • +
    • -I: Case-insensitive searches.

    • +
    • -M: Shows more detailed prompt, including file position.

    • +
    • -N: Shows line numbers (useful for source code viewing).

    • +
    • -S: Disables line wrap ("chop long lines"). Long lines can be seen by side scrolling.

    • +
    • -?: Shows help.

    • +
    </block_quote>
    +
    +
    +

    +9.3.3. Frequently Used Commands

    +<block_quote>
      +
    • [Arrows]/[Page Up]/[Page Down]/[Home]/[End]: Navigation.

    • +
    • [Space bar]: Next page.

    • +
    • b: Previous page.

    • +
    • ng: Jump to line number n. Default is the start of the file.

    • +
    • nG: Jump to line number n. Default is the end of the file.

    • +
    • /pattern: Search for pattern. Regular expressions can be used.

    • +
    • '^ or g: Go to start of file.

    • +
    • '$ or G: Go to end of file.

    • +
    • s: Save current content (got from another program like grep) in a file.

    • +
    • =: File information.

    • +
    • h: Help.

    • +
    • q: Quit.

    • +
    </block_quote>
    +
    +
    +

    +9.3.4. Examples

    +
    $ less -M readme.txt                     #Read "readme.txt."
    +$ less +F /var/log/mail.log              #Follow mode for log
    +$ file * | less                          #Easier file analysis.
    +$ grep -i void *.c | less -I -p void     #Case insensitive search                                                         for "void" in all .c files
    +
    +
    +
    +
    +

    +10. Directory Structure

    +

    In the File Hierarchy Standard (FHS) all files and directories appear under the root directory "/", even if they are stored on different physical devices. Note however that some of these directories may or may not be present on a Unix system depending on whether certain subsystems, such as the X Window System, are installed.

    +

    The majority of these directories exist in all UNIX operating systems and are generally used in much the same way; however, the descriptions here are those used specifically for the FHS, and are not considered authoritative for platforms other than Linux.

    ++++
    +
    +

    +10.1. man hier

    +

    This is the manual page on the UNIX filesystem. The syntax for this is:

    +
    $ man hier
    +
    +
    +

    +10.2. ls -l

    +

    Shows you huge amounts of information (permissions, owners, size, and when last modified) for folders and files. The syntax is

    +
    $ ls -l
    +

    This can be done after entering the required directory.

    +
    +
    +
    +

    +11. Permissions and Ownership

    +
    +

    +11.1. chmod

    +

    The chmod command (abbreviated from 'change mode') is a shell command and C language function in Unix and Unix-like environments. When executed, it can change file system modes of files and directories. The modes include permissions and special modes.A chmod command first appeared in AT&T Unix version 1, and is still used today on Unix-like machines.

    +
    +

    +11.1.1. Usage

    +

    The chmod command options are specified like this:

    +
    $ chmod [options] mode[,mode] file1 [file2 ...]
    +

    To view what the permissions currently are, type:

    +
    $ ls -l file
    +
    +
    +

    +11.1.2. Command line options

    +

    The chmod command has a number of command line options that affect its behavior. The most common options are:

    +<block_quote>
      +
    • -R: Changes the modes of directories and files recursively

    • +
    • -v: Verbose mode; lists all files as they are being processed

    • +
    </block_quote>
    +
    +11.1.2.1. Symbolic modes
    +

    To the chmod utility, all permissions and special modes are represented by its mode parameter. One way to adjust the mode of files or directories is to specify a symbolic mode. The symbolic mode is composed of three components, which are combined to form a single string of text:

    +
    $ chmod [references][operator][modes] file1 ...
    +

    The references (or classes) are used to distinguish the users to whom the permissions apply. If no references are specified it defaults to “all” but modifies only the permissions allowed by the umask. The references are represented by one or more of the following letters:

    +++++
    +

    The chmod program uses an operator to specify how the modes of a file should be adjusted. The following operators are accepted:

    ++++
    +

    The modes indicate which permissions are to be granted or taken away from the specified classes. There are three basic modes which correspond to the basic permissions:

    +++++
    +

    The combination of these three components produces a string that is understood by the chmod command. Multiple changes can be specified by separating multiple symbolic modes with commas.

    +
    +
    +
    +11.1.2.2. Symbolic examples
    +

    Add the 'read' and 'write' permissions to the 'user' and 'group' classes of a directory:

    +
    $ chmod ug+rw mydir
    +$ ls -ld mydir
    +drw-rw----   2 starwars  yoda  96 Dec 8 12:53 mydir
    +

    For a file, remove write permissions for all classes:

    +
    $ chmod a-w myfile
    +$ ls -l myfile
    +-r-xr-xr-x   2 starwars  yoda 96 Dec 8 12:53 myfile
    +

    Set the permissions for the u*ser and the *g*roup to read and execute only (no write permission) on *mydir.

    +
    $ chmod ug=rx mydir
    +$ ls -ld mydir
    +dr-xr-x---   2 starwars  yoda 96 Dec 8 12:53 mydir
    +
    +
    +
    +11.1.2.3. Octal numbers
    +

    The chmod command also accepts three and four-digit octal numbers representing modes. Using a three-digit octal number to set the modes of a file named myfile :

    +
    $ chmod 664 myfile
    +$ ls -l myfile
    +-rw-rw-r--  1   57 Jul  3 10:13  myfile
    +

    Since the setuid, setgid and sticky bits are not set, this is equivalent to:

    +
    $ chmod 0664 myfile
    +
    +
    +
    +11.1.2.4. Special modes
    +

    The chmod command is also capable of changing the additional permissions or special modes of a file or directory. The symbolic modes use s to represent the setuid and setgid modes, and t to represent the sticky mode. The modes are only applied to the appropriate classes, regardless of whether or not other classes are specified.

    +

    Most operating systems support the specification of special modes using octal modes, but some do not. On these systems, only the symbolic modes can be used.

    +
    +
    +
    +
    +
    +

    +12. Redirection and Piping

    +

    In computing, redirection is a function common to most command-line interpreters, including the various Unix shells that can redirect standard streams to user-specified locations.

    +

    Programs do redirection with the dup2(2) system call, or its less-flexible but higher-level stdio analogues, freopen(3) and popen(3).

    +
    +

    +12.1. Redirecting standard input and standard output

    +

    Redirection is usually implemented by placing certain characters between commands. Typically, the syntax of these characters is as follows:

    +
    $ command1 > file1
    +

    executes command1, placing the output in file1. Note that this will truncate any existing data in file1. To append output to the end of the file, use the >> operator.:

    +
    $ command1 < file1
    +

    executes command1, using file1 as the source of input (as opposed to the keyboard).:

    +
    $ command1 < infile > outfile
    +

    combines the two capabilities: command1 reads from infile and writes to outfile

    +
    +
    +

    +12.2. Piping

    +

    Programs can be run together such that one program reads the output from another with no need for an explicit intermediate file: +A pipeline of three programs run on a text terminal:

    +
    $ command1 | command2
    +

    executes command1, using its output as the input for command2 (commonly called piping, since the "|" character is known as a "pipe").

    +

    This is equivalent to using two redirects and a temporary file:

    +
    $ command1 > tempfile
    +$ command2 < tempfile
    +$ rm tempfile
    +

    A good example for command piping is combining echo with another command to achieve something interactive in a non-interactive shell, e.g.:

    +
    $ echo -e "user\npass" | ftp localhost
    +

    This runs the ftp client with input user, press return, then pass.

    +
    +
    +

    +12.3. Redirecting to and from the standard file handles

    +

    In Unix shells derived from the original Bourne shell, the first two actions can be further modified by placing a number (the file descriptor) immediately before the character; this will affect which stream is used for the redirection. The Unix standard I/O streams are:

    +++++
    +

    For example:

    +
    $ command1 2> file1
    +

    executes command1, directing the standard error stream to file1.

    +

    In shells derived from csh (the C shell), the syntax instead appends the & character to the redirect characters, thus achieving a similar result.

    +

    Another useful capability is to redirect one standard file handle to another. The most popular variation is to merge standard error into standard output so error messages can be processed together with (or alternately to) the usual output. Example:

    +
    $ find / -name .profile > results 2>&1
    +

    will try to find all files named .profile. Executed without redirection, it will output hits to stdout and errors (e.g. for lack of privilege to traverse protected directories) to stderr. If standard output is directed to file results, error messages appear on the console. To see both hits and error messages in file results, merge stderr (handle 2) into stdout (handle 1) using 2>&1 .

    +

    It's possible use 2>&1 before ">" but it doesn't work. In fact, when the interpreter reads 2>&1, it doesn't know yet where standard output is redirected and then standard error isn't merged.

    +

    If the merged output is to be piped into another program, the file merge sequence 2>&1 must precede the pipe symbol, thus:

    +
    $ find / -name .profile 2>&1 | less
    +

    A simplified form of the command:

    +
    $ command > file 2>&1
    +

    is:

    +
    $ command &>file
    +

    or:

    +
    $command >&file
    +
    +
    +

    +12.4. Chained pipelines

    +

    The redirection and piping tokens can be chained together to create complex commands. For example:

    +
    $ ls | grep '\.sh' | sort > shlist
    +

    lists the contents of the current directory, where this output is filtered to only contain lines which contain .sh, sort this resultant output lexicographically, and place the final output in shlist. This type of construction is used very commonly in shell scripts and batch files.

    +
    +
    +

    +12.5. Redirect to multiple outputs

    +

    The standard command tee can redirect output from a command to several destinations.

    +
    $ ls -lrt | tee xyz
    +

    This directs the file list output to both standard output as well as to the file xyz.

    +
    +
    +
    +

    +13. More Text Processing

    +
    +

    +13.1. grep

    +

    grep is a command line text search utility originally written for Unix. The name is taken from the first letters in global / regular expression / print, a series of instructions for the ed text editor. The grep command searches files or standard input globally for lines matching a given regular expression, and prints them to the program's standard output.

    +
    +

    +13.1.1. Usage

    +

    This is an example of a common grep usage:

    +
    $ grep apple fruitlist.txt
    +

    In this case, grep prints all lines containing 'apple' from the file fruitlist.txt, regardless of word boundaries; therefore lines containing 'pineapple' or 'apples' are also printed. The grep command is case sensitive by default, so this example's output does not include lines containing 'Apple' (with a capital A) unless they also contain 'apple'.

    +

    Like most Unix commands, grep accepts command line arguments to change this and many other behaviors. For example:

    +
    $ grep -i apple fruitlist.txt
    +

    This prints all lines containing 'apple' regardless of capitalization. The '-i' argument tells grep to be case insensitive, or to ignore case.

    +

    To print all lines containing 'apple' as a word ('pineapple' and 'apples' will not match):

    +
    $ grep -w apple fruitlist.txt
    +

    Regular expressions can be used to match more complicated queries.

    +
    +
    +13.1.1.1. Variations
    +

    There are countless implementations and derivatives of grep available for many operating systems. Early variants of grep included egrep and fgrep. The former applies an extended regular expression syntax that was added to Unix after Ken Thompson's original regular expression implementation. The latter searches for any of a list of 'fixed' strings using the Aho-Corasick algorithm. These variants are embodied in most modern grep implementations as command-line switches (and standardized as -E and -F in POSIX). In such combined implementations, grep may also behave differently depending on the name by which it is invoked, allowing fgrep, egrep, and grep to be links to the same program.

    +

    pcregrep is an implementation of grep that uses Perl regular expression syntax.

    +

    Other commands contain the word 'grep' to indicate that they search (usually for regular expression matches). The pgrep utility, for instance, displays the processes whose names match a given regular expression.

    +
    +
    +
    +
    +

    +13.2. tr

    +

    tr (abbreviated from translate or transliterate) is a command in Unix-like operating systems.

    +

    When executed, the program reads from the standard input and writes to the standard output. It takes as parameters two sets of characters, and replaces occurrences of the characters in the first set with the corresponding elements from the other set. For example,

    +
    $ tr 'abcd' 'jkmn'
    +

    maps 'a' to 'j', 'b' to 'k', 'c' to 'm', and 'd' to 'n'.

    +

    Sets of characters may be abbreviated by using character ranges. The previous example could be written:

    +
    $ tr 'a-d' 'jkmn'
    +

    In POSIX compliant versions of tr the set represented by a character range depends on the locale's collating order, so it is safer to avoid character ranges in scripts that might be executed in a locale different from that in which they were written. Ranges can often be replaced with POSIX character sets such as [:alpha:].

    +

    The -c flag complements the first set of characters.

    +
    $ tr -cd '[:alnum:]'
    +

    therefore removes all non-alphanumeric characters.

    +

    The -s flag causes tr to compress sequences of identical adjacent characters in its output to a single token. For example,

    +
    $ tr -s '\n' '\n'
    +

    replaces sequences of one or more newline characters with a single newline.

    +

    The -d flag causes tr to delete all tokens of the specified set of characters from its input. In this case, only a single character set argument is used. The following command removes carriage return characters, thereby converting a file in DOS/Windows format to one in Unix format.

    +
    $ tr -d '\r'
    +

    Most versions of tr, including GNU tr and classic Unix tr, operate on single byte characters and are not Unicode compliant. An exception is the Heirloom Toolchest implementation, which provides basic Unicode support.

    +

    Ruby and Perl also have an internal tr operator, which operates analogously. Tcl's string map command is more general in that it maps strings to strings while tr maps characters to characters.

    +
    +
    +
    +

    +14. Elementary Regex

    +

    In computing, regular expressions provide a concise and flexible means for identifying strings of text of interest, such as particular characters, words, or patterns of characters. A regular expression (often shortened to regex or regexp) is written in a formal language that can be interpreted by a regular expression processor, a program that either serves as a parser generator or examines text and identifies parts that match the provided specification.

    +

    Regular expressions are used by many text editors, utilities, and programming languages to search and manipulate text based on patterns. For example, Perl, Ruby and Tcl have a powerful regular expression engine built directly into their syntax. Several utilities provided by Unix distributions—including the editor ed and the filter grep — were the first to popularize the concept of regular expressions.

    +

    Traditional Unix regular expression syntax followed common conventions but often differed from tool to tool. The IEEE POSIX Basic Regular Expressions (BRE) standard (released alongside an alternative flavor called Extended Regular Expressions or ERE) was designed mostly for backward compatibility with the traditional (Simple Regular Expression) syntax but provided a common standard which has since been adopted as the default syntax of many Unix regular expression tools, though there is often some variation or additional features. Many such tools also provide support for ERE syntax with command line arguments.

    +

    In the BRE syntax, most characters are treated as literals — they match only themselves (i.e., a matches "a"). The exceptions, listed below, are called metacharacters or metasequences.

    ++++
    +
    +

    +14.1. Lazy quantification

    +

    The standard quantifiers in regular expressions are greedy, meaning they match as much as they can, only giving back as necessary to match the remainder of the regex. For example, someone new to regexes wishing to find the first instance of an item between < and > symbols in this example:

    +
    Another whale explosion occurred on <January 26>, <2004>.
    +

    ...would likely come up with the pattern <.*>, or similar. However, this pattern will actually return "<January 26>, <2004>" instead of the "<January 26>" which might be expected, because the <title_reference>*</title_reference> quantifier is greedy — it will consume as many characters as possible from the input, and "January 26>, <2004" has more characters than "January 26".

    +

    Though this problem can be avoided in a number of ways (e.g., by specifying the text that is not to be matched: <[^>]*>), modern regular expression tools allow a quantifier to be specified as lazy (also known as non-greedy, reluctant, minimal, or ungreedy) by putting a question mark after the quantifier (e.g., <.*?>), or by using a modifier which reverses the greediness of quantifiers (though changing the meaning of the standard quantifiers can be confusing). By using a lazy quantifier, the expression tries the minimal match first. Though in the previous example lazy matching is used to select one of many matching results, in some cases it can also be used to improve performance when greedy matching would require more backtracking.

    +
    +
    +
    +

    +15. One Liners

    +

    A one-liner is textual input to the command-line of an operating system shell that performs some function in just one line of input.

    +

    The one liner can be

    +<block_quote>
      +
    1. An expression written in the language of the shell.

    2. +
    3. The invocation of an interpreter together with program source for the interpreter to run.

    4. +
    5. The invocation of a compiler together with source to compile and +instructions for executing the compiled program.

    6. +
    </block_quote>

    Certain dynamic scripting languages such as AWK, sed, and perl have traditionally been adept at expressing one-liners. Specialist shell interpreters such as these Unix shells or the Windows PowerShell, allow for the construction of powerful one-liners.

    +

    The use of the phrase one-liner has been widened to also include program-source for any language that does something useful in one line.

    +

    The word One-liner has two references in the index of the book The AWK Programming Language (the book is often referred to by the abbreviation TAPL). It explains the programming language AWK, which is part of the Unix operating system. The authors explain the birth of the One-liner paradigm with their daily work on early Unix machines:

    +
    “The 1977 version had only a few built-in variables and predefined functions. It was designed for writing short programs [...] Our model was that an invocation would be one or two lines long, typed in and used immediately. Defaults were chosen to match this style [...] We, being the authors, knew how the language was supposed to be used, and so we only wrote one-liners.”
    +

    Notice that this original definition of a One-liner implies immediate execution of the program without any compilation. So, in a strict sense, only source code for interpreted languages qualifies as a One-liner. But this strict understanding of a One-liner was broadened in 1985 when the IOCCC introduced the category of Best One Liner for C, which is a compiled language.

    +

    The TAPL book contains 20 examples of One-liners (A Handful of Useful awk One-Liners) at the end of the book's first chapter.

    +

    Here are the first few of them:

    +<block_quote>
      +
    1. +

      Print the total number of input lines:

      +

      END { print NR }

      +
    2. +
    3. +

      Print the tenth input line:

      +

      NR == 10

      +
    4. +
    5. +

      Print the last field of every input line:

      +

      { print $NF }

      +
    6. +
    </block_quote>

    One-liners are also used to show off the differential expressive power of programming languages. Frequently, one-liners are used to demonstrate programming ability. Contests are often held to see who can create the most exceptional one-liner.

    +

    The following example is a C program (a winning entry in the "Best one-liner" category of the IOCCC, here split to two lines for presentation).:

    +
    main(int c,char**v){return!m(v[1],v[2]);}m(char*s,char*t){return
    +*t-42?*s?63==*t|*s==*t&&m(s+1,t+1):!*t:m(s,t+1)||*s&&m(s+1,t);}
    +

    This one-liner program is a glob pattern matcher. It understands the glob characters '*' meaning 'zero or more characters' and '?' meaning exactly one character, just like most Unix shells.

    +

    Run it with two args, the string and the glob pattern. The exit status is 0 (shell true) when the pattern matches, 1 otherwise. The glob pattern must match the whole string, so you may want to use * at the beginning and end of the pattern if you are looking for something in the middle. Examples:

    +
    $ prog foo 'f??'; echo $?
    +
    +$ prog 'best short program' '??st*o**p?*'; echo $?
    +

    Here is a one line shell script to show directories:

    +
    $ ls -R | grep ":$" | sed -e 's/:$//' -e 's/[^-][^\/]*\//--/g' -e 's/^/   /' -e 's/-/|/'
    +
    +
    +
    + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/chn11Section_5.html --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/chn11Section_5.html Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1138 @@ + + + +Section 5 + + + + + + +
    +
    + +
    +
    +

    +
    +
    + +
    +

    +1. Introducing Linux

    +

    (Attribution : A significant chunk of the content under this section is based on data from Wikipedia and the Linux Documentation Project)

    +

    Linux (usually pronounced ˈlɪnəks') is a generic term referring to Unix-like computer operating systems based on the Linux kernel, where a kernel is the intermediate layer between the hardware and the applications. The kernel is, on an abstract level, the core of (most) operating systems, that manages the various system resources. The development of the Linux OS is considered the basis for Free and Open Source Software (FOSS) collaboration since typically the underlying source code can be used, modified freely, and redistributed by anyone under the terms of the GNU (a recursive acronym for "GNU's Not Unix!") Global Public License (GPL) and other free software licences. This freedom to access and reuse various components of a system, is one of the primary reasons for the popularity of Linux.

    +

    Linux is installed on a variety of computer hardware, that include mobile phones, embedded devices and supercomputers, but is infamous for its use in servers.

    +

    The name "Linux" comes from the Linux kernel, originally written in 1991 by Linus Torvalds. The rest of the system usually comprises components such as the Apache HTTP Server, the X Window System, the GNOME and KDE desktop environments, and utilities and libraries from the GNU Project (announced in 1983 by Richard Stallman). Commonly-used applications with desktop Linux systems include the Mozilla Firefox web-browser and the OpenOffice.org office application suite. The GNU contribution is the basis for the Free Software Foundation's preferred name GNU/Linux. The kernel's mascot is a penguin named "Tux". Mozilla Firefox and OpenOffice.org are open-source projects which can be run on most Operating Systems, including proprietary ones.

    +
    +

    +1.1. Historical Background

    +
    +

    +1.1.1. Events leading to the creation

    +
      +
    • The Unix operating system was developed in the 1960s and released for public use in 1970. Its accessibility and portability caused it to be widely adopted, copied and modified by academic institutions and businesses. Its design became influential to authors of other systems. Other free operating systems include the Berkeley Software Distribution (BSD), developed at the University of California at Berkeley, and MINIX which was released by Andrew S. Tanenbaum. The development and adoption of BSD and MINIX were limited due to various reasons, and this lack of a widely-adopted and free kernel triggered Linus Torvalds into starting his project.

    • +
    • In 1983, Richard Stallman started the GNU project with the goal of creating a free UNIX-like operating system. As part of this work, he wrote the GNU General Public License (GPL). By the early 1990s there was almost enough available software to create a full operating system. However, the GNU kernel, called Hurd, failed to attract enough attention from developers leaving GNU incomplete.

    • +
    +
    +
    +

    +1.1.2. The Creation of Linux

    +

    In 1991, Linus Torvalds began a project at the University of Helsinki that later became the Linux kernel. It was initially a terminal (command-line) emulator, which Torvalds used to access the large UNIX servers of the university. He wrote the program targeting just the hardware he was using and independent of an operating system because he wanted to use the functions of his computer with an 80386 processor. Development was done on Minix using the GNU C compiler. This application is still the main choice for compiling Linux today (although the code can be built with other compilers, such as the Intel C Compiler).

    +

    Torvalds continues to direct the development of the kernel. Stallman heads the Free Software Foundation, which in turn supports the GNU components. Finally, individuals and corporations develop third-party non-GNU components, which constitute a vast body of work and including kernel modules, and user applications and libraries. Linux vendors and communities combine and distribute the kernel, GNU components, and non-GNU components, with additional package management software in the form of Linux distributions.

    +
    +
    +
    +

    +1.2. Design and Implications

    +

    A Linux-based system is a modular Unix-like operating system, deriving much of its basic design from principles established in Unix earlier. Such a system uses a monolithic kernel, called the Linux kernel, which handles process control, networking, and peripheral and file system access. Device drivers are integrated directly with the kernel. Separate projects that interface with the kernel provide much of the system's higher-level functionality. The GNU userland is an important part of most Linux-based systems, providing the most common implementation of the C library, a popular shell, and many of the common Unix tools which carry out many basic operating system tasks. The graphical user interface (or GUI) used by most Linux systems is based on the "X Window System".

    +
    +

    +1.2.1. User Interface

    +

    Users can control a Linux-based system through a command line interface (or CLI), a graphical user interface (or GUI), or through controls attached to the associated hardware (this is common for embedded systems). For desktop systems, the default mode is usually the GUI. On desktop machines, "KDE", "GNOME" and "Xfce" are the most popular user interfaces,though a variety of additional user interfaces exist. Most popular user interfaces run on top of the "X Window System" (or X), which enables a graphical application running on one machine to be displayed and controlled from another in a network.

    +

    A Linux system also provides a CLI of some sort through a shell, which is the traditional way of interacting with a Unix system. A Linux distribution specialized for servers may use the CLI as its only interface. A “headless system” (system run without even a monitor) can be controlled by the command line via a remote-control protocol such as SSH or telnet. The CLI is particularly suited for automation of repetitive or delayed tasks, and provides very simple inter-process communication. A graphical terminal emulator program is often used to access the CLI from a Linux desktop.

    +
    +
    +

    +1.2.2. Development

    +

    The primary difference between Linux and many other popular contemporary operating systems is that the Linux kernel and other components are free and open source software. Linux is not the only such operating system, although it is by far the most widely used. Some free and open source software licenses are based on the principle of "copyleft", a kind of reciprocity: any work derived from a copyleft piece of software must also be copyleft itself. The most common free software license, the GNU GPL, is a form of copyleft, and is used for the Linux kernel and many of the components from the GNU project.

    +

    Linux based distributions are intended by developers for interoperability with other operating systems and established computing standards. Linux systems adhere to POSIX, SUS, ISO and ANSI standards where possible, although to date only one Linux distribution has been POSIX.1 certified, Linux-FT.Free software projects, although developed in a collaborative fashion, are often produced independently of each other. The fact that the software licenses explicitly permit redistribution, however, provides a basis for larger scale projects that collect the software produced by stand-alone projects and make it available all at once in the form of a Linux distribution.

    +

    A Linux distribution, commonly called a "distro", is a project that manages a remote collection of system software and application software packages available for download and installation through a network connection. This allows the user to adapt the operating system to his/her specific needs. Distributions are maintained by individuals, loose-knit teams, volunteer organizations, and commercial entities. A distribution can be installed using a CD that contains distribution-specific software for initial system installation and configuration. A package manager such as Synaptic or YAST allows later package upgrades and installations. A distribution is responsible for the default configuration of the installed Linux kernel, general system security, and more generally integration of the different software packages into a coherent whole.

    +
    +
    +

    +1.2.3. Community

    +

    A distribution is largely driven by its developer and user communities. Some vendors develop and fund their distributions on a volunteer basis. Examples include Debian and the Debian-based, Ubuntu. Others maintain a community version of their commercial distributions, as Red Hat does with Fedora.

    +

    In many cities and regions, local associations known as Linux Users Groups (LUGs) seek to promote their preferred distribution and by extension free software. They hold meetings and provide free demonstrations, training, technical support, and operating system installation to new users. Many Internet communities also provide support to Linux users and developers. Most distributions and free software / open source projects have IRC (Internet Relay Chat) chatrooms or newsgroups. Online forums are another means for support. Linux distributions host mailing lists; commonly there will be a specific topic such as usage or development for a given list. All these can be found simply by running an appropriate search on Google.

    +

    Although Linux distributions are generally available without charge, several large corporations sell, support, and contribute to the development of the components of the system and of free software. These include Dell, IBM, HP, Oracle, Sun Microsystems, Novell, Nokia. A number of corporations, notably Red Hat, have built their entire business around Linux distributions.

    +
    +
    +

    +1.2.4. Can I make a profit out of running a business involving Linux?

    +

    The answer is, "Yes!". The free software licenses, on which the various software packages of a distribution built on the Linux kernel are based, explicitly accommodate and encourage commercialization; the relationship between a Linux distribution as a whole and individual vendors may be seen as symbiotic. One common business model of commercial suppliers is charging for support, especially for business users. A number of companies also offer a specialized business version of their distribution, which adds proprietary support packages and tools to administer higher numbers of installations or to simplify administrative tasks. Another business model is to give away the software in order to sell hardware. Examples of corporations that are extensively (and sometimes exclusively) open-source and Linux-powered , with successful revenue generation models involving these, are Google, SUN, Mozilla, etc.

    +
    +
    +

    +1.2.5. Programming on Linux

    +

    Most Linux distributions support dozens of programming languages. The most common collection of utilities for building both Linux applications and operating system programs is found within the GNU toolchain, which includes the GNU Compiler Collection (GCC) and the GNU build system. Amongst others, GCC provides compilers for Ada, C, C++, Java, and Fortran. The Linux kernel itself is written to be compiled with GCC. Proprietary compilers for Linux include the Intel C++ Compiler, Sun Studio, and IBM XL C/C++ Compiler.

    +

    Most distributions also include support for PHP, Perl, Ruby, Python and other dynamic languages. Examples of languages that are less common, but still supported, are C# via the Mono project, sponsored by Novell, and Scheme. A number of Java Virtual Machines and development kits run on Linux, including the original Sun Microsystems JVM (HotSpot), and IBM's J2SE RE, as well as many open-source projects like Kaffe.

    +

    The two main frameworks for developing graphical applications are those of GNOME and KDE. These projects are based on the GTK+ and Qt widget toolkits, respectively, which can also be used independently of the larger framework. Both support a wide variety of languages. There are a number of Integrated Development Environments (IDEs) available including Anjuta, Code::Blocks, Eclipse, KDevelop, Lazarus, MonoDevelop, NetBeans, and Omnis Studio while the long-established editors Vim and Emacs remain popular.

    +
    +
    +
    +

    +1.3. Reasons for Using Linux

    +
    • Linux is free:

    +

    As in "free beer". Linux can be downloaded in its entirety from the Internet completely for free. No registration fees, no costs per user, free updates, and freely available source code in case you want to change the behavior of your system. +Most of all, Linux is free as in "free speech": +The license commonly used is the GNU Public License (GPL). The license says that anybody who may want to do so, has the right to change Linux and eventually to redistribute a changed version, on the one condition that the code is still available after redistribution. In practice, you are free to grab a kernel image and sell the new code, as long as your customers can still have a copy of that code.

    +
    • Linux is portable to any hardware platform:

    +

    A vendor, who wants to sell a new type of computer and who does not know what kind of OS his/her new machine will run, can take a Linux kernel and make it work on his/her hardware, because documentation related to this activity is freely available.

    +
    • Linux was made to keep on running:

    +

    As with UNIX, a Linux system expects to run without rebooting all the time. That is why a lot of tasks are being executed at night or scheduled automatically for other times, resulting in higher availability during busier periods and a more balanced use of the hardware. This property allows for Linux to be applicable to environments where people do not have the time or the possibility to control their systems constantly.

    +
    • Linux is secure and versatile:

    +

    The security model used in Linux is based on the UNIX idea of security, which is known to be robust and of proven quality. But Linux is not only safe from attacks from the Internet: it will adapt equally to other situations, utilizing the same high standards for security.

    +
    • Linux is scalable:

    +

    From a Palmtop with 2 MB of memory to a petabyte storage cluster with hundreds of nodes: add or remove the appropriate packages and Linux fits all. One does not need a supercomputer anymore,because you can use Linux to do big things using the building blocks provided with the system. If one wants to do little things, such as making an operating system for an embedded processor or just recycling your old 486, Linux will do that as well.

    +
    • The Linux OS and Linux applications have very short debug−times:

    +

    Because Linux has been developed and tested by thousands of people, both errors and people to fix them are found very quickly. It often happens that there are only a couple of hours between discovery and fixing of a bug.

    +
    +
    +
    +

    +2. Getting Started

    +
    +

    +2.1. Logging in, activating the user interface and logging out

    +

    In order to work on a Linux system directly, one needs to provide a user name and password. You always need to authenticate to the system. Most PC−based Linux systems have two basic modes for a system to run in: either quick and clean in text console mode,which includes with mouse, multitasking and multi−user features, or in graphical console mode, which looks better but eats more system resources.

    +
    +

    +2.1.1. Graphical Mode

    +

    This is the default nowadays on most desktop computers. You know you will be connecting to the system using graphical mode when you are first asked for your user name, and then to type your password.

    +

    To log in, make sure the mouse pointer is in the login window, provide your user name and password to the system and click OK or press Enter. +It is generally considered a bad idea to connect (graphically) using the root user name, the system adminstrator's account, since the use of graphics includes running a lot of extra programs, in root's case with a lot of extra permissions. To keep all risks as low as possible, use a normal user account to connect graphically. But there are enough risks to keep this in mind as a general advice, for all use of the root account: only log in as root when extra privileges are required.

    +

    After entering your user name/password combination, it can take a little while before the graphical environment is started, depending on the CPU speed of your computer, on the software you use and on your personal settings.

    +

    To continue, you will need to open a terminal window or xterm for short (X being the name for the underlying software supporting the graphical environment). This program can be found in the Applications−>Utilities->System Tools or Internet menu, depending on what window manager you are using. There might be icons that you can use as a shortcut to get an xterm window as well, and clicking the right mouse button on the desktop background will usually present you with a menu containing a terminal window application.

    +

    While browsing the menus, you will notice that a lot of things can be done without entering commands via the keyboard. For most users, the good old point−n−click method of dealing with the computer will do. But for those who want to enter the "heart" of the system, a tool stronger than a mouse will be required to handle the various tasks. This tool is the shell, and when in graphical mode, we activate our shell by opening a terminal window.

    +

    A terminal window should always show a command prompt when you open one. This terminal shows a standard prompt, which displays the user's login name, and the current working directory, represented by the twiddle (~)

    +

    Another common form for a prompt is this one: +[

    +
    +

    user@host
    +

    dir]

    +

    In the above example, user will be your login name, hosts the name of the machine you are working on, and dir an indication of your current location in the file system. Prompts can display all kinds of information, but they are not part of the commands you are giving to your system. To disconnect from the system in graphical mode, you need to close all terminal windows and other applications. After that, hit the logout icon or find Log Out in the menu. Closing everything is not really necessary, and the system can do this for you, but session management might put all currently open applications back on your screen when you connect again, which takes longer and is not always the desired effect. However, this behavior is configurable.

    +

    When you see the login screen again, asking to enter user name and password, logout was successful.

    +
    +
    +

    +2.1.2. Text Mode

    +

    One is in text mode when the whole screen is black, showing (in most cases white) characters. A text mode login screen typically shows some information about the machine you are working on, the name of the machine and a prompt waiting for you to log in.

    +

    The login is different from a graphical login, in that you have to hit the Enter key after providing your user name, because there are no buttons on the screen that you can click with the mouse. Then you should type your password, followed by another Enter. You will not see any indication that you are entering something, not even an asterisk, and you won't see the cursor move. But this is normal on Linux and is done for security +reasons.

    +

    When the system has accepted you as a valid user, you may get some more information, called the message of the day, which can be anything. Additionally, it is popular on UNIX systems to display a fortune cookie, which contains some general wise or unwise (this is up to you) thoughts. After that, you will be given a shell, indicated with the same prompt that you would get in graphical mode.

    +

    Also in text mode: log in as root only to do setup and configuration that absolutely requires administrator privileges, such as adding users, installing software packages, and performing network and other system configuration. Once you are finished, immediately leave the special account and resume your work as a non−privileged user.

    +

    Logging out is done by entering the logout command, followed by Enter. You are successfully disconnected from the system when you see the login screen again.Don't power−off the computer after logging out. It is not meant to be shut off without application of the proper procedures for halting the system. Powering it off without going through the halting process might cause severe damage!

    +
    +
    +
    +
    +

    +3. Basic Commands

    +
    +

    +3.1. ls

    +

    When invoked without any arguments, ls lists the files in the current working directory. A directory that is not the current working directory can be specified and ls will list the files there. The user also may specify any list of files and directories. In this case, all files and all contents of specified directories will be listed. The name ls is derived from list segments which was used in earlier systems.

    +

    Files whose names start with "." are not listed, unless the -a flag is specified or the files are specified explicitly.

    +

    Without options, ls displays files in a bare format. This bare format however makes it difficult to establish the type, permissions, and size of the files. The most common options to reveal this information or change the list of files are:

    +<block_quote>
      +
    • -l long format, displaying Unix file types, permissions, number of hard links, owner, group, size, date, and filename

    • +
    • -F appends a character revealing the nature of a file, for example, * for an executable, or / for a directory. Regular files have no suffix.

    • +
    • -a lists all files in the given directory, including those whose names start with "." (which are hidden files in Unix). By default, these files are excluded from the list.

    • +
    • -R recursively lists subdirectories. The command ls -R / would therefore list all files.

    • +
    • -d shows information about a symbolic link or directory, rather than about the link's target or listing the contents of a directory.

    • +
    • -t sort the list of files by modification time.

    • +
    • -h print sizes in human readable format. (e.g., 1K, 234M, 2G, etc.)

    • +
    </block_quote>

    In some environments, providing the option --color (for GNU ls) or -G (FreeBSD ls) causes ls to highlight different types of files with different colors, instead of with characters as -F would. To determine what color to use for a file, GNU ls checks the Unix file type, the file permissions, and the file extension, while FreeBSD ls checks only the Unix file type and file permissions.:

    +
    $ ls
    +jeeves.rst psmith.html blandings.html
    +$ ls -l
    +drwxr--r--   1 plum  editors   4096  jeeves
    +-rw-r--r--   1 plum  editors  30405  psmith
    +-r-xr-xr-x   1 plum  plum      8460  blandings
    +

    Here "$" actually is the beginning of the prompt. This is typical in most Unix-based systems.

    +
    +
    +

    +3.2. date

    +

    The Unix date command displays the time and date. The super-user can use it to set the system clock.

    +

    With no options, the date command displays the current date and time, including the abbreviated day name, abbreviated month name, day of the month, the time separated by colons, the timezone name, and the year. For example:

    +
    $date
    +Tue Sep  8 12:01:45 IST 2009
    +

    On some systems to set the current date and time to September 8, 2004 01:22 you type:

    +
    $date --set="20040908 01:22"
    +

    In order to view the various options for the date command, type:

    +
    $man date
    +

    This will take you to the "Manual" page comprising of all the details on the date command. You can return to the terminal from the "man" page by pressing the Esc key in the keyboard and typing ":q" in that order.

    +
    +
    +

    +3.3. cd

    +

    This stands for "change directory". When one wants to go up to the parent directory, bypassing the tree of directories one has entered, “ cd ..” can be used.

    +

    One dot '.' represents the current directory while two dots '..' represent the parent directory.

    +

    “ cd -” will return you to the previous directory (a bit like an “undo”).

    +

    You can also use cd absolute path or cd relative path (see below):

    +

    Absolute paths:

    +<block_quote>

    An “ absolute path” is easily recognised from the leading forward slash, /. The / means that you start at the top level directory and continue down.

    </block_quote>

    For example to get to /boot/grub you would type:

    +
    $cd /boot/grub
    +

    This is an absolute path because you start at the top of the hierarchy and go downwards from there (it doesn't matter where in the filesystem you were when you typed the command).

    +

    Relative paths:

    +<block_quote>

    A “ relative path” doesn't have a preceding slash. Use a relative path when you start from a directory below the top level directory structure. This is dependent on where you are in the filesystem.

    +

    For example if you are in root's home directory and want to get to /root/music, you type:

    +
    $ cd music
    </block_quote>

    Please note that there is no / using the above cd command. Using a / would cause this to be an absolute path, working from the top of the hierarchy downward.

    +
    +
    +

    +3.4. who

    +

    The standard Unix command who displays a list of users who are currently logged into a computer.

    +

    The who command is related to the command w, which provides the same information but also displays additional data and statistics.:

    +
    $who
    +beeblebrox tty7         2009-09-08 10:50 (:0)
    +beeblebrox pts/0        2009-09-08 11:25 (:0.0)
    +dumbledore pts/1        2009-09-08 18:11 (potter.xyz.in)
    +beeblebrox pts/2        2009-09-08 18:53 (:0.0)
    +

    The command can be invoked with the arguments am i or am I (so it is invoked as who am i or * who am I*), showing information about the current terminal only (see the -m option below, of which this invocation is equivalent).

    +

    In order to find out the various options that can be appended to the who command, check the man page by typing out the following in the terminal:

    +
    $man who
    +

    This will take you to the "Manual" page containing details about the who command

    +
    +
    +

    +3.5. mkdir

    +

    This command is used to make a new directory. Normal usage is as straightforward as follows:

    +
    $mkdir name_of_directory
    +

    Where name_of_directory is the name of the directory one wants to create. When typed as above (ie. normal usage), the new directory would be created within the current directory. On Unix, multiple directories can be specified, and mkdir will try to create all of them.

    +
    +

    +3.5.1. Options

    +

    On Unix-like operating systems, mkdir takes options. Three of the most common options are:

    +<block_quote>
      +
    • -p: will also create all directories leading up to the given directory that do not exist already. If the given directory already exists, ignore the error.

    • +
    • -v: display each directory that mkdir creates. Most often used with -p.

    • +
    • -m: specify the octal permissions of directories created by mkdir.

    • +
    </block_quote>

    -p is most often used when using mkdir to build up complex directory hierarchies, in case a necessary directory is missing or already there. -m is commonly used to lock down temporary directories used by shell scripts.

    +
    +
    +

    +3.5.2. Examples

    +

    An example of -p in action is:

    +
    $mkdir -p /tmp/a/b/c
    +

    If /tmp/a exists but /tmp/a/b does not, mkdir will create /tmp/a/b before creating /tmp/a/b/c.

    +

    And an even more powerful command, creating a full tree at once (this however is a Shell extension, nothing mkdir does itself):

    +
    $mkdir -p tmpdir/{trunk/sources/{includes,docs},branches,tags}
    +

    This will create:

    +<definition_list><definition_list_item><term>tmpdir - branches</term><definition>
      +
    • tag

    • +
    • <definition_list><definition_list_item><term>trunk - sources - includes</term><definition>
      • docs

      </definition>
      </definition_list_item>
      </definition_list>
    • +
    </definition>
    </definition_list_item>
    </definition_list>
    +
    +
    +
    +
    +

    +4. Getting Help

    +
    +

    +4.1. apropos and whatis

    +

    This is a command to search the manual pages files in Unix and Unix-like operating systems.

    +
    $ apropos grep
    +egrep       egrep (1)       Search a file for a pattern using full regular expressions
    +fgrep       fgrep (1)       Search a file for a fixed-character string
    +fmlgrep     fmlgrep (1)     Search a file for a pattern
    +grep        grep (1)        Search a file for a pattern
    +gzgrep      gzgrep (1)      Search a possibly compressed file for a regular expression
    +nisgrep     nismatch (1)    Utilities for searching NIS+ tables
    +pgrep       pgrep (1)       Find or signal a process by name or other attribute
    +zgrep       zgrep (1)       Search a possibly compressed file for a regular expression
    +...
    +

    In this example, the user uses apropos to search for the string "grep", and apropos returns the indicated man pages that include the term "grep".

    +

    A short index of explanations for commands is available using the whatis command, like in the examples below:

    +
    $whatis ls
    +ls (1)           - list directory contents
    +

    This displays short information about a command, and the first section in the collection of man pages that contains an appropriate page.

    +

    If you don't know where to get started and which man page to read, apropos gives more information. Say that you do not know how to start a browser, then you could enter the following command:

    +
    $apropos browser
    +gmusicbrowser (1)    - Jukebox for large collections of audio files
    +infobrowser (1)      - read Info documents
    +libsmbclient (7)     - An extension library for browsers and that               can be used...
    +opera (1)            - a standards-compliant graphical Web browser
    +sensible-browser (1) - sensible editing, paging, and web browsing
    +smbtree (1)          - A text based smb network browser
    +tvtk_doc (1)         - A GUI based TVTK documentation search browser.
    +viewres (1)          - graphical class browser for Xt
    +w3m (1)              - a text based Web browser and pager
    +www-browser (1)      - a text based Web browser and pager
    +...
    +
    +
    +

    +4.2. man

    +

    Man pages (short for "manual pages") are the extensive documentation that comes preinstalled with almost all substantial Unix and Unix-like operating systems. The Unix command used to display them is man. Each page is a self-contained document.

    +

    To read a manual page for a Unix command, one can use:

    +
    $ man <command_name>
    +

    at a shell prompt; for example, "man ftp". In order to simplify navigation through the output, man generally uses the less terminal pager.

    +

    Pages are traditionally referred to using the notation "name(section)"; for example, ftp(1). The same page name may appear in more than one section of the manual, this can occur when the names of system calls, user commands, or macro packages coincide. Two examples are man(1) and man(7), or exit(2) and exit(3). The syntax for accessing the non-default manual section varies between different man implementations. On Linux and BSD, for example, the syntax for reading *printf(3) is:

    +
    $man 3 printf
    +

    Another example:

    +
    $man man
    +

    The previous example will take you to the "Manual" page entry about manual pages!

    +
    +

    +4.2.1. Layout

    +

    All man pages follow a common layout that is optimized for presentation on a simple ASCII text display, possibly without any form of highlighting or font control. Sections present may include:

    +<definition_list><definition_list_item><term>NAME</term><definition>

    The name of the command or function, followed by a one-line description of what it does.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SYNOPSIS</term><definition>

    In the case of a command, you get a formal description of how to run it and what command line options it takes. For program functions, a list of the parameters the function takes and which header file contains its definition. For experienced users, this may be all the documentation they need.

    </definition>
    </definition_list_item>
    <definition_list_item><term>DESCRIPTION</term><definition>

    A textual description of the functioning of the command or function.

    </definition>
    </definition_list_item>
    <definition_list_item><term>EXAMPLES</term><definition>

    Some examples of common usage.

    </definition>
    </definition_list_item>
    <definition_list_item><term>SEE ALSO</term><definition>

    A list of related commands or functions.

    </definition>
    </definition_list_item>
    </definition_list>

    Other sections may be present, but these are not well standardized across man pages. Common examples include: OPTIONS, EXIT STATUS, ENVIRONMENT, KNOWN BUGS, FILES, AUTHOR, REPORTING BUGS, HISTORY and COPYRIGHT.

    +

    These days virtually every Unix command line application comes with its man page, and many Unix users perceive a lack of man pages as a sign of low quality; indeed, some projects, such as Debian, go out of their way to write man pages for programs lacking one. Few alternatives to man have enjoyed much popularity, with the possible exception of the GNU project's "info" system, an early and simple hypertext system.

    +

    However, the format of a single page for each application, the lack of classification within the sections and the relatively unsophisticated formatting facilities have motivated the development of alternative documentation systems, such as the previously mentioned "info" system.

    +

    Most Unix GUI applications (particularly those built using the GNOME and KDE development environments) now provide end-user documentation in HTML and include embedded HTML viewers such as yelp for reading the help within the application.

    +

    Usually the man pages are written in English. Translations into other languages can be also available on the system.

    +

    The default format of the man pages is troff, with either the macro package man (appearance oriented) or on some systems mdoc (semantic oriented). This makes it possible to typeset a man page to PostScript, PDF and various other formats for viewing or printing.

    +
    +
    +
    +

    +4.3. info

    +

    info is a software utility which forms a hypertextual, multipage documentation and help viewer working on a command line interface, useful when there is no GUI available.

    +

    The syntax is

    +
    $ info <command_name>
    +

    info processes info files, which are Texinfo formatted files, and presents the documentation as a tree, with simple commands to traverse the tree and to follow cross references. For instance

    +<block_quote>
      +
    • n goes to the next page.

    • +
    • p goes to the previous page.

    • +
    • u goes to the upper page.

    • +
    • l goes to the last(visited) node

    • +
    • To follow a cross reference, the cursor can be moved over a link (a word preceded by a <title_reference>*</title_reference>) and enter pressed.

    • +
    </block_quote>

    info was initially written for use with GNU/Linux and then ported to other Unix-like operating systems.

    +
    +
    +

    +4.4. --help

    +

    Most GNU commands support the --help, which gives a short explanation about how to use the command and a list of available options. Below is the output of this option with the cat command:

    +
    $ userprompt@host: cat --help
    +Usage: cat [OPTION] [FILE]...
    +Concatenate FILE(s), or standard input, to standard output.
    +
    +  -A, --show-all           equivalent to -vET
    +  -b, --number-nonblank    number nonempty output lines
    +  -e                       equivalent to -vE
    +  -E, --show-ends          display $ at end of each line
    +  -n, --number             number all output lines
    +  -s, --squeeze-blank      suppress repeated empty output lines
    +  -t                       equivalent to -vT
    +  -T, --show-tabs          display TAB characters as ^I
    +  -u                       (ignored)
    +  -v, --show-nonprinting   use ^ and M- notation, except for LFD and              TAB
    +  --help     display this help and exit
    +  --version  output version information and exit
    +
    +With no FILE, or when FILE is -, read standard input.
    +
    +Examples:
    +  cat f - g  Output f's contents, then standard input, then g's           contents.
    +  cat        Copy standard input to standard output.
    +
    +Report bugs to <bug-coreutils@gnu.org>.
    +
    +
    +
    +

    +5. Basic file handling

    +
    +

    +5.1. cp

    +

    cp is the command entered in a Unix shell to copy a file from one place to another, possibly on a different filesystem. The original file remains unchanged, and the new file may have the same or a different name.

    +
    +

    +5.1.1. Usage

    +

    To copy a file to another file:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ][ -- ] SourceFile TargetFile
    +

    To copy a file to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -r | -R ] [ -- ] SourceFile ...              TargetDirectory
    +

    To copy a directory to a directory:

    +
    $ cp [ -f ] [ -H ] [ -i ] [ -p ] [ -- ] { -r | -R }
    +SourceDirectory ... TargetDirectory
    +
    +
    +

    +5.1.2. Flags

    +

    -f (force) – specifies removal of the target file if it cannot be opened for write operations. The removal precedes any copying performed by the cp command.

    +

    -P – makes the cp command copy symbolic links. The default is to follow symbolic links, that is, to copy files to which symbolic links point.

    +

    -i (interactive) – prompts you with the name of a file to be overwritten. This occurs if the TargetDirectory or TargetFile parameter contains a file with the same name as a file specified in the SourceFile or SourceDirectory parameter. If you enter y or the locale's equivalent of y, the cp command continues. Any other answer prevents the cp command from overwriting the file.

    +

    -p (preserve) – duplicates the following characteristics of each SourceFile/SourceDirectory in the corresponding TargetFile and/or TargetDirectory:

    +<block_quote>
      +
    • The time of the last data modification and the time of the last access.

    • +
    • The user ID and group ID (only if it has permissions to do this)

    • +
    • The file permission bits and the SUID and SGID bits.

    • +
    </block_quote>

    -R (recursive) – copy directories (recursively copying all the contents)

    +
    +
    +

    +5.1.3. Examples

    +

    To make a copy of a file in the current directory, enter:

    +
    $ cp prog.c prog.bak
    +

    This copies prog.c to prog.bak. If the prog.bak file does not already exist, the cp command creates it. If it does exist, the cp command replaces it with a copy of the prog.c file.

    +

    To copy a file in your current directory into another directory, enter:

    +
    $ cp zaphod /home/books/hhgg
    +

    This copies the jones file to /home/books/hhgg/zaphod.

    +

    To copy a file to a new file and preserve the modification date, time, and access control list associated with the source file, enter:

    +
    $ cp -p martin_luther_king martin_luther_king.jr
    +

    This copies the martin_luther_king file to the martin_luther_king.jr file. Instead of creating the file with the current date and time stamp, the system gives the martin_luther_king.jr file the same date and time as the martin_luther_king file. The martin_luther_king.jr file also inherits the martin_luther_king file's access control protection.

    +

    To copy all the files in a directory to a new directory, enter:

    +
    $ cp /home/galactica/clients/* /home/hhgg/customers
    +

    This copies only the files in the clients directory to the customers directory.

    +

    To copy a directory, including all its files and subdirectories, to another directory, enter:

    +<block_quote>

    $ cp -R /home/hhgg/clients /home/hhgg/customers

    </block_quote>

    This copies the clients directory, including all its files, subdirectories, and the files in those subdirectories, to the customers/clients directory.

    +

    To copy a specific set of files of any extension to another directory, enter:

    +
    $ cp zaphod arthur ford /home/hhgg/clients
    +

    This copies the zaphod, arthur, and ford files in your current working directory to the /home/hhgg/clients directory.

    +

    To use pattern-matching characters to copy files, enter:

    +
    $ cp programs/*.py .
    +

    This copies the files in the programs directory that end with .py to the current directory, signified by the single "." (dot). You must type a space between the py and the final dot.

    +
    +
    +
    +

    +5.2. mv

    +

    mv (short for move) is a Unix command that moves one or more files or directories from one place to another. The original file is deleted, and the new file may have the same or a different name. If possible (i.e. when the original and new files are on the same file system), mv will rename the file instead. Write permission is required on all directories being modified.

    +
    +

    +5.2.1. Conflicting existing file

    +

    In all cases, when a file is moved to have the name of an existing file (in the same directory), the existing file is deleted. If the existing file is not writable but is in a directory that is writable, then the mv command asks for confirmation if possible (i.e. if run from a terminal) before proceeding, unless the -f (force) option is used.

    +
    +
    +

    +5.2.2. Differences with copy and delete

    +

    Note that, usually, when moving files within the same volume, moving (and/or renaming) is not the same as simply copying and then deleting the original. When moving a file, the link is simply removed from the old parent directory and added to the new parent directory. However, the file itself is untouched (i.e. it has the same inodes and resides at the same place on the disk). For example, you cannot copy a file you cannot read, but you can move (and/or rename) it (provided you have write permission to its old and new parent directories). Also, suppose there is a non-empty directory you do not have write permission to. You cannot delete this directory (since you cannot delete its contents); but you can move (and/or rename) it. Also, since moving between filenames on a single volume does not involve copying, it is faster and does not place strain of lots of reads and writes on the disk. Moving files across different volumes, however, does necessitate copying and deleting.

    +
    +
    +

    +5.2.3. Examples

    +
    $ mv myfile mynewfilename    renames a file
    +$ mv myfile otherfilename    renames a file and deletes the existing            file "myfile"
    +$ mv myfile /myfile          moves 'myfile' from the current            directory to the root directory
    +$ mv myfile dir/myfile       moves 'myfile' to 'dir/myfile' relative            to the current directory
    +$ mv myfile dir              same as the previous command (the          filename is implied to be the same)
    +$ mv myfile dir/myfile2      moves 'myfile' to dir and renames it to            'myfile2'
    +$ mv foo bar baz dir         moves multiple files to directory dir
    +$ mv --help                  shows a very concise help about the                syntax of the command
    +$ man mv                     prints an extensive user manual for                'mv' in the terminal
    +

    In all cases, the file or files being moved or renamed can be a directory.

    +

    Note that when the command is called with two arguments (as mv name1 name2 or mv name1 /dir/name2), it can have three different effects, depending on whether name2 does not exist, is an existing file, or is an existing directory. If the user intends to refer to an existing directory, /. (or in some Unix versions / is sufficient) may be appended to the name to force the system to check this. To move a file to a new directory, the directory must be created first.

    +
    +
    +
    +

    +5.3. rm

    +

    rm (short for "remove") is one of several basic Unix command lines that operates on files. It is used to delete files from a filesystem. The data is not actually destroyed. Only the index listing where the file is stored is destroyed, and the storage is made available for reuse. There are undelete utilities that will attempt to reconstruct the index and can bring the file back if the parts were not reused.

    +

    Here's example to remove a file named "foo" from a directory, here shown with the -i option:

    +
    $ rm -i foo
    +remove foo? y
    +
    +

    +5.3.1. Options

    +

    Common options that rm accepts include:

    +<block_quote>
      +
    • -r, which removes directories, removing the contents recursively beforehand (so as not to leave files without a directory to reside in) ("recursive")

    • +
    • -i, which asks for every deletion to be confirmed ("interactive")

    • +
    • -f, which ignores non-existent files and overrides any confirmation prompts ("force")

    • +
    • -v, which shows what is being removed as it happens ("verbose")

    • +
    </block_quote>

    rm is often aliased to "rm -i" so as to avoid accidental deletion of files. If a user still wishes to delete a large number of files without confirmation, they can manually cancel out the -i argument by adding the -f option (as the option specified later on the expanded command line "rm -i -f" takes precedence).

    +

    rm -rf (variously, rm -rf /, rm -rf <title_reference>*</title_reference>, and others) is frequently used in jokes and anecdotes about Unix disasters. The rm -rf variant of the command, if run by a superuser on the root directory, would cause the contents of every writable mounted filesystem on the computer to be deleted.

    +

    rm is often used in conjunction with xargs to supply a list of files to delete:

    +
    xargs rm < filelist
    +

    When rm is used on a symbolic link, it deletes the link, but does not affect the target of the link.

    +
    +
    +

    +5.3.2. Permissions

    +

    Usually, on most filesystems, deleting a file requires write permission on the parent directory (and execute permission, in order to enter the directory in the first place). (Note that, confusingly for beginners, permissions on the file itself are irrelevant. However, GNU rm asks for confirmation if a write-protected file is to be deleted, unless the -f option is used.)

    +

    To delete a directory (with rm -r), one must delete all of its contents recursively. This requires that one must have read and write and execute permission to that directory (if it's not empty) and all non-empty subdirectories recursively (if there are any). The read permissions are needed to list the contents of the directory in order to delete them. This sometimes leads to an odd situation where a non-empty directory cannot be deleted because one doesn't have write permission to it and so cannot delete its contents; but if the same directory were empty, one would be able to delete it.

    +

    If a file resides in a directory with the sticky bit set, then deleting the file requires one to be the owner of the file.

    +
    +
    +
    +
    +

    +6. Command Line Arguments

    +

    In computer command line interfaces, a command line argument is an argument sent to a program being called. In general, a program can take any number of command line arguments, which may be necessary for the program to run, or may even be ignored, depending on the function of that program.

    +

    For example, in Unix and Unix-like environments, an example of a command-line argument is:

    +
    rm file.s
    +

    "file.s" is a command line argument which tells the program rm to remove the file "file.s".

    +

    Programming languages such as C, C++ and Java allow a program to interpret the command line arguments by handling them as string parameters in the main function.

    +

    A command line option or simply option (also known as a command line parameter, flag, or a switch) is an indication by a user that a computer program should change its default output.

    +

    Long options are introduced via "--", and are typically whole words. For example, ls --long --classify --all. Arguments to long options are provided with "=", as ls --block-size=1024. Some Unix programs use long options with single dashes, for example MPlayer as in mplayer -nosound.

    +

    Linux also uses "--" to terminate option lists. For example, an attempt to delete a file called -file1 by using rm -file1 may produce an error, since rm may interpret -file1 as a command line switch. Using rm -- -file1 removes ambiguity.

    +
    +
    +

    +7. Basic Text Processing

    +
    +

    +7.1. head

    +

    head is a program on Unix and Unix-like systems used to display the first few lines of a text file or piped data. The command syntax is:

    +
    $ head [options] <file_name>
    +

    By default, head will print the first 10 lines of its input to the standard output. The number of lines printed may be changed with a command line option. The following example shows the first 20 lines of filename:

    +
    $ head -n 20 filename
    +

    This displays the first 5 lines of all files starting with foo:

    +
    $ head -n 5 foo*
    +

    Some versions omit the n and just let you say -5.

    +
    +

    +7.1.1. Flags

    +
    -c <x number of bytes> Copy first x number of bytes.
    +

    Other options: sed

    +

    Many early versions of Unix did not have this command, and so documentation and books had sed do this job:

    +
    sed 5q foo
    +

    This says to print every line (implicit), and quit after the fifth.

    +
    +
    +
    +

    +7.2. tail

    +

    tail is a program on Unix and Unix-like systems used to display the last few lines of a text file or piped data.

    +

    The command-syntax is:

    +
    $ tail [options] <file_name>
    +

    By default, tail will print the last 10 lines of its input to the standard output. With command line options the number of lines printed and the printing units (lines, blocks or bytes) may be changed. The following example shows the last 20 lines of filename:

    +
    $ tail -n 20 filename
    +

    This example shows the last 15 bytes of all files starting with foo:

    +
    $ tail -c 15 foo*
    +

    This example shows all lines of filename from the second line onwards:

    +
    $ tail -n +2 filename
    +

    Using an older syntax (still used in Sun Solaris as the -n option is not supported), the last 20 lines and the last 50 bytes of filename can be shown with the following command:

    +
    $ tail -20 filename
    +$ tail -50c filename
    +

    However this syntax is now obsolete and does not conform with the POSIX 1003.1-2001 standard. Even if still supported in current versions, when used with other options (like -f, see below), these switches could not work at all.

    +
    +

    +7.2.1. File monitoring

    +

    tail has a special command line option -f (follow) that allows a file to be monitored. Instead of displaying the last few lines and exiting, tail displays the lines and then monitors the file. As new lines are added to the file by another process, tail updates the display. This is particularly useful for monitoring log files. The following command will display the last 10 lines of messages and append new lines to the display as new lines are added to messages:

    +
    $ tail -f /var/adm/messages
    +

    To interrupt tail while it is monitoring, break-in with Ctrl+C. This command can be run "in the background" with &, see job control.

    +

    If you have a command's result to monitor, you can use the watch command.

    +
    +
    +
    +

    +7.3. cut

    +

    In computing, cut is a Unix command line utility which is used to extract sections from each line of input — usually from a file.

    +

    Extraction of line segments can typically be done by bytes (-b), characters (-c), or fields (-f) separated by a delimiter (-d — the tab character by default). A range must be provided in each case which consists of one of N, N-M, N- (N to the end of the line), or -M (beginning of the line to M), where N and M are counted from 1 (there is no zeroth value). Since version 6, an error is thrown if you include a zeroth value. Prior to this the value was ignored and assumed to be 1.

    +

    Assuming a file named file containing the lines:

    +
    foo:bar:baz:qux:quux
    +one:two:three:four:five:six:seven
    +alpha:beta:gamma:delta:epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +

    To output the fourth through tenth characters of each line:

    +
    $ cut -c 4-10 file
    +

    This gives the output:

    +
    :bar:ba
    +:two:th
    +ha:beta
    +

    To output the fifth field through the end of the line of each line using the colon character as the field delimiter:

    +
    $ cut -d : -f 5- file
    +

    This gives the output:

    +
    quux
    +five:six:seven
    +epsilon:zeta:eta:teta:iota:kappa:lambda:mu
    +
    +
    +

    +7.4. paste

    +

    paste is a Unix command line utility which is used to join files horizontally (parallel merging) by outputting lines consisting of the sequentially corresponding lines of each file specified, separated by tabs, to the standard output. It is effectively the horizontal equivalent to the utility cat command which operates on the vertical plane of two or more files.

    +

    To paste several columns of data together into the file www from files who, where, and when:

    +
    $ paste who where when > www
    +

    If the files contain:

    +++++
    +

    This creates the file named www containing:

    +
    Batman            GothamCity       January 3
    +Trillian          Andromeda        February 4
    +Jeeves            London           March 19
    +
    +
    +
    +

    +8. Shell Meta Characters

    +

    Unix recognizes certain special characters, called "meta characters," as command directives. The shell meta characters are recognized anywhere they appear in the command line, even if they are not surrounded by blank space. For that reason, it is safest to only use the characters A-Z, a-z, 0-9, and the period, dash, and underscore characters when naming files and directories on Unix. If your file or directory has a shell meta character in the name, you will find it difficult to use the name in a shell command.

    +

    The shell meta characters include:

    +

    / < > ! $ % ^ & * | { } [ ] " ' ` ~ ;

    +

    Different shells may differ in the meta characters recognized.

    +

    As an example,

    +
    $ ls file.*
    +

    run on a directory containing the files file, file.c, file.lst, and myfile would list the files file.c and file.lst. However,:

    +
    $ ls file.?
    +

    run on the same directory would only list file.c because the ? only matches one character, no more, no less. This can save you a great deal of typing time. For example, if there is a file called california_cornish_hens_with_wild_rice and no other files whose names begin with 'c', you could view the file without typing the whole name by typing this:

    +
    $ more c*
    +

    because the c* matches that long file name.

    +

    Filenames containing metacharacters can pose many problems and should never be intentionally created. If you do find that you've created a file with metacharacters, and you would like to remove it, you have three options. You may use wildcards to match metacharacter, use the to directly enter the filename, or put the command in double quotes (except in the case of double quotes within the file name, these must be captured with one of the first two methods). For example, deleting a file named <title_reference>"``*`|more</title_reference>"` can be accomplished with:

    +
    $ rm ??more
    +

    or:

    +
    $ rm $\backslash$*$\backslash$|more
    +

    or:

    +
    $ rm ''*|more''
    +
    +
    +

    +9. Looking At Files

    +
    +

    +9.1. cat

    +

    The cat command is a standard Unix program used to concatenate and display files. The name is from "catenate", a synonym of concatenate.

    +

    The Single Unix Specification specifies the behavior that the contents of each of the files given in sequence as arguments will be written to the standard output in the same sequence, and mandates one option, -u, where each byte is printed as it is read.

    +

    If the filename is specified as -, then cat will read from standard input at that point in the sequence. If no files are specified, cat will read from standard input entered.

    +
    +

    +9.1.1. Jargon File Definition

    +

    The Jargon File version 4.4.7 lists this as the definition of cat:

    +
    1. To spew an entire file to the screen or some other output sink without
    +     pause (syn. blast).
    +
    +2. By extension, to dump large amounts of data at an unprepared target or
    +     with no intention of browsing it carefully. Usage: considered silly.
    +     Rare outside Unix sites. See also dd, BLT.
    +
    +     Among Unix fans, *cat(1)* is considered an excellent example of
    +     user-interface design, because it delivers the file contents without
    +     such verbosity as spacing or headers between the files, and because
    +     it does not require the files to consist of lines of text, but works
    +     with any sort of data.
    +
    +     Among Unix critics, *cat(1)* is considered the canonical example of
    +     bad user-interface design, because of its woefully unobvious name.
    +     It is far more often used to blast a single file to standard output
    +     than to concatenate two or more files. The name cat for the former
    +     operation is just as unintuitive as, say, LISP's cdr.
    +
    +     Of such oppositions are holy wars made...
    +
    +
    +

    +9.1.2. Useless Use of 'cat'

    +

    UUOC (from comp.unix.shell on Usenet) stands for “Useless Use of cat”. As it is observed on comp.unix.shell, “The purpose of cat is to concatenate (or 'catenate') files. If it's only one file, concatenating it with nothing at all is a waste of time, and costs you a process.”

    +

    Nevertheless one sees people doing:

    +
    $ cat file | some_command and its args ...
    +

    instead of the equivalent and cheaper:

    +
    <file some_command and its args ...
    +

    or (equivalently and more classically):

    +
    some_command and its args ... <file
    +

    Since 1995, occasional awards for UUOC have been given out. The activity of fixing instances of UUOC is sometimes called 'demoggification'.

    +

    Amongst many, it is still considered safer to use cat for such cases given that the < and > keys are next to each other in many popular keyboard mappings. While the risk might be low, the impact of using > instead of < can be high and prohibitive.

    +
    +
    +

    +9.1.3. zcat

    +

    zcat is a Unix program similar to cat, that decompresses individual files and concatenates them to standard output. Traditionally zcat operated on files compressed by compress but today it is usually able to operate on gzip or even bzip2 archives. On such systems, it is equivalent to gunzip -c

    +
    +
    +
    +

    +9.2. more

    +

    In computing, more is a command to view (but not modify) the contents of a text file one screen at a time (terminal pager). It is available on Unix and Unix-like systems, DOS, OS/2 and Microsoft Windows. Programs of this sort are called pagers.

    +
    +

    +9.2.1. Usage

    +

    The command-syntax is:

    +
    $ more [options] [file_name]
    +

    If no file name is provided, more looks for input from stdin.

    +

    Once more has obtained input, it displays as much as can fit on the current screen and waits for user input to advance, with the exception that a form feed (^L) will also cause more to wait at that line, regardless of the amount of text on the screen. In the lower-left corner of the screen is displayed the text "--More--" and a percentage, representing the percent of the file that more has paged through. (This percentage includes the text displayed on the current screen.) When more reaches the end of a file (100%) it exits. The most common methods of navigating through a file are Enter, which advances the output by one line, and Space, which advances the output by one screen.

    +

    There are also other commands that can be used while navigating through the document; consult more's man page for more details.

    +

    Options are typically entered before the file name, but can also be entered in the environment variable $MORE. Options entered in the actual command line will override those entered in the $MORE environment variable. Available options may vary between Unix systems.

    +
    +
    +
    +

    +9.3. less

    +

    less is a terminal pager program on Unix, Windows and Unix-like systems used to view (but not change) the contents of a text file one screen at a time. It is similar to more, but has the extended capability of allowing both forward and backward navigation through the file. Unlike most Unix text editors/viewers, less does not need to read the entire file before starting, resulting in faster load times with large files.

    +
    +

    +9.3.1. Usage

    +

    less can be invoked with options to change its behaviour, for example, the number of lines to display on the screen. A few options vary depending on the operating system. While less is displaying the file, various commands can be used to navigate through the file. These commands are based on those used by both more and vi. It is also possible to search for character patterns in the file.

    +

    By default, less displays the contents of the file to the standard output (one screen at a time). If the file name argument is omitted, it displays the contents from standard input (usually the output of another command through a pipe). If the output is redirected to anything other than a terminal, for example a pipe to another command, less behaves like cat.

    +

    The command-syntax is:

    +
    $ less [options] file_name
    +
    +
    +

    +9.3.2. Frequently Used Options

    +<block_quote>
      +
    • -g: Highlights just the current match of any searched string.

    • +
    • -I: Case-insensitive searches.

    • +
    • -M: Shows more detailed prompt, including file position.

    • +
    • -N: Shows line numbers (useful for source code viewing).

    • +
    • -S: Disables line wrap ("chop long lines"). Long lines can be seen by side scrolling.

    • +
    • -?: Shows help.

    • +
    </block_quote>
    +
    +
    +

    +9.3.3. Frequently Used Commands

    +<block_quote>
      +
    • [Arrows]/[Page Up]/[Page Down]/[Home]/[End]: Navigation.

    • +
    • [Space bar]: Next page.

    • +
    • b: Previous page.

    • +
    • ng: Jump to line number n. Default is the start of the file.

    • +
    • nG: Jump to line number n. Default is the end of the file.

    • +
    • /pattern: Search for pattern. Regular expressions can be used.

    • +
    • '^ or g: Go to start of file.

    • +
    • '$ or G: Go to end of file.

    • +
    • s: Save current content (got from another program like grep) in a file.

    • +
    • =: File information.

    • +
    • h: Help.

    • +
    • q: Quit.

    • +
    </block_quote>
    +
    +
    +

    +9.3.4. Examples

    +
    $ less -M readme.txt                     #Read "readme.txt."
    +$ less +F /var/log/mail.log              #Follow mode for log
    +$ file * | less                          #Easier file analysis.
    +$ grep -i void *.c | less -I -p void     #Case insensitive search                                                         for "void" in all .c files
    +
    +
    +
    +
    +

    +10. Directory Structure

    +

    In the File Hierarchy Standard (FHS) all files and directories appear under the root directory "/", even if they are stored on different physical devices. Note however that some of these directories may or may not be present on a Unix system depending on whether certain subsystems, such as the X Window System, are installed.

    +

    The majority of these directories exist in all UNIX operating systems and are generally used in much the same way; however, the descriptions here are those used specifically for the FHS, and are not considered authoritative for platforms other than Linux.

    ++++
    +
    +

    +10.1. man hier

    +

    This is the manual page on the UNIX filesystem. The syntax for this is:

    +
    $ man hier
    +
    +
    +

    +10.2. ls -l

    +

    Shows you huge amounts of information (permissions, owners, size, and when last modified) for folders and files. The syntax is

    +
    $ ls -l
    +

    This can be done after entering the required directory.

    +
    +
    +
    +

    +11. Permissions and Ownership

    +
    +

    +11.1. chmod

    +

    The chmod command (abbreviated from 'change mode') is a shell command and C language function in Unix and Unix-like environments. When executed, it can change file system modes of files and directories. The modes include permissions and special modes.A chmod command first appeared in AT&T Unix version 1, and is still used today on Unix-like machines.

    +
    +

    +11.1.1. Usage

    +

    The chmod command options are specified like this:

    +
    $ chmod [options] mode[,mode] file1 [file2 ...]
    +

    To view what the permissions currently are, type:

    +
    $ ls -l file
    +
    +
    +

    +11.1.2. Command line options

    +

    The chmod command has a number of command line options that affect its behavior. The most common options are:

    +<block_quote>
      +
    • -R: Changes the modes of directories and files recursively

    • +
    • -v: Verbose mode; lists all files as they are being processed

    • +
    </block_quote>
    +
    +11.1.2.1. Symbolic modes
    +

    To the chmod utility, all permissions and special modes are represented by its mode parameter. One way to adjust the mode of files or directories is to specify a symbolic mode. The symbolic mode is composed of three components, which are combined to form a single string of text:

    +
    $ chmod [references][operator][modes] file1 ...
    +

    The references (or classes) are used to distinguish the users to whom the permissions apply. If no references are specified it defaults to “all” but modifies only the permissions allowed by the umask. The references are represented by one or more of the following letters:

    +++++
    +

    The chmod program uses an operator to specify how the modes of a file should be adjusted. The following operators are accepted:

    ++++
    +

    The modes indicate which permissions are to be granted or taken away from the specified classes. There are three basic modes which correspond to the basic permissions:

    +++++
    +

    The combination of these three components produces a string that is understood by the chmod command. Multiple changes can be specified by separating multiple symbolic modes with commas.

    +
    +
    +
    +11.1.2.2. Symbolic examples
    +

    Add the 'read' and 'write' permissions to the 'user' and 'group' classes of a directory:

    +
    $ chmod ug+rw mydir
    +$ ls -ld mydir
    +drw-rw----   2 starwars  yoda  96 Dec 8 12:53 mydir
    +

    For a file, remove write permissions for all classes:

    +
    $ chmod a-w myfile
    +$ ls -l myfile
    +-r-xr-xr-x   2 starwars  yoda 96 Dec 8 12:53 myfile
    +

    Set the permissions for the u*ser and the *g*roup to read and execute only (no write permission) on *mydir.

    +
    $ chmod ug=rx mydir
    +$ ls -ld mydir
    +dr-xr-x---   2 starwars  yoda 96 Dec 8 12:53 mydir
    +
    +
    +
    +11.1.2.3. Octal numbers
    +

    The chmod command also accepts three and four-digit octal numbers representing modes. Using a three-digit octal number to set the modes of a file named myfile :

    +
    $ chmod 664 myfile
    +$ ls -l myfile
    +-rw-rw-r--  1   57 Jul  3 10:13  myfile
    +

    Since the setuid, setgid and sticky bits are not set, this is equivalent to:

    +
    $ chmod 0664 myfile
    +
    +
    +
    +11.1.2.4. Special modes
    +

    The chmod command is also capable of changing the additional permissions or special modes of a file or directory. The symbolic modes use s to represent the setuid and setgid modes, and t to represent the sticky mode. The modes are only applied to the appropriate classes, regardless of whether or not other classes are specified.

    +

    Most operating systems support the specification of special modes using octal modes, but some do not. On these systems, only the symbolic modes can be used.

    +
    +
    +
    +
    +
    +

    +12. Redirection and Piping

    +

    In computing, redirection is a function common to most command-line interpreters, including the various Unix shells that can redirect standard streams to user-specified locations.

    +

    Programs do redirection with the dup2(2) system call, or its less-flexible but higher-level stdio analogues, freopen(3) and popen(3).

    +
    +

    +12.1. Redirecting standard input and standard output

    +

    Redirection is usually implemented by placing certain characters between commands. Typically, the syntax of these characters is as follows:

    +
    $ command1 > file1
    +

    executes command1, placing the output in file1. Note that this will truncate any existing data in file1. To append output to the end of the file, use the >> operator.:

    +
    $ command1 < file1
    +

    executes command1, using file1 as the source of input (as opposed to the keyboard).:

    +
    $ command1 < infile > outfile
    +

    combines the two capabilities: command1 reads from infile and writes to outfile

    +
    +
    +

    +12.2. Piping

    +

    Programs can be run together such that one program reads the output from another with no need for an explicit intermediate file: +A pipeline of three programs run on a text terminal:

    +
    $ command1 | command2
    +

    executes command1, using its output as the input for command2 (commonly called piping, since the "|" character is known as a "pipe").

    +

    This is equivalent to using two redirects and a temporary file:

    +
    $ command1 > tempfile
    +$ command2 < tempfile
    +$ rm tempfile
    +

    A good example for command piping is combining echo with another command to achieve something interactive in a non-interactive shell, e.g.:

    +
    $ echo -e "user\npass" | ftp localhost
    +

    This runs the ftp client with input user, press return, then pass.

    +
    +
    +

    +12.3. Redirecting to and from the standard file handles

    +

    In Unix shells derived from the original Bourne shell, the first two actions can be further modified by placing a number (the file descriptor) immediately before the character; this will affect which stream is used for the redirection. The Unix standard I/O streams are:

    +++++
    +

    For example:

    +
    $ command1 2> file1
    +

    executes command1, directing the standard error stream to file1.

    +

    In shells derived from csh (the C shell), the syntax instead appends the & character to the redirect characters, thus achieving a similar result.

    +

    Another useful capability is to redirect one standard file handle to another. The most popular variation is to merge standard error into standard output so error messages can be processed together with (or alternately to) the usual output. Example:

    +
    $ find / -name .profile > results 2>&1
    +

    will try to find all files named .profile. Executed without redirection, it will output hits to stdout and errors (e.g. for lack of privilege to traverse protected directories) to stderr. If standard output is directed to file results, error messages appear on the console. To see both hits and error messages in file results, merge stderr (handle 2) into stdout (handle 1) using 2>&1 .

    +

    It's possible use 2>&1 before ">" but it doesn't work. In fact, when the interpreter reads 2>&1, it doesn't know yet where standard output is redirected and then standard error isn't merged.

    +

    If the merged output is to be piped into another program, the file merge sequence 2>&1 must precede the pipe symbol, thus:

    +
    $ find / -name .profile 2>&1 | less
    +

    A simplified form of the command:

    +
    $ command > file 2>&1
    +

    is:

    +
    $ command &>file
    +

    or:

    +
    $command >&file
    +
    +
    +

    +12.4. Chained pipelines

    +

    The redirection and piping tokens can be chained together to create complex commands. For example:

    +
    $ ls | grep '\.sh' | sort > shlist
    +

    lists the contents of the current directory, where this output is filtered to only contain lines which contain .sh, sort this resultant output lexicographically, and place the final output in shlist. This type of construction is used very commonly in shell scripts and batch files.

    +
    +
    +

    +12.5. Redirect to multiple outputs

    +

    The standard command tee can redirect output from a command to several destinations.

    +
    $ ls -lrt | tee xyz
    +

    This directs the file list output to both standard output as well as to the file xyz.

    +
    +
    +
    +

    +13. More Text Processing

    +
    +

    +13.1. grep

    +

    grep is a command line text search utility originally written for Unix. The name is taken from the first letters in global / regular expression / print, a series of instructions for the ed text editor. The grep command searches files or standard input globally for lines matching a given regular expression, and prints them to the program's standard output.

    +
    +

    +13.1.1. Usage

    +

    This is an example of a common grep usage:

    +
    $ grep apple fruitlist.txt
    +

    In this case, grep prints all lines containing 'apple' from the file fruitlist.txt, regardless of word boundaries; therefore lines containing 'pineapple' or 'apples' are also printed. The grep command is case sensitive by default, so this example's output does not include lines containing 'Apple' (with a capital A) unless they also contain 'apple'.

    +

    Like most Unix commands, grep accepts command line arguments to change this and many other behaviors. For example:

    +
    $ grep -i apple fruitlist.txt
    +

    This prints all lines containing 'apple' regardless of capitalization. The '-i' argument tells grep to be case insensitive, or to ignore case.

    +

    To print all lines containing 'apple' as a word ('pineapple' and 'apples' will not match):

    +
    $ grep -w apple fruitlist.txt
    +

    Regular expressions can be used to match more complicated queries.

    +
    +
    +13.1.1.1. Variations
    +

    There are countless implementations and derivatives of grep available for many operating systems. Early variants of grep included egrep and fgrep. The former applies an extended regular expression syntax that was added to Unix after Ken Thompson's original regular expression implementation. The latter searches for any of a list of 'fixed' strings using the Aho-Corasick algorithm. These variants are embodied in most modern grep implementations as command-line switches (and standardized as -E and -F in POSIX). In such combined implementations, grep may also behave differently depending on the name by which it is invoked, allowing fgrep, egrep, and grep to be links to the same program.

    +

    pcregrep is an implementation of grep that uses Perl regular expression syntax.

    +

    Other commands contain the word 'grep' to indicate that they search (usually for regular expression matches). The pgrep utility, for instance, displays the processes whose names match a given regular expression.

    +
    +
    +
    +
    +

    +13.2. tr

    +

    tr (abbreviated from translate or transliterate) is a command in Unix-like operating systems.

    +

    When executed, the program reads from the standard input and writes to the standard output. It takes as parameters two sets of characters, and replaces occurrences of the characters in the first set with the corresponding elements from the other set. For example,

    +
    $ tr 'abcd' 'jkmn'
    +

    maps 'a' to 'j', 'b' to 'k', 'c' to 'm', and 'd' to 'n'.

    +

    Sets of characters may be abbreviated by using character ranges. The previous example could be written:

    +
    $ tr 'a-d' 'jkmn'
    +

    In POSIX compliant versions of tr the set represented by a character range depends on the locale's collating order, so it is safer to avoid character ranges in scripts that might be executed in a locale different from that in which they were written. Ranges can often be replaced with POSIX character sets such as [:alpha:].

    +

    The -c flag complements the first set of characters.

    +
    $ tr -cd '[:alnum:]'
    +

    therefore removes all non-alphanumeric characters.

    +

    The -s flag causes tr to compress sequences of identical adjacent characters in its output to a single token. For example,

    +
    $ tr -s '\n' '\n'
    +

    replaces sequences of one or more newline characters with a single newline.

    +

    The -d flag causes tr to delete all tokens of the specified set of characters from its input. In this case, only a single character set argument is used. The following command removes carriage return characters, thereby converting a file in DOS/Windows format to one in Unix format.

    +
    $ tr -d '\r'
    +

    Most versions of tr, including GNU tr and classic Unix tr, operate on single byte characters and are not Unicode compliant. An exception is the Heirloom Toolchest implementation, which provides basic Unicode support.

    +

    Ruby and Perl also have an internal tr operator, which operates analogously. Tcl's string map command is more general in that it maps strings to strings while tr maps characters to characters.

    +
    +
    +
    +

    +14. Elementary Regex

    +

    In computing, regular expressions provide a concise and flexible means for identifying strings of text of interest, such as particular characters, words, or patterns of characters. A regular expression (often shortened to regex or regexp) is written in a formal language that can be interpreted by a regular expression processor, a program that either serves as a parser generator or examines text and identifies parts that match the provided specification.

    +

    Regular expressions are used by many text editors, utilities, and programming languages to search and manipulate text based on patterns. For example, Perl, Ruby and Tcl have a powerful regular expression engine built directly into their syntax. Several utilities provided by Unix distributions—including the editor ed and the filter grep — were the first to popularize the concept of regular expressions.

    +

    Traditional Unix regular expression syntax followed common conventions but often differed from tool to tool. The IEEE POSIX Basic Regular Expressions (BRE) standard (released alongside an alternative flavor called Extended Regular Expressions or ERE) was designed mostly for backward compatibility with the traditional (Simple Regular Expression) syntax but provided a common standard which has since been adopted as the default syntax of many Unix regular expression tools, though there is often some variation or additional features. Many such tools also provide support for ERE syntax with command line arguments.

    +

    In the BRE syntax, most characters are treated as literals — they match only themselves (i.e., a matches "a"). The exceptions, listed below, are called metacharacters or metasequences.

    ++++
    +
    +

    +14.1. Lazy quantification

    +

    The standard quantifiers in regular expressions are greedy, meaning they match as much as they can, only giving back as necessary to match the remainder of the regex. For example, someone new to regexes wishing to find the first instance of an item between < and > symbols in this example:

    +
    Another whale explosion occurred on <January 26>, <2004>.
    +

    ...would likely come up with the pattern <.*>, or similar. However, this pattern will actually return "<January 26>, <2004>" instead of the "<January 26>" which might be expected, because the <title_reference>*</title_reference> quantifier is greedy — it will consume as many characters as possible from the input, and "January 26>, <2004" has more characters than "January 26".

    +

    Though this problem can be avoided in a number of ways (e.g., by specifying the text that is not to be matched: <[^>]*>), modern regular expression tools allow a quantifier to be specified as lazy (also known as non-greedy, reluctant, minimal, or ungreedy) by putting a question mark after the quantifier (e.g., <.*?>), or by using a modifier which reverses the greediness of quantifiers (though changing the meaning of the standard quantifiers can be confusing). By using a lazy quantifier, the expression tries the minimal match first. Though in the previous example lazy matching is used to select one of many matching results, in some cases it can also be used to improve performance when greedy matching would require more backtracking.

    +
    +
    +
    +

    +15. One Liners

    +

    A one-liner is textual input to the command-line of an operating system shell that performs some function in just one line of input.

    +

    The one liner can be

    +<block_quote>
      +
    1. An expression written in the language of the shell.

    2. +
    3. The invocation of an interpreter together with program source for the interpreter to run.

    4. +
    5. The invocation of a compiler together with source to compile and +instructions for executing the compiled program.

    6. +
    </block_quote>

    Certain dynamic scripting languages such as AWK, sed, and perl have traditionally been adept at expressing one-liners. Specialist shell interpreters such as these Unix shells or the Windows PowerShell, allow for the construction of powerful one-liners.

    +

    The use of the phrase one-liner has been widened to also include program-source for any language that does something useful in one line.

    +

    The word One-liner has two references in the index of the book The AWK Programming Language (the book is often referred to by the abbreviation TAPL). It explains the programming language AWK, which is part of the Unix operating system. The authors explain the birth of the One-liner paradigm with their daily work on early Unix machines:

    +
    “The 1977 version had only a few built-in variables and predefined functions. It was designed for writing short programs [...] Our model was that an invocation would be one or two lines long, typed in and used immediately. Defaults were chosen to match this style [...] We, being the authors, knew how the language was supposed to be used, and so we only wrote one-liners.”
    +

    Notice that this original definition of a One-liner implies immediate execution of the program without any compilation. So, in a strict sense, only source code for interpreted languages qualifies as a One-liner. But this strict understanding of a One-liner was broadened in 1985 when the IOCCC introduced the category of Best One Liner for C, which is a compiled language.

    +

    The TAPL book contains 20 examples of One-liners (A Handful of Useful awk One-Liners) at the end of the book's first chapter.

    +

    Here are the first few of them:

    +<block_quote>
      +
    1. +

      Print the total number of input lines:

      +

      END { print NR }

      +
    2. +
    3. +

      Print the tenth input line:

      +

      NR == 10

      +
    4. +
    5. +

      Print the last field of every input line:

      +

      { print $NF }

      +
    6. +
    </block_quote>

    One-liners are also used to show off the differential expressive power of programming languages. Frequently, one-liners are used to demonstrate programming ability. Contests are often held to see who can create the most exceptional one-liner.

    +

    The following example is a C program (a winning entry in the "Best one-liner" category of the IOCCC, here split to two lines for presentation).:

    +
    main(int c,char**v){return!m(v[1],v[2]);}m(char*s,char*t){return
    +*t-42?*s?63==*t|*s==*t&&m(s+1,t+1):!*t:m(s,t+1)||*s&&m(s+1,t);}
    +

    This one-liner program is a glob pattern matcher. It understands the glob characters '*' meaning 'zero or more characters' and '?' meaning exactly one character, just like most Unix shells.

    +

    Run it with two args, the string and the glob pattern. The exit status is 0 (shell true) when the pattern matches, 1 otherwise. The glob pattern must match the whole string, so you may want to use * at the beginning and end of the pattern if you are looking for something in the middle. Examples:

    +
    $ prog foo 'f??'; echo $?
    +
    +$ prog 'best short program' '??st*o**p?*'; echo $?
    +

    Here is a one line shell script to show directories:

    +
    $ ls -R | grep ":$" | sed -e 's/:$//' -e 's/[^-][^\/]*\//--/g' -e 's/^/   /' -e 's/-/|/'
    +
    +
    +
    + diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/bad-merge-1.png Binary file SEESenv/web/html/support/figs/bad-merge-1.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/bad-merge-2.png Binary file SEESenv/web/html/support/figs/bad-merge-2.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/bad-merge-3.png Binary file SEESenv/web/html/support/figs/bad-merge-3.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/bad-merge-4.png Binary file SEESenv/web/html/support/figs/bad-merge-4.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/bad-merge-5.png Binary file SEESenv/web/html/support/figs/bad-merge-5.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/caution.png Binary file SEESenv/web/html/support/figs/caution.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/feature-branches.png Binary file SEESenv/web/html/support/figs/feature-branches.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/filelog.png Binary file SEESenv/web/html/support/figs/filelog.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/kdiff3.png Binary file SEESenv/web/html/support/figs/kdiff3.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/metadata.png Binary file SEESenv/web/html/support/figs/metadata.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/mq-stack.png Binary file SEESenv/web/html/support/figs/mq-stack.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/note.png Binary file SEESenv/web/html/support/figs/note.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/revlog.png Binary file SEESenv/web/html/support/figs/revlog.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/rss.png Binary file SEESenv/web/html/support/figs/rss.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/snapshot.png Binary file SEESenv/web/html/support/figs/snapshot.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/tip.png Binary file SEESenv/web/html/support/figs/tip.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/tour-history.png Binary file SEESenv/web/html/support/figs/tour-history.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/tour-merge-conflict.png Binary file SEESenv/web/html/support/figs/tour-merge-conflict.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/tour-merge-merge.png Binary file SEESenv/web/html/support/figs/tour-merge-merge.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/tour-merge-pull.png Binary file SEESenv/web/html/support/figs/tour-merge-pull.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/tour-merge-sep-repos.png Binary file SEESenv/web/html/support/figs/tour-merge-sep-repos.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/undo-manual-merge.png Binary file SEESenv/web/html/support/figs/undo-manual-merge.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/undo-manual.png Binary file SEESenv/web/html/support/figs/undo-manual.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/undo-non-tip.png Binary file SEESenv/web/html/support/figs/undo-non-tip.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/undo-simple.png Binary file SEESenv/web/html/support/figs/undo-simple.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/wdir-after-commit.png Binary file SEESenv/web/html/support/figs/wdir-after-commit.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/wdir-branch.png Binary file SEESenv/web/html/support/figs/wdir-branch.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/wdir-merge.png Binary file SEESenv/web/html/support/figs/wdir-merge.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/wdir-pre-branch.png Binary file SEESenv/web/html/support/figs/wdir-pre-branch.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/figs/wdir.png Binary file SEESenv/web/html/support/figs/wdir.png has changed diff -r 672eaaab9204 -r 52d12eb31c30 SEESenv/web/html/support/form-min.js --- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/SEESenv/web/html/support/form-min.js Fri Feb 12 01:11:21 2010 +0530 @@ -0,0 +1,1 @@ +(function($){$.fn.ajaxSubmit=function(_2){if(typeof _2=="function"){_2={success:_2};}_2=$.extend({url:this.attr("action")||window.location,type:this.attr("method")||"GET"},_2||{});var _3={};$.event.trigger("form.pre.serialize",[this,_2,_3]);if(_3.veto){return this;}var a=this.formToArray(_2.semantic);if(_2.data){for(var n in _2.data){a.push({name:n,value:_2.data[n]});}}if(_2.beforeSubmit&&_2.beforeSubmit(a,this,_2)===false){return this;}$.event.trigger("form.submit.validate",[a,this,_2,_3]);if(_3.veto){return this;}var q=$.param(a);if(_2.type.toUpperCase()=="GET"){_2.url+=(_2.url.indexOf("?")>=0?"&":"?")+q;_2.data=null;}else{_2.data=q;}var _7=this,callbacks=[];if(_2.resetForm){callbacks.push(function(){_7.resetForm();});}if(_2.clearForm){callbacks.push(function(){_7.clearForm();});}if(!_2.dataType&&_2.target){var _8=_2.success||function(){};callbacks.push(function(_9){if(this.evalScripts){$(_2.target).attr("innerHTML",_9).evalScripts().each(_8,arguments);}else{$(_2.target).html(_9).each(_8,arguments);}});}else{if(_2.success){callbacks.push(_2.success);}}_2.success=function(_a,_b){for(var i=0,max=callbacks.length;i");var io=$io[0];var op8=$.browser.opera&&window.opera.version()<9;if($.browser.msie||op8){io.src="javascript:false;document.write(\"\");";}$io.css({position:"absolute",top:"-1000px",left:"-1000px"});var xhr={responseText:null,responseXML:null,status:0,statusText:"n/a",getAllResponseHeaders:function(){},getResponseHeader:function(){},setRequestHeader:function(){}};var g=_11.global;if(g&&!$.active++){$.event.trigger("ajaxStart");}if(g){$.event.trigger("ajaxSend",[xhr,_11]);}var _18=0;var _19=0;setTimeout(function(){$io.appendTo("body");io.attachEvent?io.attachEvent("onload",cb):io.addEventListener("load",cb,false);var _1a=_10.encoding?"encoding":"enctype";var t=_7.attr("target");_7.attr({target:id,method:"POST",action:_11.url});_10[_1a]="multipart/form-data";if(_11.timeout){setTimeout(function(){_19=true;cb();},_11.timeout);}_10.submit();_7.attr("target",t);},10);function cb(){if(_18++){return;}io.detachEvent?io.detachEvent("onload",cb):io.removeEventListener("load",cb,false);var ok=true;try{if(_19){throw "timeout";}var _1d,doc;doc=io.contentWindow?io.contentWindow.document:io.contentDocument?io.contentDocument:io.document;xhr.responseText=doc.body?doc.body.innerHTML:null;xhr.responseXML=doc.XMLDocument?doc.XMLDocument:doc;if(_11.dataType=="json"||_11.dataType=="script"){var ta=doc.getElementsByTagName("textarea")[0];_1d=ta?ta.value:xhr.responseText;if(_11.dataType=="json"){eval("data = "+_1d);}else{$.globalEval(_1d);}}else{if(_11.dataType=="xml"){_1d=xhr.responseXML;if(!_1d&&xhr.responseText!=null){_1d=toXml(xhr.responseText);}}else{_1d=xhr.responseText;}}}catch(e){ok=false;$.handleError(_11,xhr,"error",e);}if(ok){_11.success(_1d,"success");if(g){$.event.trigger("ajaxSuccess",[xhr,_11]);}}if(g){$.event.trigger("ajaxComplete",[xhr,_11]);}if(g&&!--$.active){$.event.trigger("ajaxStop");}if(_11.complete){_11.complete(xhr,ok?"success":"error");}setTimeout(function(){$io.remove();xhr.responseXML=null;},100);}function toXml(s,doc){if(window.ActiveXObject){doc=new ActiveXObject("Microsoft.XMLDOM");doc.async="false";doc.loadXML(s);}else{doc=(new DOMParser()).parseFromString(s,"text/xml");}return (doc&&doc.documentElement&&doc.documentElement.tagName!="parsererror")?doc:null;}}};$.fn.ajaxSubmit.counter=0;$.fn.ajaxForm=function(_21){return this.ajaxFormUnbind().submit(submitHandler).each(function(){this.formPluginId=$.fn.ajaxForm.counter++;$.fn.ajaxForm.optionHash[this.formPluginId]=_21;$(":submit,input:image",this).click(clickHandler);});};$.fn.ajaxForm.counter=1;$.fn.ajaxForm.optionHash={};function clickHandler(e){var _23=this.form;_23.clk=this;if(this.type=="image"){if(e.offsetX!=undefined){_23.clk_x=e.offsetX;_23.clk_y=e.offsetY;}else{if(typeof $.fn.offset=="function"){var _24=$(this).offset();_23.clk_x=e.pageX-_24.left;_23.clk_y=e.pageY-_24.top;}else{_23.clk_x=e.pageX-this.offsetLeft;_23.clk_y=e.pageY-this.offsetTop;}}}setTimeout(function(){_23.clk=_23.clk_x=_23.clk_y=null;},10);}function submitHandler(){var id=this.formPluginId;var _26=$.fn.ajaxForm.optionHash[id];$(this).ajaxSubmit(_26);return false;}$.fn.ajaxFormUnbind=function(){this.unbind("submit",submitHandler);return this.each(function(){$(":submit,input:image",this).unbind("click",clickHandler);});};$.fn.formToArray=function(_27){var a=[];if(this.length==0){return a;}var _29=this[0];var els=_27?_29.getElementsByTagName("*"):_29.elements;if(!els){return a;}for(var i=0,max=els.length;i= 0 ? '&' : '?') + q; + options.data = null; // data is null for 'get' + } + else + options.data = q; // data is the query string for 'post' + + var $form = this, callbacks = []; + if (options.resetForm) callbacks.push(function() { $form.resetForm(); }); + if (options.clearForm) callbacks.push(function() { $form.clearForm(); }); + + // perform a load on the target only if dataType is not provided + if (!options.dataType && options.target) { + var oldSuccess = options.success || function(){}; + callbacks.push(function(data) { + if (this.evalScripts) + $(options.target).attr("innerHTML", data).evalScripts().each(oldSuccess, arguments); + else // jQuery v1.1.4 + $(options.target).html(data).each(oldSuccess, arguments); + }); + } + else if (options.success) + callbacks.push(options.success); + + options.success = function(data, status) { + for (var i=0, max=callbacks.length; i < max; i++) + callbacks[i](data, status, $form); + }; + + // are there files to upload? + var files = $('input:file', this).fieldValue(); + var found = false; + for (var j=0; j < files.length; j++) + if (files[j]) + found = true; + + if (options.iframe || found) // options.iframe allows user to force iframe mode + fileUpload(); + else + $.ajax(options); + + // fire 'notify' event + $.event.trigger('form.submit.notify', [this, options]); + return this; + + + // private function for handling file uploads (hat tip to YAHOO!) + function fileUpload() { + var form = $form[0]; + var opts = $.extend({}, $.ajaxSettings, options); + + var id = 'jqFormIO' + $.fn.ajaxSubmit.counter++; + var $io = $('