330pics-shell scripts_fourth
来源:互联网 发布:织梦cms模板安装 编辑:程序博客网 时间:2024/04/29 04:26
***@@@grp.sh@@@!!!************************************************************************************
#!/bin/bash
# grp.sh: 一个非常粗糙的'grep'命令的实现.
E_BADARGS=65
if [ -z "$1" ] # 检查传递给脚本的参数.
then
echo "Usage: `basename $0` pattern"
exit $E_BADARGS
fi
echo
for file in * # 遍历$PWD下的所有文件.
do
output=$(sed -n /"$1"/p $file) # 命令替换.
if [ ! -z "$output" ] # 如果"$output"不加双引号将会发生什么?
then
echo -n "$file: "
echo $output
fi # sed -ne "/$1/s|^|${file}: |p" 这句与上边这段等价.
echo
done
echo
exit 0
# 练习:
# -----
# 1) 在任何给定的文件中,如果有超过一个匹配的话, 在输出中添加新行.
# 2) 添加一些特征.
%%%&&&grp.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@hanoi.bash@@@!!!************************************************************************************
#! /bin/bash
#
# 汉诺塔
# Bash script
# Copyright (C) 2000 Amit Singh. All Rights Reserved.
# http://hanoi.kernelthread.com
#
# 在bash version 2.05b.0(13)-release下通过测试
#
# 经过脚本原作者同意
#+ 可以使用在"Advanced Bash Scripting Guide"中.
# 本书作者对此脚本做了少许修改.
#=================================================================#
# 汉诺塔是由Edouard Lucas提出的数学谜题,
#+ 他是19世纪的法国数学家.
#
# 有三个直立的柱子竖在地面上.
# 第一个柱子上有一组盘子套在上面.
# 这些盘子是平的, 中间有孔,
#+ 可以套在柱子上面.
# 这些盘子的直径不同, 它们从下到上,
#+ 按照尺寸递减的顺序摆放.
# 也就是说, 最小的在最上边, 最大的在最下面.
#
# 现在的任务是要把这组盘子
#+ 从一个柱子上全部搬到另一个柱子上.
# 你每次只能将一个盘子从一个柱子移动到另一个柱子上.
# 你也可以把盘子从其他的柱子上移回到原来的柱子上.
# 你只能把小的盘子放到大的盘子上,
#+ 反过来就*不*行.
# 切记, 这是规则, 绝对不能把大盘子放到小盘子的上面.
#
# 如果盘子的数量比较少, 那么移不了几次就能完成.
#+ 但是随着盘子数量的增加,
#+ 移动次数几乎成倍的增长,
#+ 而且移动的"策略"也会变得越来越复杂.
#
# 想了解更多信息的话, 请访问http://hanoi.kernelthread.com.
#
#
# ... ... ...
# | | | | | |
# _|_|_ | | | |
# |_____| | | | |
# |_______| | | | |
# |_________| | | | |
# |___________| | | | |
# | | | | | |
# .--------------------------------------------------------------.
# |**************************************************************|
# #1 #2 #3
#
#=================================================================#
E_NOPARAM=66 # 没有参数传给脚本.
E_BADPARAM=67 # 传给脚本的盘子个数不符合要求.
Moves= # 保存移动次数的全局变量.
# 这里修改了原来的脚本.
dohanoi() { # 递归函数.
case $1 in
0)
;;
*)
dohanoi "$(($1-1))" $2 $4 $3
echo move $2 "-->" $3
let "Moves += 1" # 这里修改了原脚本.
dohanoi "$(($1-1))" $4 $3 $2
;;
esac
}
case $# in
1)
case $(($1>0)) in # 至少要有一个盘子.
1)
dohanoi $1 1 3 2
echo "Total moves = $Moves"
exit 0;
;;
*)
echo "$0: illegal value for number of disks";
exit $E_BADPARAM;
;;
esac
;;
*)
echo "usage: $0 N"
echo " Where /"N/" is the number of disks."
exit $E_NOPARAM;
;;
esac
# 练习:
# -----
# 1) 这个位置以下的代码会不会被执行?
# 为什么不? (容易)
# 2) 解释一下这个运行的"dohanoi"函数的运行原理.
# (比较难)
%%%&&&hanoi.bash&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@ha.sh@@@!!!************************************************************************************
#!/bin/bash
# $Id: ha.sh,v 1.2 2005/04/21 23:24:26 oliver Exp $
# Copyright 2005 Oliver Beckstein
# Released under the GNU Public License
# Author of script granted permission for inclusion in ABS Guide.
# (Thank you!)
#----------------------------------------------------------------
# pseudo hash based on indirect parameter expansion
# API: access through functions:
#
# create the hash:
#
# newhash Lovers
#
# add entries (note single quotes for spaces)
#
# addhash Lovers Tristan Isolde
# addhash Lovers 'Romeo Montague' 'Juliet Capulet'
#
# access value by key
#
# gethash Lovers Tristan ----> Isolde
#
# show all keys
#
# keyshash Lovers ----> 'Tristan' 'Romeo Montague'
#
#
# convention: instead of perls' foo{bar} = boing' syntax,
# use
# '_foo_bar=boing' (two underscores, no spaces)
#
# 1) store key in _NAME_keys[]
# 2) store value in _NAME_values[] using the same integer index
# The integer index for the last entry is _NAME_ptr
#
# NOTE: No error or sanity checks, just bare bones.
function _inihash () {
# private function
# call at the beginning of each procedure
# defines: _keys _values _ptr
#
# usage: _inihash NAME
local name=$1
_keys=_${name}_keys
_values=_${name}_values
_ptr=_${name}_ptr
}
function newhash () {
# usage: newhash NAME
# NAME should not contain spaces or '.';
# actually: it must be a legal name for a bash variable
# We rely on bash automatically recognising arrays.
local name=$1
local _keys _values _ptr
_inihash ${name}
eval ${_ptr}=0
}
function addhash () {
# usage: addhash NAME KEY 'VALUE with spaces'
# arguments with spaces need to be quoted with single quotes ''
local name=$1 k="$2" v="$3"
local _keys _values _ptr
_inihash ${name}
#echo "DEBUG(addhash): ${_ptr}=${!_ptr}"
eval let ${_ptr}=${_ptr}+1
eval "$_keys[${!_ptr}]=/"${k}/""
eval "$_values[${!_ptr}]=/"${v}/""
}
function gethash () {
# usage: gethash NAME KEY
# returns boing
# ERR=0 if entry found, 1 otherwise
# Thats not a proper hash---we simply linearly search through the keys
local name=$1 key="$2"
local _keys _values _ptr
local k v i found h
_inihash ${name}
# _ptr holds the highest index in the hash
found=0
for i in $(seq 1 ${!_ptr}); do
h="/${${_keys}[${i}]}" # safer to do it in two steps
eval k=${h} # (especially when quoting for spaces)
if [ "${k}" = "${key}" ]; then found=1; break; fi
done;
[ ${found} = 0 ] && return 1;
# else: i is the index that matches the key
h="/${${_values}[${i}]}"
eval echo "${h}"
return 0;
}
function keyshash () {
# usage: keyshash NAME
# returns list of all keys defined for hash name
local name=$1 key="$2"
local _keys _values _ptr
local k i h
_inihash ${name}
# _ptr holds the highest index in the hash
for i in $(seq 1 ${!_ptr}); do
h="/${${_keys}[${i}]}" # Safer to do it in two steps
eval k=${h} # (especially when quoting for spaces)
echo -n "'${k}' "
done;
}
# --------------------------------------------------------------------
# Now, let's test it.
# (Per comments at the beginning of the script.)
newhash Lovers
addhash Lovers Tristan Isolde
addhash Lovers 'Romeo Montague' 'Juliet Capulet'
# Output results.
echo
gethash Lovers Tristan # Isolde
echo
keyshash Lovers # 'Tristan' 'Romeo Montague'
echo; echo
exit 0
# Exercise: Add error checks to the functions.
%%%&&&ha.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@hash-example.sh@@@!!!************************************************************************************
#!/bin/bash
# hash-example.sh: Colorizing text.
# Author: Mariusz Gniazdowski <mgniazd-at-gmail.com>
. Hash.lib # Load the library of functions.
hash_set colors red "/033[0;31m"
hash_set colors blue "/033[0;34m"
hash_set colors light_blue "/033[1;34m"
hash_set colors light_red "/033[1;31m"
hash_set colors cyan "/033[0;36m"
hash_set colors light_green "/033[1;32m"
hash_set colors light_gray "/033[0;37m"
hash_set colors green "/033[0;32m"
hash_set colors yellow "/033[1;33m"
hash_set colors light_purple "/033[1;35m"
hash_set colors purple "/033[0;35m"
hash_set colors reset_color "/033[0;00m"
# $1 - keyname
# $2 - value
try_colors() {
echo -en "$2"
echo "This line is $1."
}
hash_foreach colors try_colors
hash_echo colors reset_color -en
echo -e '/nLet us overwrite some colors with yellow./n'
# It's hard to read yellow text on some terminals.
hash_dup colors yellow red light_green blue green light_gray cyan
hash_foreach colors try_colors
hash_echo colors reset_color -en
echo -e '/nLet us delete them and try colors once more . . ./n'
for i in red light_green blue green light_gray cyan; do
hash_unset colors $i
done
hash_foreach colors try_colors
hash_echo colors reset_color -en
hash_set other txt "Other examples . . ."
hash_echo other txt
hash_get_into other txt text
echo $text
hash_set other my_fun try_colors
hash_call other my_fun purple "`hash_echo colors purple`"
hash_echo colors reset_color -en
echo; echo "Back to normal?"; echo
exit $?
# On some terminals, the "light" colors print in bold,
# and end up looking darker than the normal ones.
# Why is this?
%%%&&&hash-example.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@hello.sh@@@!!!************************************************************************************
#!/bin/bash
# hello.sh: 显示"hello"还是"goodbye"
#+ 依赖于脚本是如何被调用的.
# 在当前目录下($PWD)为这个脚本创建一个链接:
# ln -s hello.sh goodbye
# 现在, 通过如下两种方法来调用这个脚本:
# ./hello.sh
# ./goodbye
HELLO_CALL=65
GOODBYE_CALL=66
if [ $0 = "./goodbye" ]
then
echo "Good-bye!"
# 当然, 在这里你也可以添加一些其他的goodbye类型的命令.
exit $GOODBYE_CALL
fi
echo "Hello!"
# 当然, 在这里你也可以添加一些其他的hello类型的命令.
exit $HELLO_CALL
%%%&&&hello.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@here-function.sh@@@!!!************************************************************************************
#!/bin/bash
# here-function.sh
GetPersonalData ()
{
read firstname
read lastname
read address
read city
read state
read zipcode
} # 这个函数看起来就是一个交互函数, 但是...
# 给上边的函数提供输入.
GetPersonalData <<RECORD001
Bozo
Bozeman
2726 Nondescript Dr.
Baltimore
MD
21226
RECORD001
echo
echo "$firstname $lastname"
echo "$address"
echo "$city, $state $zipcode"
echo
exit 0
%%%&&&here-function.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@hexconvert.sh@@@!!!************************************************************************************
#!/bin/bash
# hexconvert.sh: 将10进制数字转换为16进制数字.
E_NOARGS=65 # 缺少命令行参数错误.
BASE=16 # 16进制.
if [ -z "$1" ]
then
echo "Usage: $0 number"
exit $E_NOARGS
# 需要一个命令行参数.
fi
# 练习: 添加命令行参数检查.
hexcvt ()
{
if [ -z "$1" ]
then
echo 0
return # 如果没有参数传递到这个函数中的话就"return" 0.
fi
echo ""$1" "$BASE" o p" | dc
# "o" 设置输出的基数(数制).
# "p" 打印栈顶.
# 参考dc的man页来了解其他的选项.
return
}
hexcvt "$1"
exit 0
%%%&&&hexconvert.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@horserace.sh@@@!!!************************************************************************************
#!/bin/bash
# horserace.sh: 一个非常简单的模拟赛马的游戏.
# 作者: Stefano Palmeri
# 经过授权可以在本书中使用.
################################################################
# 脚本目的:
# 使用转义序列和终端颜色进行游戏.
#
# 练习:
# 编辑脚本, 使它运行起来更具随机性,
#+ 建立一个假的赌场 . . .
# 嗯 . . . 嗯 . . . 这种开场让我联想起一部电影 . . .
#
# 脚本将会给每匹马分配一个随机障碍.
# 按照马的障碍来计算几率,
#+ 并且使用一种欧洲(?)的风格表达出来.
# 比如: 几率(odds)=3.75的话, 那就意味着如果你押$1,
#+ 你就会赢得$3.75.
#
# 此脚本已经在GNU/Linux操作系统上测试过,
#+ 测试终端有xterm, rxvt, 和konsole.
# 测试机器上安装的是AMD 900 MHz处理器,
#+ 平均比赛时间为75秒.
# 如果使用更快的机器, 那么比赛用时会更少.
# 所以, 如果你想增加比赛的悬念, 可以重置变量USLEEP_ARG.
#
# 本脚本由Stefano Palmeri编写.
################################################################
E_RUNERR=65
# 检查一下md5sum和bc是否已经被安装.
if ! which bc &> /dev/null; then
echo bc is not installed.
echo "Can/'t run . . . "
exit $E_RUNERR
fi
if ! which md5sum &> /dev/null; then
echo md5sum is not installed.
echo "Can/'t run . . . "
exit $E_RUNERR
fi
# 设置下面的变量将会降低脚本的运行速度.
# 它会作为参数, 传递给usleep命令(man usleep),
#+ 并且单位是微秒(500000微秒 = 半秒).
USLEEP_ARG=0
# 如果脚本被Ctl-C中断, 那就清除临时目录,
#+ 恢复终端光标和终端颜色.
trap 'echo -en "/E[?25h"; echo -en "/E[0m"; stty echo;/
tput cup 20 0; rm -fr $HORSE_RACE_TMP_DIR' TERM EXIT
# 请参考与调试相关的章节, 可以获得'trap'命令的详细用法.
# 为脚本设置一个唯一的(实际上不是绝对唯一)临时目录名.
HORSE_RACE_TMP_DIR=$HOME/.horserace-`date +%s`-`head -c10 /dev/urandom | md5sum | head -c30`
# 创建临时目录, 并移动到该目录下.
mkdir $HORSE_RACE_TMP_DIR
cd $HORSE_RACE_TMP_DIR
# 这个函数将会把光标移动到行为$1, 列为$2的位置上, 然后打印$3.
# 例如: "move_and_echo 5 10 linux"等价与"tput cup 4 9; echo linux",
#+ 但是使用一个命令代替了两个命令.
# 注意: "tput cup"定义0 0位置, 为终端左上角,
#+ 而echo定义1 1位置, 为终端左上角.
move_and_echo() {
echo -ne "/E[${1};${2}H""$3"
}
# 此函数用来产生1-9之间的伪随机数.
random_1_9 () {
head -c10 /dev/urandom | md5sum | tr -d [a-z] | tr -d 0 | cut -c1
}
# 在画马的时候, 这两个函数用来模拟"移动".
draw_horse_one() {
echo -n " "//$MOVE_HORSE//
}
draw_horse_two(){
echo -n " "////$MOVE_HORSE////
}
# 定义当前终端尺寸.
N_COLS=`tput cols`
N_LINES=`tput lines`
# 至少需要一个20(行) X 80(列)的终端. 检查一下.
if [ $N_COLS -lt 80 ] || [ $N_LINES -lt 20 ]; then
echo "`basename $0` needs a 80-cols X 20-lines terminal."
echo "Your terminal is ${N_COLS}-cols X ${N_LINES}-lines."
exit $E_RUNERR
fi
# 开始画赛场.
# 需要一个80字符的字符串. 见下面.
BLANK80=`seq -s "" 100 | head -c80`
clear
# 将前景色与背景色设为白.
echo -ne '/E[37;47m'
# 将光标移动到终端的左上角.
tput cup 0 0
# 画6条白线.
for n in `seq 5`; do
echo $BLANK80 # 用这个80字符的字符串将终端变为彩色的.
done
# 将前景色设为黑色.
echo -ne '/E[30m'
move_and_echo 3 1 "START 1"
move_and_echo 3 75 FINISH
move_and_echo 1 5 "|"
move_and_echo 1 80 "|"
move_and_echo 2 5 "|"
move_and_echo 2 80 "|"
move_and_echo 4 5 "| 2"
move_and_echo 4 80 "|"
move_and_echo 5 5 "V 3"
move_and_echo 5 80 "V"
# 将前景色设置为红色.
echo -ne '/E[31m'
# 一些ASCII的艺术效果.
move_and_echo 1 8 "..@@@..@@@@@...@@@@@.@...@..@@@@..."
move_and_echo 2 8 ".@...@...@.......@...@...@.@......."
move_and_echo 3 8 ".@@@@@...@.......@...@@@@@.@@@@...."
move_and_echo 4 8 ".@...@...@.......@...@...@.@......."
move_and_echo 5 8 ".@...@...@.......@...@...@..@@@@..."
move_and_echo 1 43 "@@@@...@@@...@@@@..@@@@..@@@@."
move_and_echo 2 43 "@...@.@...@.@.....@.....@....."
move_and_echo 3 43 "@@@@..@@@@@.@.....@@@@...@@@.."
move_and_echo 4 43 "@..@..@...@.@.....@.........@."
move_and_echo 5 43 "@...@.@...@..@@@@..@@@@.@@@@.."
# 将前景色和背景色设为绿色.
echo -ne '/E[32;42m'
# 画11行绿线.
tput cup 5 0
for n in `seq 11`; do
echo $BLANK80
done
# 将前景色设为黑色.
echo -ne '/E[30m'
tput cup 5 0
# 画栅栏.
echo "++++++++++++++++++++++++++++++++++++++/
++++++++++++++++++++++++++++++++++++++++++"
tput cup 15 0
echo "++++++++++++++++++++++++++++++++++++++/
++++++++++++++++++++++++++++++++++++++++++"
# 将前景色和背景色设回白色.
echo -ne '/E[37;47m'
# 画3条白线.
for n in `seq 3`; do
echo $BLANK80
done
# 将前景色设为黑色.
echo -ne '/E[30m'
# 创建9个文件, 用来保存障碍物.
for n in `seq 10 7 68`; do
touch $n
done
# 将脚本所要画的"马"设置为第一种类型.
HORSE_TYPE=2
# 为每匹"马"创建位置文件和几率文件.
#+ 在这些文件中, 保存马的当前位置,
#+ 类型和几率.
for HN in `seq 9`; do
touch horse_${HN}_position
touch odds_${HN}
echo /-1 > horse_${HN}_position
echo $HORSE_TYPE >> horse_${HN}_position
# 给马定义随机障碍物.
HANDICAP=`random_1_9`
# 检查函数random_1_9是否返回一个有效值.
while ! echo $HANDICAP | grep [1-9] &> /dev/null; do
HANDICAP=`random_1_9`
done
# 给马定义最后一个障碍物的位置.
LHP=`expr $HANDICAP /* 7 + 3`
for FILE in `seq 10 7 $LHP`; do
echo $HN >> $FILE
done
# 计算几率.
case $HANDICAP in
1) ODDS=`echo $HANDICAP /* 0.25 + 1.25 | bc`
echo $ODDS > odds_${HN}
;;
2 | 3) ODDS=`echo $HANDICAP /* 0.40 + 1.25 | bc`
echo $ODDS > odds_${HN}
;;
4 | 5 | 6) ODDS=`echo $HANDICAP /* 0.55 + 1.25 | bc`
echo $ODDS > odds_${HN}
;;
7 | 8) ODDS=`echo $HANDICAP /* 0.75 + 1.25 | bc`
echo $ODDS > odds_${HN}
;;
9) ODDS=`echo $HANDICAP /* 0.90 + 1.25 | bc`
echo $ODDS > odds_${HN}
esac
done
# 打印几率.
print_odds() {
tput cup 6 0
echo -ne '/E[30;42m'
for HN in `seq 9`; do
echo "#$HN odds->" `cat odds_${HN}`
done
}
# 在起跑线上把马画出来.
draw_horses() {
tput cup 6 0
echo -ne '/E[30;42m'
for HN in `seq 9`; do
echo ///$HN///" "
done
}
print_odds
echo -ne '/E[47m'
# 等待按下回车键, 按下之后就开始比赛.
# 转义序列'/E[?25l'禁用光标.
tput cup 17 0
echo -e '/E[?25l'Press [enter] key to start the race...
read -s
# 禁用了终端的常规echo功能.
# 这么做用来避免在比赛中,
#+ 按键所导致的"花"屏.
stty -echo
# --------------------------------------------------------
# 开始比赛.
draw_horses
echo -ne '/E[37;47m'
move_and_echo 18 1 $BLANK80
echo -ne '/E[30m'
move_and_echo 18 1 Starting...
sleep 1
# 设置终点线的列号.
WINNING_POS=74
# 定义比赛开始的时间.
START_TIME=`date +%s`
# 下面的"while"结构需要使用COL变量.
COL=0
while [ $COL -lt $WINNING_POS ]; do
MOVE_HORSE=0
# 检查random_1_9函数是否返回了有效值.
while ! echo $MOVE_HORSE | grep [1-9] &> /dev/null; do
MOVE_HORSE=`random_1_9`
done
# 定义"随机抽取的马"的原来类型和位置.
HORSE_TYPE=`cat horse_${MOVE_HORSE}_position | tail -1`
COL=$(expr `cat horse_${MOVE_HORSE}_position | head -1`)
ADD_POS=1
# 判断当前位置是否存在障碍物.
if seq 10 7 68 | grep -w $COL &> /dev/null; then
if grep -w $MOVE_HORSE $COL &> /dev/null; then
ADD_POS=0
grep -v -w $MOVE_HORSE $COL > ${COL}_new
rm -f $COL
mv -f ${COL}_new $COL
else ADD_POS=1
fi
else ADD_POS=1
fi
COL=`expr $COL + $ADD_POS`
echo $COL > horse_${MOVE_HORSE}_position # 保存新位置.
# 选择要画出来的马的类型.
case $HORSE_TYPE in
1) HORSE_TYPE=2; DRAW_HORSE=draw_horse_two
;;
2) HORSE_TYPE=1; DRAW_HORSE=draw_horse_one
esac
echo $HORSE_TYPE >> horse_${MOVE_HORSE}_position # 保存当前类型.
# 将前景色设为黑, 背景色设为绿.
echo -ne '/E[30;42m'
# 将光标移动到马的新位置.
tput cup `expr $MOVE_HORSE + 5` `cat horse_${MOVE_HORSE}_position | head -1`
# 画马.
$DRAW_HORSE
usleep $USLEEP_ARG
# 当所有的马都越过第15行之后, 再次打印几率.
touch fieldline15
if [ $COL = 15 ]; then
echo $MOVE_HORSE >> fieldline15
fi
if [ `wc -l fieldline15 | cut -f1 -d " "` = 9 ]; then
print_odds
: > fieldline15
fi
# 取得领头的马.
HIGHEST_POS=`cat *position | sort -n | tail -1`
# 将背景色设为白.
echo -ne '/E[47m'
tput cup 17 0
echo -n Current leader: `grep -w $HIGHEST_POS *position | cut -c7`" "
done
# 定义比赛结束的时间.
FINISH_TIME=`date +%s`
# 将背景色设置为绿色, 并且开启闪烁文本的功能.
echo -ne '/E[30;42m'
echo -en '/E[5m'
# 让获胜的马闪烁.
tput cup `expr $MOVE_HORSE + 5` `cat horse_${MOVE_HORSE}_position | head -1`
$DRAW_HORSE
# 禁用闪烁文本.
echo -en '/E[25m'
# 将前景色和背景色设置为白色.
echo -ne '/E[37;47m'
move_and_echo 18 1 $BLANK80
# 将前景色设置为黑色.
echo -ne '/E[30m'
# 让获胜的马闪烁.
tput cup 17 0
echo -e "/E[5mWINNER: $MOVE_HORSE/E[25m"" Odds: `cat odds_${MOVE_HORSE}`"/
" Race time: `expr $FINISH_TIME - $START_TIME` secs"
# 恢复光标, 恢复原来的颜色.
echo -en "/E[?25h"
echo -en "/E[0m"
# 恢复打印功能.
stty echo
# 删除掉和赛马有关的临时文件.
rm -rf $HORSE_RACE_TMP_DIR
tput cup 19 0
exit 0
%%%&&&horserace.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@hypotenuse.sh@@@!!!************************************************************************************
#!/bin/bash
# hypotenuse.sh: 返回直角三角形的斜边.
# (直角边长的平方和,然后对和取平方根)
ARGS=2 # 需要将2个直角边作为参数传递进来.
E_BADARGS=65 # 错误的参数值.
if [ $# -ne "$ARGS" ] # 测试传递到脚本中的参数值.
then
echo "Usage: `basename $0` side_1 side_2"
exit $E_BADARGS
fi
AWKSCRIPT=' { printf( "%3.7f/n", sqrt($1*$1 + $2*$2) ) } '
# 命令 / 传递给awk的参数
# 现在, 将参数通过管道传递给awk.
echo -n "Hypotenuse of $1 and $2 = "
echo $1 $2 | awk "$AWKSCRIPT"
exit 0
%%%&&&hypotenuse.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@idelete.sh@@@!!!************************************************************************************
#!/bin/bash
# idelete.sh: 通过文件的inode号来删除文件.
# 当文件名以一个非法字符开头的时候, 这就非常有用了,
#+ 比如 ? 或 -.
ARGCOUNT=1 # 文件名参数必须被传递到脚本中.
E_WRONGARGS=70
E_FILE_NOT_EXIST=71
E_CHANGED_MIND=72
if [ $# -ne "$ARGCOUNT" ]
then
echo "Usage: `basename $0` filename"
exit $E_WRONGARGS
fi
if [ ! -e "$1" ]
then
echo "File /""$1"/" does not exist."
exit $E_FILE_NOT_EXIST
fi
inum=`ls -i | grep "$1" | awk '{print $1}'`
# inum = inode 文件的(索引节点)号.
# --------------------------------------------------------
# 每个文件都有一个inode号, 这个号用来记录文件物理地址信息.
# --------------------------------------------------------
echo; echo -n "Are you absolutely sure you want to delete /"$1/" (y/n)? "
# 'rm' 命令的 '-v' 选项得询问也会出现这句话.
read answer
case "$answer" in
[nN]) echo "Changed your mind, huh?"
exit $E_CHANGED_MIND
;;
*) echo "Deleting file /"$1/".";;
esac
find . -inum $inum -exec rm {} /;
# ^^
# 大括号就是"find"命令
#+ 用来替换文本输出的地方.
echo "File "/"$1"/" deleted!"
exit 0
%%%&&&idelete.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@ifs-empty.sh@@@!!!************************************************************************************
#!/bin/bash
# 如果$IFS被设置, 但其值为空,
#+ 那么"$*"和"$@"将不会像期望的那样显示位置参数.
mecho () # 打印位置参数.
{
echo "$1,$2,$3";
}
IFS="" # 设置了, 但值为空.
set a b c # 位置参数.
mecho "$*" # abc,,
mecho $* # a,b,c
mecho $@ # a,b,c
mecho "$@" # a,b,c
# 当$IFS值为空时, $*和$@的行为依赖于
#+ 正在运行的Bash或者sh的版本.
# 因此在脚本中使用这种"特性"是不明智的.
# 感谢, Stephane Chazelas.
exit 0
%%%&&&ifs-empty.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@ifs.sh@@@!!!************************************************************************************
#!/bin/bash
# $IFS 处理空白与处理其他字符不同.
output_args_one_per_line()
{
for arg
do echo "[$arg]"
done
}
echo; echo "IFS=/" /""
echo "-------"
IFS=" "
var=" a b c "
output_args_one_per_line $var # output_args_one_per_line `echo " a b c "`
#
# [a]
# [b]
# [c]
echo; echo "IFS=:"
echo "-----"
IFS=:
var=":a::b:c:::" # 与上边一样, 但是用" "替换了":".
output_args_one_per_line $var
#
# []
# [a]
# []
# [b]
# [c]
# []
# []
# []
# 同样的事情也会发生在awk的"FS"域中.
# 感谢, Stephane Chazelas.
echo
exit 0
%%%&&&ifs.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@incompat.sh@@@!!!************************************************************************************
#!/bin/bash
# 内部Bash变量"$*"和"$@"的古怪行为,
#+ 都依赖于它们是否被双引号引用起来.
# 单词拆分与换行的不一致的处理.
set -- "First one" "second" "third:one" "" "Fifth: :one"
# 设置这个脚本的参数, $1, $2, 等等.
echo
echo 'IFS unchanged, using "$*"'
c=0
for i in "$*" # 引用起来
do echo "$((c+=1)): [$i]" # 这行在下边每个例子中都一样.
# 打印参数.
done
echo ---
echo 'IFS unchanged, using $*'
c=0
for i in $* # 未引用
do echo "$((c+=1)): [$i]"
done
echo ---
echo 'IFS unchanged, using "$@"'
c=0
for i in "$@"
do echo "$((c+=1)): [$i]"
done
echo ---
echo 'IFS unchanged, using $@'
c=0
for i in $@
do echo "$((c+=1)): [$i]"
done
echo ---
IFS=:
echo 'IFS=":", using "$*"'
c=0
for i in "$*"
do echo "$((c+=1)): [$i]"
done
echo ---
echo 'IFS=":", using $*'
c=0
for i in $*
do echo "$((c+=1)): [$i]"
done
echo ---
var=$*
echo 'IFS=":", using "$var" (var=$*)'
c=0
for i in "$var"
do echo "$((c+=1)): [$i]"
done
echo ---
echo 'IFS=":", using $var (var=$*)'
c=0
for i in $var
do echo "$((c+=1)): [$i]"
done
echo ---
var="$*"
echo 'IFS=":", using $var (var="$*")'
c=0
for i in $var
do echo "$((c+=1)): [$i]"
done
echo ---
echo 'IFS=":", using "$var" (var="$*")'
c=0
for i in "$var"
do echo "$((c+=1)): [$i]"
done
echo ---
echo 'IFS=":", using "$@"'
c=0
for i in "$@"
do echo "$((c+=1)): [$i]"
done
echo ---
echo 'IFS=":", using $@'
c=0
for i in $@
do echo "$((c+=1)): [$i]"
done
echo ---
var=$@
echo 'IFS=":", using $var (var=$@)'
c=0
for i in $var
do echo "$((c+=1)): [$i]"
done
echo ---
echo 'IFS=":", using "$var" (var=$@)'
c=0
for i in "$var"
do echo "$((c+=1)): [$i]"
done
echo ---
var="$@"
echo 'IFS=":", using "$var" (var="$@")'
c=0
for i in "$var"
do echo "$((c+=1)): [$i]"
done
echo ---
echo 'IFS=":", using $var (var="$@")'
c=0
for i in $var
do echo "$((c+=1)): [$i]"
done
echo
# 使用ksh或者zsh -y来试试这个脚本.
exit 0
# 这个例子脚本是由Stephane Chazelas所编写,
# 并且本书作者做了轻微改动.
%%%&&&incompat.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@ind-func.sh@@@!!!************************************************************************************
#!/bin/bash
# ind-func.sh: 将一个间接引用传递给函数.
echo_var ()
{
echo "$1"
}
message=Hello
Hello=Goodbye
echo_var "$message" # Hello
# 现在,让我们传递一个间接引用给函数.
echo_var "${!message}" # Goodbye
echo "-------------"
# 如果我们改变"hello"变量的值会发生什么?
Hello="Hello, again!"
echo_var "$message" # Hello
echo_var "${!message}" # Hello, again!
exit 0
%%%&&&ind-func.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@ind-ref.sh@@@!!!************************************************************************************
#!/bin/bash
# ind-ref.sh: 间接变量引用.
# 访问一个以另一个变量内容作为名字的变量的值.(译者注: 怎么译都不顺)
a=letter_of_alphabet # 变量"a"的值是另一个变量的名字.
letter_of_alphabet=z
echo
# 直接引用.
echo "a = $a" # a = letter_of_alphabet
# 间接引用.
eval a=/$$a
echo "Now a = $a" # 现在 a = z
echo
# 现在, 让我们试试修改第二个引用的值.
t=table_cell_3
table_cell_3=24
echo "/"table_cell_3/" = $table_cell_3" # "table_cell_3" = 24
echo -n "dereferenced /"t/" = "; eval echo /$$t # 解引用 "t" = 24
# 在这个简单的例子中, 下面的表达式也能正常工作么(为什么?).
# eval t=/$$t; echo "/"t/" = $t"
echo
t=table_cell_3
NEW_VAL=387
table_cell_3=$NEW_VAL
echo "Changing value of /"table_cell_3/" to $NEW_VAL."
echo "/"table_cell_3/" now $table_cell_3"
echo -n "dereferenced /"t/" now "; eval echo /$$t
# "eval" 带有两个参数 "echo" 和 "/$$t" (与$table_cell_3等价)
echo
# (感谢, Stephane Chazelas, 澄清了上边语句的行为.)
# 另一个方法是使用${!t}符号, 见"Bash, 版本2"小节的讨论.
# 也请参考 ex78.sh.
exit 0
%%%&&&ind-ref.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@int-or-string.sh@@@!!!************************************************************************************
#!/bin/bash
# int-or-string.sh: 整型还是字符串?
a=2334 # 整型.
let "a += 1"
echo "a = $a " # a = 2335
echo # 还是整型.
b=${a/23/BB} # 将"23"替换成"BB".
# 这将把变量b从整型变为字符串.
echo "b = $b" # b = BB35
declare -i b # 即使使用declare命令也不会对此有任何帮助.
echo "b = $b" # b = BB35
let "b += 1" # BB35 + 1 =
echo "b = $b" # b = 1
echo
c=BB34
echo "c = $c" # c = BB34
d=${c/BB/23} # 将"BB"替换成"23".
# 这使得变量$d变为一个整形.
echo "d = $d" # d = 2334
let "d += 1" # 2334 + 1 =
echo "d = $d" # d = 2335
echo
# null变量会如何呢?
e=""
echo "e = $e" # e =
let "e += 1" # 算术操作允许一个null变量?
echo "e = $e" # e = 1
echo # null变量将被转换成一个整型变量.
# 如果没有声明变量会怎样?
echo "f = $f" # f =
let "f += 1" # 算术操作能通过么?
echo "f = $f" # f = 1
echo # 未声明的变量将转换成一个整型变量.
# 所以说Bash中的变量都是不区分类型的.
exit 0
%%%&&&int-or-string.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@isalpha.sh@@@!!!************************************************************************************
#!/bin/bash
# isalpha.sh: 使用"case"结构来过滤字符串.
SUCCESS=0
FAILURE=-1
isalpha () # 检查输入的 *第一个字符* 是不是字母表上的字符.
{
if [ -z "$1" ] # 没有参数传进来?
then
return $FAILURE
fi
case "$1" in
[a-zA-Z]*) return $SUCCESS;; # 以一个字母开头?
* ) return $FAILURE;;
esac
} # 同C语言的"isalpha ()"函数比较一下.
isalpha2 () # 测试 *整个字符串* 是否都是字母表上的字符.
{
[ $# -eq 1 ] || return $FAILURE
case $1 in
*[!a-zA-Z]*|"") return $FAILURE;;
*) return $SUCCESS;;
esac
}
isdigit () # 测试 *整个字符串* 是否都是数字.
{ # 换句话说, 就是测试一下是否是整数变量.
[ $# -eq 1 ] || return $FAILURE
case $1 in
*[!0-9]*|"") return $FAILURE;;
*) return $SUCCESS;;
esac
}
check_var () # 测试isalpha().
{
if isalpha "$@"
then
echo "/"$*/" begins with an alpha character."
if isalpha2 "$@"
then # 不需要测试第一个字符是否是non-alpha.
echo "/"$*/" contains only alpha characters."
else
echo "/"$*/" contains at least one non-alpha character."
fi
else
echo "/"$*/" begins with a non-alpha character."
# 如果没有参数传递进来, 也是"non-alpha".
fi
echo
}
digit_check () # 测试isdigit().
{
if isdigit "$@"
then
echo "/"$*/" contains only digits [0 - 9]."
else
echo "/"$*/" has at least one non-digit character."
fi
echo
}
a=23skidoo
b=H3llo
c=-What?
d=What?
e=`echo $b` # 命令替换.
f=AbcDef
g=27234
h=27a34
i=27.34
check_var $a
check_var $b
check_var $c
check_var $d
check_var $e
check_var $f
check_var # 没有参数传递进来, 将会发生什么?
#
digit_check $g
digit_check $h
digit_check $i
exit 0 # S.C改进了这个脚本.
# 练习:
# -----
# 编写一个'isfloat ()'函数来测试浮点数.
# 暗示: 这个函数基本上与'isdigit ()'相同,
#+ 但是要添加一些小数点部分的处理.
%%%&&&isalpha.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@iscan.sh@@@!!!************************************************************************************
#! /bin/sh
## 使用netcat工具写的和DaveG写的ident-scan扫描器有同等功能的东西. 噢, 他会被气死的.
## 参数: target port [port port port ...]
## 标准输出和标准输入被混到一块.
##
## 优点: 运行起来比ident-scan慢, 这样使远程机器inetd进程更不易注意而不会产生警告,
##+ 并且只有很少的知名端口会被指定.
## 缺点: 要求数字端口参数, 输出中无法区分标准输出和标准错误,
##+ 并且当远程服务监听在很高的端口时无法工作.
# 脚本作者: Hobbit <hobbit@avian.org>
# 已征得作者同意在ABS指南中使用.
# ---------------------------------------------------
E_BADARGS=65 # 至少需要两个参数.
TWO_WINKS=2 # 睡眠多长时间.
THREE_WINKS=3
IDPORT=113 # indent协议的认证端口.
RAND1=999
RAND2=31337
TIMEOUT0=9
TIMEOUT1=8
TIMEOUT2=4
# ---------------------------------------------------
case "${2}" in
"" ) echo "Need HOST and at least one PORT." ; exit $E_BADARGS ;;
esac
# 测试目标主机看是否运行了identd守护进程.
nc -z -w $TIMEOUT0 "$1" $IDPORT || { echo "Oops, $1 isn't running identd." ; exit 0 ; }
# -z 选项扫描监听进程.
# -w $TIMEOUT = 尝试连接多长时间.
# 产生一个随机的本地起点源端口.
RP=`expr $$ % $RAND1 + $RAND2`
TRG="$1"
shift
while test "$1" ; do
nc -v -w $TIMEOUT1 -p ${RP} "$TRG" ${1} < /dev/null > /dev/null &
PROC=$!
sleep $THREE_WINKS
echo "${1},${RP}" | nc -w $TIMEOUT2 -r "$TRG" $IDPORT 2>&1
sleep $TWO_WINKS
# 这看上去是不是像个残疾的脚本或是其他类似的东西... ?
# ABS作者评注 : "这不是真的那么糟糕,
#+ 事实上, 做得非常聪明."
kill -HUP $PROC
RP=`expr ${RP} + 1`
shift
done
exit $?
# 注意事项:
# ---------
# 试着把第30行去掉,
#+ 然后以"localhost.localdomain 25"为参数来运行脚本.
# 关于Hobbit写的更多'nc'例子脚本,
#+ 可以在以下文档中找到:
#+ /usr/share/doc/nc-X.XX/scripts 目录.
%%%&&&iscan.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@is_spammer.bash@@@!!!************************************************************************************
#!/bin/bash
# $Id: is_spammer.bash,v 1.12.2.11 2004/10/01 21:42:33 mszick Exp $
# Above line is RCS info.
# The latest version of this script is available from http://www.morethan.org.
#
# Spammer-identification
# by Michael S. Zick
# Used in the ABS Guide with permission.
#######################################################
# Documentation
# See also "Quickstart" at end of script.
#######################################################
:<<-'__is_spammer_Doc_'
Copyright (c) Michael S. Zick, 2004
License: Unrestricted reuse in any form, for any purpose.
Warranty: None -{Its a script; the user is on their own.}-
Impatient?
Application code: goto "# # # Hunt the Spammer' program code # # #"
Example output: ":<<-'_is_spammer_outputs_'"
How to use: Enter script name without arguments.
Or goto "Quickstart" at end of script.
Provides
Given a domain name or IP(v4) address as input:
Does an exhaustive set of queries to find the associated
network resources (short of recursing into TLDs).
Checks the IP(v4) addresses found against Blacklist
nameservers.
If found to be a blacklisted IP(v4) address,
reports the blacklist text records.
(Usually hyper-links to the specific report.)
Requires
A working Internet connection.
(Exercise: Add check and/or abort if not on-line when running script.)
Bash with arrays (2.05b+).
The external program 'dig' --
a utility program provided with the 'bind' set of programs.
Specifically, the version which is part of Bind series 9.x
See: http://www.isc.org
All usages of 'dig' are limited to wrapper functions,
which may be rewritten as required.
See: dig_wrappers.bash for details.
("Additional documentation" -- below)
Usage
Script requires a single argument, which may be:
1) A domain name;
2) An IP(v4) address;
3) A filename, with one name or address per line.
Script accepts an optional second argument, which may be:
1) A Blacklist server name;
2) A filename, with one Blacklist server name per line.
If the second argument is not provided, the script uses
a built-in set of (free) Blacklist servers.
See also, the Quickstart at the end of this script (after 'exit').
Return Codes
0 - All OK
1 - Script failure
2 - Something is Blacklisted
Optional environment variables
SPAMMER_TRACE
If set to a writable file,
script will log an execution flow trace.
SPAMMER_DATA
If set to a writable file, script will dump its
discovered data in the form of GraphViz file.
See: http://www.research.att.com/sw/tools/graphviz
SPAMMER_LIMIT
Limits the depth of resource tracing.
Default is 2 levels.
A setting of 0 (zero) means 'unlimited' . . .
Caution: script might recurse the whole Internet!
A limit of 1 or 2 is most useful when processing
a file of domain names and addresses.
A higher limit can be useful when hunting spam gangs.
Additional documentation
Download the archived set of scripts
explaining and illustrating the function contained within this script.
http://personal.riverusers.com/mszick_clf.tar.bz2
Study notes
This script uses a large number of functions.
Nearly all general functions have their own example script.
Each of the example scripts have tutorial level comments.
Scripting project
Add support for IP(v6) addresses.
IP(v6) addresses are recognized but not processed.
Advanced project
Add the reverse lookup detail to the discovered information.
Report the delegation chain and abuse contacts.
Modify the GraphViz file output to include the
newly discovered information.
__is_spammer_Doc_
#######################################################
#### Special IFS settings used for string parsing. ####
# Whitespace == :Space:Tab:Line Feed:Carriage Return:
WSP_IFS=$'/x20'$'/x09'$'/x0A'$'/x0D'
# No Whitespace == Line Feed:Carriage Return
NO_WSP=$'/x0A'$'/x0D'
# Field separator for dotted decimal IP addresses
ADR_IFS=${NO_WSP}'.'
# Array to dotted string conversions
DOT_IFS='.'${WSP_IFS}
# # # Pending operations stack machine # # #
# This set of functions described in func_stack.bash.
# (See "Additional documentation" above.)
# # #
# Global stack of pending operations.
declare -f -a _pending_
# Global sentinel for stack runners
declare -i _p_ctrl_
# Global holder for currently executing function
declare -f _pend_current_
# # # Debug version only - remove for regular use # # #
#
# The function stored in _pend_hook_ is called
# immediately before each pending function is
# evaluated. Stack clean, _pend_current_ set.
#
# This thingy demonstrated in pend_hook.bash.
declare -f _pend_hook_
# # #
# The do nothing function
pend_dummy() { : ; }
# Clear and initialize the function stack.
pend_init() {
unset _pending_[@]
pend_func pend_stop_mark
_pend_hook_='pend_dummy' # Debug only.
}
# Discard the top function on the stack.
pend_pop() {
if [ ${#_pending_[@]} -gt 0 ]
then
local -i _top_
_top_=${#_pending_[@]}-1
unset _pending_[$_top_]
fi
}
# pend_func function_name [$(printf '%q/n' arguments)]
pend_func() {
local IFS=${NO_WSP}
set -f
_pending_[${#_pending_[@]}]=$@
set +f
}
# The function which stops the release:
pend_stop_mark() {
_p_ctrl_=0
}
pend_mark() {
pend_func pend_stop_mark
}
# Execute functions until 'pend_stop_mark' . . .
pend_release() {
local -i _top_ # Declare _top_ as integer.
_p_ctrl_=${#_pending_[@]}
while [ ${_p_ctrl_} -gt 0 ]
do
_top_=${#_pending_[@]}-1
_pend_current_=${_pending_[$_top_]}
unset _pending_[$_top_]
$_pend_hook_ # Debug only.
eval $_pend_current_
done
}
# Drop functions until 'pend_stop_mark' . . .
pend_drop() {
local -i _top_
local _pd_ctrl_=${#_pending_[@]}
while [ ${_pd_ctrl_} -gt 0 ]
do
_top_=$_pd_ctrl_-1
if [ "${_pending_[$_top_]}" == 'pend_stop_mark' ]
then
unset _pending_[$_top_]
break
else
unset _pending_[$_top_]
_pd_ctrl_=$_top_
fi
done
if [ ${#_pending_[@]} -eq 0 ]
then
pend_func pend_stop_mark
fi
}
#### Array editors ####
# This function described in edit_exact.bash.
# (See "Additional documentation," above.)
# edit_exact <excludes_array_name> <target_array_name>
edit_exact() {
[ $# -eq 2 ] ||
[ $# -eq 3 ] || return 1
local -a _ee_Excludes
local -a _ee_Target
local _ee_x
local _ee_t
local IFS=${NO_WSP}
set -f
eval _ee_Excludes=/( /$/{$1/[@/]/} /)
eval _ee_Target=/( /$/{$2/[@/]/} /)
local _ee_len=${#_ee_Target[@]} # Original length.
local _ee_cnt=${#_ee_Excludes[@]} # Exclude list length.
[ ${_ee_len} -ne 0 ] || return 0 # Can't edit zero length.
[ ${_ee_cnt} -ne 0 ] || return 0 # Can't edit zero length.
for (( x = 0; x < ${_ee_cnt} ; x++ ))
do
_ee_x=${_ee_Excludes[$x]}
for (( n = 0 ; n < ${_ee_len} ; n++ ))
do
_ee_t=${_ee_Target[$n]}
if [ x"${_ee_t}" == x"${_ee_x}" ]
then
unset _ee_Target[$n] # Discard match.
[ $# -eq 2 ] && break # If 2 arguments, then done.
fi
done
done
eval $2=/( /$/{_ee_Target/[@/]/} /)
set +f
return 0
}
# This function described in edit_by_glob.bash.
# edit_by_glob <excludes_array_name> <target_array_name>
edit_by_glob() {
[ $# -eq 2 ] ||
[ $# -eq 3 ] || return 1
local -a _ebg_Excludes
local -a _ebg_Target
local _ebg_x
local _ebg_t
local IFS=${NO_WSP}
set -f
eval _ebg_Excludes=/( /$/{$1/[@/]/} /)
eval _ebg_Target=/( /$/{$2/[@/]/} /)
local _ebg_len=${#_ebg_Target[@]}
local _ebg_cnt=${#_ebg_Excludes[@]}
[ ${_ebg_len} -ne 0 ] || return 0
[ ${_ebg_cnt} -ne 0 ] || return 0
for (( x = 0; x < ${_ebg_cnt} ; x++ ))
do
_ebg_x=${_ebg_Excludes[$x]}
for (( n = 0 ; n < ${_ebg_len} ; n++ ))
do
[ $# -eq 3 ] && _ebg_x=${_ebg_x}'*' # Do prefix edit
if [ ${_ebg_Target[$n]:=} ] #+ if defined & set.
then
_ebg_t=${_ebg_Target[$n]/#${_ebg_x}/}
[ ${#_ebg_t} -eq 0 ] && unset _ebg_Target[$n]
fi
done
done
eval $2=/( /$/{_ebg_Target/[@/]/} /)
set +f
return 0
}
# This function described in unique_lines.bash.
# unique_lines <in_name> <out_name>
unique_lines() {
[ $# -eq 2 ] || return 1
local -a _ul_in
local -a _ul_out
local -i _ul_cnt
local -i _ul_pos
local _ul_tmp
local IFS=${NO_WSP}
set -f
eval _ul_in=/( /$/{$1/[@/]/} /)
_ul_cnt=${#_ul_in[@]}
for (( _ul_pos = 0 ; _ul_pos < ${_ul_cnt} ; _ul_pos++ ))
do
if [ ${_ul_in[${_ul_pos}]:=} ] # If defined & not empty
then
_ul_tmp=${_ul_in[${_ul_pos}]}
_ul_out[${#_ul_out[@]}]=${_ul_tmp}
for (( zap = _ul_pos ; zap < ${_ul_cnt} ; zap++ ))
do
[ ${_ul_in[${zap}]:=} ] &&
[ 'x'${_ul_in[${zap}]} == 'x'${_ul_tmp} ] &&
unset _ul_in[${zap}]
done
fi
done
eval $2=/( /$/{_ul_out/[@/]/} /)
set +f
return 0
}
# This function described in char_convert.bash.
# to_lower <string>
to_lower() {
[ $# -eq 1 ] || return 1
local _tl_out
_tl_out=${1//A/a}
_tl_out=${_tl_out//B/b}
_tl_out=${_tl_out//C/c}
_tl_out=${_tl_out//D/d}
_tl_out=${_tl_out//E/e}
_tl_out=${_tl_out//F/f}
_tl_out=${_tl_out//G/g}
_tl_out=${_tl_out//H/h}
_tl_out=${_tl_out//I/i}
_tl_out=${_tl_out//J/j}
_tl_out=${_tl_out//K/k}
_tl_out=${_tl_out//L/l}
_tl_out=${_tl_out//M/m}
_tl_out=${_tl_out//N/n}
_tl_out=${_tl_out//O/o}
_tl_out=${_tl_out//P/p}
_tl_out=${_tl_out//Q/q}
_tl_out=${_tl_out//R/r}
_tl_out=${_tl_out//S/s}
_tl_out=${_tl_out//T/t}
_tl_out=${_tl_out//U/u}
_tl_out=${_tl_out//V/v}
_tl_out=${_tl_out//W/w}
_tl_out=${_tl_out//X/x}
_tl_out=${_tl_out//Y/y}
_tl_out=${_tl_out//Z/z}
echo ${_tl_out}
return 0
}
#### Application helper functions ####
# Not everybody uses dots as separators (APNIC, for example).
# This function described in to_dot.bash
# to_dot <string>
to_dot() {
[ $# -eq 1 ] || return 1
echo ${1//[#|@|%]/.}
return 0
}
# This function described in is_number.bash.
# is_number <input>
is_number() {
[ "$#" -eq 1 ] || return 1 # is blank?
[ x"$1" == 'x0' ] && return 0 # is zero?
local -i tst
let tst=$1 2>/dev/null # else is numeric!
return $?
}
# This function described in is_address.bash.
# is_address <input>
is_address() {
[ $# -eq 1 ] || return 1 # Blank ==> false
local -a _ia_input
local IFS=${ADR_IFS}
_ia_input=( $1 )
if [ ${#_ia_input[@]} -eq 4 ] &&
is_number ${_ia_input[0]} &&
is_number ${_ia_input[1]} &&
is_number ${_ia_input[2]} &&
is_number ${_ia_input[3]} &&
[ ${_ia_input[0]} -lt 256 ] &&
[ ${_ia_input[1]} -lt 256 ] &&
[ ${_ia_input[2]} -lt 256 ] &&
[ ${_ia_input[3]} -lt 256 ]
then
return 0
else
return 1
fi
}
# This function described in split_ip.bash.
# split_ip <IP_address> <array_name_norm> [<array_name_rev>]
split_ip() {
[ $# -eq 3 ] || # Either three
[ $# -eq 2 ] || return 1 #+ or two arguments
local -a _si_input
local IFS=${ADR_IFS}
_si_input=( $1 )
IFS=${WSP_IFS}
eval $2=/(/ /$/{_si_input/[@/]/}/ /)
if [ $# -eq 3 ]
then
# Build query order array.
local -a _dns_ip
_dns_ip[0]=${_si_input[3]}
_dns_ip[1]=${_si_input[2]}
_dns_ip[2]=${_si_input[1]}
_dns_ip[3]=${_si_input[0]}
eval $3=/(/ /$/{_dns_ip/[@/]/}/ /)
fi
return 0
}
# This function described in dot_array.bash.
# dot_array <array_name>
dot_array() {
[ $# -eq 1 ] || return 1 # Single argument required.
local -a _da_input
eval _da_input=/(/ /$/{$1/[@/]/}/ /)
local IFS=${DOT_IFS}
local _da_output=${_da_input[@]}
IFS=${WSP_IFS}
echo ${_da_output}
return 0
}
# This function described in file_to_array.bash
# file_to_array <file_name> <line_array_name>
file_to_array() {
[ $# -eq 2 ] || return 1 # Two arguments required.
local IFS=${NO_WSP}
local -a _fta_tmp_
_fta_tmp_=( $(cat $1) )
eval $2=/( /$/{_fta_tmp_/[@/]/} /)
return 0
}
# Columnized print of an array of multi-field strings.
# col_print <array_name> <min_space> <tab_stop [tab_stops]>
col_print() {
[ $# -gt 2 ] || return 0
local -a _cp_inp
local -a _cp_spc
local -a _cp_line
local _cp_min
local _cp_mcnt
local _cp_pos
local _cp_cnt
local _cp_tab
local -i _cp
local -i _cpf
local _cp_fld
# WARNING: FOLLOWING LINE NOT BLANK -- IT IS QUOTED SPACES.
local _cp_max=' '
set -f
local IFS=${NO_WSP}
eval _cp_inp=/(/ /$/{$1/[@/]/}/ /)
[ ${#_cp_inp[@]} -gt 0 ] || return 0 # Empty is easy.
_cp_mcnt=$2
_cp_min=${_cp_max:1:${_cp_mcnt}}
shift
shift
_cp_cnt=$#
for (( _cp = 0 ; _cp < _cp_cnt ; _cp++ ))
do
_cp_spc[${#_cp_spc[@]}]="${_cp_max:2:$1}" #"
shift
done
_cp_cnt=${#_cp_inp[@]}
for (( _cp = 0 ; _cp < _cp_cnt ; _cp++ ))
do
_cp_pos=1
IFS=${NO_WSP}$'/x20'
_cp_line=( ${_cp_inp[${_cp}]} )
IFS=${NO_WSP}
for (( _cpf = 0 ; _cpf < ${#_cp_line[@]} ; _cpf++ ))
do
_cp_tab=${_cp_spc[${_cpf}]:${_cp_pos}}
if [ ${#_cp_tab} -lt ${_cp_mcnt} ]
then
_cp_tab="${_cp_min}"
fi
echo -n "${_cp_tab}"
(( _cp_pos = ${_cp_pos} + ${#_cp_tab} ))
_cp_fld="${_cp_line[${_cpf}]}"
echo -n ${_cp_fld}
(( _cp_pos = ${_cp_pos} + ${#_cp_fld} ))
done
echo
done
set +f
return 0
}
# # # # 'Hunt the Spammer' data flow # # # #
# Application return code
declare -i _hs_RC
# Original input, from which IP addresses are removed
# After which, domain names to check
declare -a uc_name
# Original input IP addresses are moved here
# After which, IP addresses to check
declare -a uc_address
# Names against which address expansion run
# Ready for name detail lookup
declare -a chk_name
# Addresses against which name expansion run
# Ready for address detail lookup
declare -a chk_address
# Recursion is depth-first-by-name.
# The expand_input_address maintains this list
#+ to prohibit looking up addresses twice during
#+ domain name recursion.
declare -a been_there_addr
been_there_addr=( '127.0.0.1' ) # Whitelist localhost
# Names which we have checked (or given up on)
declare -a known_name
# Addresses which we have checked (or given up on)
declare -a known_address
# List of zero or more Blacklist servers to check.
# Each 'known_address' will be checked against each server,
#+ with negative replies and failures suppressed.
declare -a list_server
# Indirection limit - set to zero == no limit
indirect=${SPAMMER_LIMIT:=2}
# # # # 'Hunt the Spammer' information output data # # # #
# Any domain name may have multiple IP addresses.
# Any IP address may have multiple domain names.
# Therefore, track unique address-name pairs.
declare -a known_pair
declare -a reverse_pair
# In addition to the data flow variables; known_address
#+ known_name and list_server, the following are output to the
#+ external graphics interface file.
# Authority chain, parent -> SOA fields.
declare -a auth_chain
# Reference chain, parent name -> child name
declare -a ref_chain
# DNS chain - domain name -> address
declare -a name_address
# Name and service pairs - domain name -> service
declare -a name_srvc
# Name and resource pairs - domain name -> Resource Record
declare -a name_resource
# Parent and Child pairs - parent name -> child name
# This MAY NOT be the same as the ref_chain followed!
declare -a parent_child
# Address and Blacklist hit pairs - address->server
declare -a address_hits
# Dump interface file data
declare -f _dot_dump
_dot_dump=pend_dummy # Initially a no-op
# Data dump is enabled by setting the environment variable SPAMMER_DATA
#+ to the name of a writable file.
declare _dot_file
# Helper function for the dump-to-dot-file function
# dump_to_dot <array_name> <prefix>
dump_to_dot() {
local -a _dda_tmp
local -i _dda_cnt
local _dda_form=' '${2}'%04u %s/n'
local IFS=${NO_WSP}
eval _dda_tmp=/(/ /$/{$1/[@/]/}/ /)
_dda_cnt=${#_dda_tmp[@]}
if [ ${_dda_cnt} -gt 0 ]
then
for (( _dda = 0 ; _dda < _dda_cnt ; _dda++ ))
do
printf "${_dda_form}" /
"${_dda}" "${_dda_tmp[${_dda}]}" >>${_dot_file}
done
fi
}
# Which will also set _dot_dump to this function . . .
dump_dot() {
local -i _dd_cnt
echo '# Data vintage: '$(date -R) >${_dot_file}
echo '# ABS Guide: is_spammer.bash; v2, 2004-msz' >>${_dot_file}
echo >>${_dot_file}
echo 'digraph G {' >>${_dot_file}
if [ ${#known_name[@]} -gt 0 ]
then
echo >>${_dot_file}
echo '# Known domain name nodes' >>${_dot_file}
_dd_cnt=${#known_name[@]}
for (( _dd = 0 ; _dd < _dd_cnt ; _dd++ ))
do
printf ' N%04u [label="%s"] ;/n' /
"${_dd}" "${known_name[${_dd}]}" >>${_dot_file}
done
fi
if [ ${#known_address[@]} -gt 0 ]
then
echo >>${_dot_file}
echo '# Known address nodes' >>${_dot_file}
_dd_cnt=${#known_address[@]}
for (( _dd = 0 ; _dd < _dd_cnt ; _dd++ ))
do
printf ' A%04u [label="%s"] ;/n' /
"${_dd}" "${known_address[${_dd}]}" >>${_dot_file}
done
fi
echo >>${_dot_file}
echo '/*' >>${_dot_file}
echo ' * Known relationships :: User conversion to' >>${_dot_file}
echo ' * graphic form by hand or program required.' >>${_dot_file}
echo ' *' >>${_dot_file}
if [ ${#auth_chain[@]} -gt 0 ]
then
echo >>${_dot_file}
echo '# Authority reference edges followed and field source.' >>${_dot_file}
dump_to_dot auth_chain AC
fi
if [ ${#ref_chain[@]} -gt 0 ]
then
echo >>${_dot_file}
echo '# Name reference edges followed and field source.' >>${_dot_file}
dump_to_dot ref_chain RC
fi
if [ ${#name_address[@]} -gt 0 ]
then
echo >>${_dot_file}
echo '# Known name->address edges' >>${_dot_file}
dump_to_dot name_address NA
fi
if [ ${#name_srvc[@]} -gt 0 ]
then
echo >>${_dot_file}
echo '# Known name->service edges' >>${_dot_file}
dump_to_dot name_srvc NS
fi
if [ ${#name_resource[@]} -gt 0 ]
then
echo >>${_dot_file}
echo '# Known name->resource edges' >>${_dot_file}
dump_to_dot name_resource NR
fi
if [ ${#parent_child[@]} -gt 0 ]
then
echo >>${_dot_file}
echo '# Known parent->child edges' >>${_dot_file}
dump_to_dot parent_child PC
fi
if [ ${#list_server[@]} -gt 0 ]
then
echo >>${_dot_file}
echo '# Known Blacklist nodes' >>${_dot_file}
_dd_cnt=${#list_server[@]}
for (( _dd = 0 ; _dd < _dd_cnt ; _dd++ ))
do
printf ' LS%04u [label="%s"] ;/n' /
"${_dd}" "${list_server[${_dd}]}" >>${_dot_file}
done
fi
unique_lines address_hits address_hits
if [ ${#address_hits[@]} -gt 0 ]
then
echo >>${_dot_file}
echo '# Known address->Blacklist_hit edges' >>${_dot_file}
echo '# CAUTION: dig warnings can trigger false hits.' >>${_dot_file}
dump_to_dot address_hits AH
fi
echo >>${_dot_file}
echo ' *' >>${_dot_file}
echo ' * That is a lot of relationships. Happy graphing.' >>${_dot_file}
echo ' */' >>${_dot_file}
echo '}' >>${_dot_file}
return 0
}
# # # # 'Hunt the Spammer' execution flow # # # #
# Execution trace is enabled by setting the
#+ environment variable SPAMMER_TRACE to the name of a writable file.
declare -a _trace_log
declare _log_file
# Function to fill the trace log
trace_logger() {
_trace_log[${#_trace_log[@]}]=${_pend_current_}
}
# Dump trace log to file function variable.
declare -f _log_dump
_log_dump=pend_dummy # Initially a no-op.
# Dump the trace log to a file.
dump_log() {
local -i _dl_cnt
_dl_cnt=${#_trace_log[@]}
for (( _dl = 0 ; _dl < _dl_cnt ; _dl++ ))
do
echo ${_trace_log[${_dl}]} >> ${_log_file}
done
_dl_cnt=${#_pending_[@]}
if [ ${_dl_cnt} -gt 0 ]
then
_dl_cnt=${_dl_cnt}-1
echo '# # # Operations stack not empty # # #' >> ${_log_file}
for (( _dl = ${_dl_cnt} ; _dl >= 0 ; _dl-- ))
do
echo ${_pending_[${_dl}]} >> ${_log_file}
done
fi
}
# # # Utility program 'dig' wrappers # # #
#
# These wrappers are derived from the
#+ examples shown in dig_wrappers.bash.
#
# The major difference is these return
#+ their results as a list in an array.
#
# See dig_wrappers.bash for details and
#+ use that script to develop any changes.
#
# # #
# Short form answer: 'dig' parses answer.
# Forward lookup :: Name -> Address
# short_fwd <domain_name> <array_name>
short_fwd() {
local -a _sf_reply
local -i _sf_rc
local -i _sf_cnt
IFS=${NO_WSP}
echo -n '.'
# echo 'sfwd: '${1}
_sf_reply=( $(dig +short ${1} -c in -t a 2>/dev/null) )
_sf_rc=$?
if [ ${_sf_rc} -ne 0 ]
then
_trace_log[${#_trace_log[@]}]='# # # Lookup error '${_sf_rc}' on '${1}' # # #'
# [ ${_sf_rc} -ne 9 ] && pend_drop
return ${_sf_rc}
else
# Some versions of 'dig' return warnings on stdout.
_sf_cnt=${#_sf_reply[@]}
for (( _sf = 0 ; _sf < ${_sf_cnt} ; _sf++ ))
do
[ 'x'${_sf_reply[${_sf}]:0:2} == 'x;;' ] &&
unset _sf_reply[${_sf}]
done
eval $2=/( /$/{_sf_reply/[@/]/} /)
fi
return 0
}
# Reverse lookup :: Address -> Name
# short_rev <ip_address> <array_name>
short_rev() {
local -a _sr_reply
local -i _sr_rc
local -i _sr_cnt
IFS=${NO_WSP}
echo -n '.'
# echo 'srev: '${1}
_sr_reply=( $(dig +short -x ${1} 2>/dev/null) )
_sr_rc=$?
if [ ${_sr_rc} -ne 0 ]
then
_trace_log[${#_trace_log[@]}]='# # # Lookup error '${_sr_rc}' on '${1}' # # #'
# [ ${_sr_rc} -ne 9 ] && pend_drop
return ${_sr_rc}
else
# Some versions of 'dig' return warnings on stdout.
_sr_cnt=${#_sr_reply[@]}
for (( _sr = 0 ; _sr < ${_sr_cnt} ; _sr++ ))
do
[ 'x'${_sr_reply[${_sr}]:0:2} == 'x;;' ] &&
unset _sr_reply[${_sr}]
done
eval $2=/( /$/{_sr_reply/[@/]/} /)
fi
return 0
}
# Special format lookup used to query blacklist servers.
# short_text <ip_address> <array_name>
short_text() {
local -a _st_reply
local -i _st_rc
local -i _st_cnt
IFS=${NO_WSP}
# echo 'stxt: '${1}
_st_reply=( $(dig +short ${1} -c in -t txt 2>/dev/null) )
_st_rc=$?
if [ ${_st_rc} -ne 0 ]
then
_trace_log[${#_trace_log[@]}]='# # # Text lookup error '${_st_rc}' on '${1}' # # #'
# [ ${_st_rc} -ne 9 ] && pend_drop
return ${_st_rc}
else
# Some versions of 'dig' return warnings on stdout.
_st_cnt=${#_st_reply[@]}
for (( _st = 0 ; _st < ${#_st_cnt} ; _st++ ))
do
[ 'x'${_st_reply[${_st}]:0:2} == 'x;;' ] &&
unset _st_reply[${_st}]
done
eval $2=/( /$/{_st_reply/[@/]/} /)
fi
return 0
}
# The long forms, a.k.a., the parse it yourself versions
# RFC 2782 Service lookups
# dig +noall +nofail +answer _ldap._tcp.openldap.org -t srv
# _<service>._<protocol>.<domain_name>
# _ldap._tcp.openldap.org. 3600 IN SRV 0 0 389 ldap.openldap.org.
# domain TTL Class SRV Priority Weight Port Target
# Forward lookup :: Name -> poor man's zone transfer
# long_fwd <domain_name> <array_name>
long_fwd() {
local -a _lf_reply
local -i _lf_rc
local -i _lf_cnt
IFS=${NO_WSP}
echo -n ':'
# echo 'lfwd: '${1}
_lf_reply=( $(
dig +noall +nofail +answer +authority +additional /
${1} -t soa ${1} -t mx ${1} -t any 2>/dev/null) )
_lf_rc=$?
if [ ${_lf_rc} -ne 0 ]
then
_trace_log[${#_trace_log[@]}]='# # # Zone lookup error '${_lf_rc}' on '${1}' # # #'
# [ ${_lf_rc} -ne 9 ] && pend_drop
return ${_lf_rc}
else
# Some versions of 'dig' return warnings on stdout.
_lf_cnt=${#_lf_reply[@]}
for (( _lf = 0 ; _lf < ${_lf_cnt} ; _lf++ ))
do
[ 'x'${_lf_reply[${_lf}]:0:2} == 'x;;' ] &&
unset _lf_reply[${_lf}]
done
eval $2=/( /$/{_lf_reply/[@/]/} /)
fi
return 0
}
# The reverse lookup domain name corresponding to the IPv6 address:
# 4321:0:1:2:3:4:567:89ab
# would be (nibble, I.E: Hexdigit) reversed:
# b.a.9.8.7.6.5.0.4.0.0.0.3.0.0.0.2.0.0.0.1.0.0.0.0.0.0.0.1.2.3.4.IP6.ARPA.
# Reverse lookup :: Address -> poor man's delegation chain
# long_rev <rev_ip_address> <array_name>
long_rev() {
local -a _lr_reply
local -i _lr_rc
local -i _lr_cnt
local _lr_dns
_lr_dns=${1}'.in-addr.arpa.'
IFS=${NO_WSP}
echo -n ':'
# echo 'lrev: '${1}
_lr_reply=( $(
dig +noall +nofail +answer +authority +additional /
${_lr_dns} -t soa ${_lr_dns} -t any 2>/dev/null) )
_lr_rc=$?
if [ ${_lr_rc} -ne 0 ]
then
_trace_log[${#_trace_log[@]}]='# # # Delegation lookup error '${_lr_rc}' on '${1}' # # #'
# [ ${_lr_rc} -ne 9 ] && pend_drop
return ${_lr_rc}
else
# Some versions of 'dig' return warnings on stdout.
_lr_cnt=${#_lr_reply[@]}
for (( _lr = 0 ; _lr < ${_lr_cnt} ; _lr++ ))
do
[ 'x'${_lr_reply[${_lr}]:0:2} == 'x;;' ] &&
unset _lr_reply[${_lr}]
done
eval $2=/( /$/{_lr_reply/[@/]/} /)
fi
return 0
}
# # # Application specific functions # # #
# Mung a possible name; suppresses root and TLDs.
# name_fixup <string>
name_fixup(){
local -a _nf_tmp
local -i _nf_end
local _nf_str
local IFS
_nf_str=$(to_lower ${1})
_nf_str=$(to_dot ${_nf_str})
_nf_end=${#_nf_str}-1
[ ${_nf_str:${_nf_end}} != '.' ] &&
_nf_str=${_nf_str}'.'
IFS=${ADR_IFS}
_nf_tmp=( ${_nf_str} )
IFS=${WSP_IFS}
_nf_end=${#_nf_tmp[@]}
case ${_nf_end} in
0) # No dots, only dots.
echo
return 1
;;
1) # Only a TLD.
echo
return 1
;;
2) # Maybe okay.
echo ${_nf_str}
return 0
# Needs a lookup table?
if [ ${#_nf_tmp[1]} -eq 2 ]
then # Country coded TLD.
echo
return 1
else
echo ${_nf_str}
return 0
fi
;;
esac
echo ${_nf_str}
return 0
}
# Grope and mung original input(s).
split_input() {
[ ${#uc_name[@]} -gt 0 ] || return 0
local -i _si_cnt
local -i _si_len
local _si_str
unique_lines uc_name uc_name
_si_cnt=${#uc_name[@]}
for (( _si = 0 ; _si < _si_cnt ; _si++ ))
do
_si_str=${uc_name[$_si]}
if is_address ${_si_str}
then
uc_address[${#uc_address[@]}]=${_si_str}
unset uc_name[$_si]
else
if ! uc_name[$_si]=$(name_fixup ${_si_str})
then
unset ucname[$_si]
fi
fi
done
uc_name=( ${uc_name[@]} )
_si_cnt=${#uc_name[@]}
_trace_log[${#_trace_log[@]}]='# # # Input '${_si_cnt}' unchecked name input(s). # # #'
_si_cnt=${#uc_address[@]}
_trace_log[${#_trace_log[@]}]='# # # Input '${_si_cnt}' unchecked address input(s). # # #'
return 0
}
# # # Discovery functions -- recursively interlocked by external data # # #
# # # The leading 'if list is empty; return 0' in each is required. # # #
# Recursion limiter
# limit_chk() <next_level>
limit_chk() {
local -i _lc_lmt
# Check indirection limit.
if [ ${indirect} -eq 0 ] || [ $# -eq 0 ]
then
# The 'do-forever' choice
echo 1 # Any value will do.
return 0 # OK to continue.
else
# Limiting is in effect.
if [ ${indirect} -lt ${1} ]
then
echo ${1} # Whatever.
return 1 # Stop here.
else
_lc_lmt=${1}+1 # Bump the given limit.
echo ${_lc_lmt} # Echo it.
return 0 # OK to continue.
fi
fi
}
# For each name in uc_name:
# Move name to chk_name.
# Add addresses to uc_address.
# Pend expand_input_address.
# Repeat until nothing new found.
# expand_input_name <indirection_limit>
expand_input_name() {
[ ${#uc_name[@]} -gt 0 ] || return 0
local -a _ein_addr
local -a _ein_new
local -i _ucn_cnt
local -i _ein_cnt
local _ein_tst
_ucn_cnt=${#uc_name[@]}
if ! _ein_cnt=$(limit_chk ${1})
then
return 0
fi
for (( _ein = 0 ; _ein < _ucn_cnt ; _ein++ ))
do
if short_fwd ${uc_name[${_ein}]} _ein_new
then
for (( _ein_cnt = 0 ; _ein_cnt < ${#_ein_new[@]}; _ein_cnt++ ))
do
_ein_tst=${_ein_new[${_ein_cnt}]}
if is_address ${_ein_tst}
then
_ein_addr[${#_ein_addr[@]}]=${_ein_tst}
fi
done
fi
done
unique_lines _ein_addr _ein_addr # Scrub duplicates.
edit_exact chk_address _ein_addr # Scrub pending detail.
edit_exact known_address _ein_addr # Scrub already detailed.
if [ ${#_ein_addr[@]} -gt 0 ] # Anything new?
then
uc_address=( ${uc_address[@]} ${_ein_addr[@]} )
pend_func expand_input_address ${1}
_trace_log[${#_trace_log[@]}]='# # # Added '${#_ein_addr[@]}' unchecked address input(s). # # #'
fi
edit_exact chk_name uc_name # Scrub pending detail.
edit_exact known_name uc_name # Scrub already detailed.
if [ ${#uc_name[@]} -gt 0 ]
then
chk_name=( ${chk_name[@]} ${uc_name[@]} )
pend_func detail_each_name ${1}
fi
unset uc_name[@]
return 0
}
# For each address in uc_address:
# Move address to chk_address.
# Add names to uc_name.
# Pend expand_input_name.
# Repeat until nothing new found.
# expand_input_address <indirection_limit>
expand_input_address() {
[ ${#uc_address[@]} -gt 0 ] || return 0
local -a _eia_addr
local -a _eia_name
local -a _eia_new
local -i _uca_cnt
local -i _eia_cnt
local _eia_tst
unique_lines uc_address _eia_addr
unset uc_address[@]
edit_exact been_there_addr _eia_addr
_uca_cnt=${#_eia_addr[@]}
[ ${_uca_cnt} -gt 0 ] &&
been_there_addr=( ${been_there_addr[@]} ${_eia_addr[@]} )
for (( _eia = 0 ; _eia < _uca_cnt ; _eia++ ))
do
if short_rev ${_eia_addr[${_eia}]} _eia_new
then
for (( _eia_cnt = 0 ; _eia_cnt < ${#_eia_new[@]} ; _eia_cnt++ ))
do
_eia_tst=${_eia_new[${_eia_cnt}]}
if _eia_tst=$(name_fixup ${_eia_tst})
then
_eia_name[${#_eia_name[@]}]=${_eia_tst}
fi
done
fi
done
unique_lines _eia_name _eia_name # Scrub duplicates.
edit_exact chk_name _eia_name # Scrub pending detail.
edit_exact known_name _eia_name # Scrub already detailed.
if [ ${#_eia_name[@]} -gt 0 ] # Anything new?
then
uc_name=( ${uc_name[@]} ${_eia_name[@]} )
pend_func expand_input_name ${1}
_trace_log[${#_trace_log[@]}]='# # # Added '${#_eia_name[@]}' unchecked name input(s). # # #'
fi
edit_exact chk_address _eia_addr # Scrub pending detail.
edit_exact known_address _eia_addr # Scrub already detailed.
if [ ${#_eia_addr[@]} -gt 0 ] # Anything new?
then
chk_address=( ${chk_address[@]} ${_eia_addr[@]} )
pend_func detail_each_address ${1}
fi
return 0
}
# The parse-it-yourself zone reply.
# The input is the chk_name list.
# detail_each_name <indirection_limit>
detail_each_name() {
[ ${#chk_name[@]} -gt 0 ] || return 0
local -a _den_chk # Names to check
local -a _den_name # Names found here
local -a _den_address # Addresses found here
local -a _den_pair # Pairs found here
local -a _den_rev # Reverse pairs found here
local -a _den_tmp # Line being parsed
local -a _den_auth # SOA contact being parsed
local -a _den_new # The zone reply
local -a _den_pc # Parent-Child gets big fast
local -a _den_ref # So does reference chain
local -a _den_nr # Name-Resource can be big
local -a _den_na # Name-Address
local -a _den_ns # Name-Service
local -a _den_achn # Chain of Authority
local -i _den_cnt # Count of names to detail
local -i _den_lmt # Indirection limit
local _den_who # Named being processed
local _den_rec # Record type being processed
local _den_cont # Contact domain
local _den_str # Fixed up name string
local _den_str2 # Fixed up reverse
local IFS=${WSP_IFS}
# Local, unique copy of names to check
unique_lines chk_name _den_chk
unset chk_name[@] # Done with globals.
# Less any names already known
edit_exact known_name _den_chk
_den_cnt=${#_den_chk[@]}
# If anything left, add to known_name.
[ ${_den_cnt} -gt 0 ] &&
known_name=( ${known_name[@]} ${_den_chk[@]} )
# for the list of (previously) unknown names . . .
for (( _den = 0 ; _den < _den_cnt ; _den++ ))
do
_den_who=${_den_chk[${_den}]}
if long_fwd ${_den_who} _den_new
then
unique_lines _den_new _den_new
if [ ${#_den_new[@]} -eq 0 ]
then
_den_pair[${#_den_pair[@]}]='0.0.0.0 '${_den_who}
fi
# Parse each line in the reply.
for (( _line = 0 ; _line < ${#_den_new[@]} ; _line++ ))
do
IFS=${NO_WSP}$'/x09'$'/x20'
_den_tmp=( ${_den_new[${_line}]} )
IFS=${WSP_IFS}
# If usable record and not a warning message . . .
if [ ${#_den_tmp[@]} -gt 4 ] && [ 'x'${_den_tmp[0]} != 'x;;' ]
then
_den_rec=${_den_tmp[3]}
_den_nr[${#_den_nr[@]}]=${_den_who}' '${_den_rec}
# Begin at RFC1033 (+++)
case ${_den_rec} in
#<name> [<ttl>] [<class>] SOA <origin> <person>
SOA) # Start Of Authority
if _den_str=$(name_fixup ${_den_tmp[0]})
then
_den_name[${#_den_name[@]}]=${_den_str}
_den_achn[${#_den_achn[@]}]=${_den_who}' '${_den_str}' SOA'
# SOA origin -- domain name of master zone record
if _den_str2=$(name_fixup ${_den_tmp[4]})
then
_den_name[${#_den_name[@]}]=${_den_str2}
_den_achn[${#_den_achn[@]}]=${_den_who}' '${_den_str2}' SOA.O'
fi
# Responsible party e-mail address (possibly bogus).
# Possibility of first.last@domain.name ignored.
set -f
if _den_str2=$(name_fixup ${_den_tmp[5]})
then
IFS=${ADR_IFS}
_den_auth=( ${_den_str2} )
IFS=${WSP_IFS}
if [ ${#_den_auth[@]} -gt 2 ]
then
_den_cont=${_den_auth[1]}
for (( _auth = 2 ; _auth < ${#_den_auth[@]} ; _auth++ ))
do
_den_cont=${_den_cont}'.'${_den_auth[${_auth}]}
done
_den_name[${#_den_name[@]}]=${_den_cont}'.'
_den_achn[${#_den_achn[@]}]=${_den_who}' '${_den_cont}'. SOA.C'
fi
fi
set +f
fi
;;
A) # IP(v4) Address Record
if _den_str=$(name_fixup ${_den_tmp[0]})
then
_den_name[${#_den_name[@]}]=${_den_str}
_den_pair[${#_den_pair[@]}]=${_den_tmp[4]}' '${_den_str}
_den_na[${#_den_na[@]}]=${_den_str}' '${_den_tmp[4]}
_den_ref[${#_den_ref[@]}]=${_den_who}' '${_den_str}' A'
else
_den_pair[${#_den_pair[@]}]=${_den_tmp[4]}' unknown.domain'
_den_na[${#_den_na[@]}]='unknown.domain '${_den_tmp[4]}
_den_ref[${#_den_ref[@]}]=${_den_who}' unknown.domain A'
fi
_den_address[${#_den_address[@]}]=${_den_tmp[4]}
_den_pc[${#_den_pc[@]}]=${_den_who}' '${_den_tmp[4]}
;;
NS) # Name Server Record
# Domain name being serviced (may be other than current)
if _den_str=$(name_fixup ${_den_tmp[0]})
then
_den_name[${#_den_name[@]}]=${_den_str}
_den_ref[${#_den_ref[@]}]=${_den_who}' '${_den_str}' NS'
# Domain name of service provider
if _den_str2=$(name_fixup ${_den_tmp[4]})
then
_den_name[${#_den_name[@]}]=${_den_str2}
_den_ref[${#_den_ref[@]}]=${_den_who}' '${_den_str2}' NSH'
_den_ns[${#_den_ns[@]}]=${_den_str2}' NS'
_den_pc[${#_den_pc[@]}]=${_den_str}' '${_den_str2}
fi
fi
;;
MX) # Mail Server Record
# Domain name being serviced (wildcards not handled here)
if _den_str=$(name_fixup ${_den_tmp[0]})
then
_den_name[${#_den_name[@]}]=${_den_str}
_den_ref[${#_den_ref[@]}]=${_den_who}' '${_den_str}' MX'
fi
# Domain name of service provider
if _den_str=$(name_fixup ${_den_tmp[5]})
then
_den_name[${#_den_name[@]}]=${_den_str}
_den_ref[${#_den_ref[@]}]=${_den_who}' '${_den_str}' MXH'
_den_ns[${#_den_ns[@]}]=${_den_str}' MX'
_den_pc[${#_den_pc[@]}]=${_den_who}' '${_den_str}
fi
;;
PTR) # Reverse address record
# Special name
if _den_str=$(name_fixup ${_den_tmp[0]})
then
_den_ref[${#_den_ref[@]}]=${_den_who}' '${_den_str}' PTR'
# Host name (not a CNAME)
if _den_str2=$(name_fixup ${_den_tmp[4]})
then
_den_rev[${#_den_rev[@]}]=${_den_str}' '${_den_str2}
_den_ref[${#_den_ref[@]}]=${_den_who}' '${_den_str2}' PTRH'
_den_pc[${#_den_pc[@]}]=${_den_who}' '${_den_str}
fi
fi
;;
AAAA) # IP(v6) Address Record
if _den_str=$(name_fixup ${_den_tmp[0]})
then
_den_name[${#_den_name[@]}]=${_den_str}
_den_pair[${#_den_pair[@]}]=${_den_tmp[4]}' '${_den_str}
_den_na[${#_den_na[@]}]=${_den_str}' '${_den_tmp[4]}
_den_ref[${#_den_ref[@]}]=${_den_who}' '${_den_str}' AAAA'
else
_den_pair[${#_den_pair[@]}]=${_den_tmp[4]}' unknown.domain'
_den_na[${#_den_na[@]}]='unknown.domain '${_den_tmp[4]}
_den_ref[${#_den_ref[@]}]=${_den_who}' unknown.domain'
fi
# No processing for IPv6 addresses
_den_pc[${#_den_pc[@]}]=${_den_who}' '${_den_tmp[4]}
;;
CNAME) # Alias name record
# Nickname
if _den_str=$(name_fixup ${_den_tmp[0]})
then
_den_name[${#_den_name[@]}]=${_den_str}
_den_ref[${#_den_ref[@]}]=${_den_who}' '${_den_str}' CNAME'
_den_pc[${#_den_pc[@]}]=${_den_who}' '${_den_str}
fi
# Hostname
if _den_str=$(name_fixup ${_den_tmp[4]})
then
_den_name[${#_den_name[@]}]=${_den_str}
_den_ref[${#_den_ref[@]}]=${_den_who}' '${_den_str}' CHOST'
_den_pc[${#_den_pc[@]}]=${_den_who}' '${_den_str}
fi
;;
# TXT)
# ;;
esac
fi
done
else # Lookup error == 'A' record 'unknown address'
_den_pair[${#_den_pair[@]}]='0.0.0.0 '${_den_who}
fi
done
# Control dot array growth.
unique_lines _den_achn _den_achn # Works best, all the same.
edit_exact auth_chain _den_achn # Works best, unique items.
if [ ${#_den_achn[@]} -gt 0 ]
then
IFS=${NO_WSP}
auth_chain=( ${auth_chain[@]} ${_den_achn[@]} )
IFS=${WSP_IFS}
fi
unique_lines _den_ref _den_ref # Works best, all the same.
edit_exact ref_chain _den_ref # Works best, unique items.
if [ ${#_den_ref[@]} -gt 0 ]
then
IFS=${NO_WSP}
ref_chain=( ${ref_chain[@]} ${_den_ref[@]} )
IFS=${WSP_IFS}
fi
unique_lines _den_na _den_na
edit_exact name_address _den_na
if [ ${#_den_na[@]} -gt 0 ]
then
IFS=${NO_WSP}
name_address=( ${name_address[@]} ${_den_na[@]} )
IFS=${WSP_IFS}
fi
unique_lines _den_ns _den_ns
edit_exact name_srvc _den_ns
if [ ${#_den_ns[@]} -gt 0 ]
then
IFS=${NO_WSP}
name_srvc=( ${name_srvc[@]} ${_den_ns[@]} )
IFS=${WSP_IFS}
fi
unique_lines _den_nr _den_nr
edit_exact name_resource _den_nr
if [ ${#_den_nr[@]} -gt 0 ]
then
IFS=${NO_WSP}
name_resource=( ${name_resource[@]} ${_den_nr[@]} )
IFS=${WSP_IFS}
fi
unique_lines _den_pc _den_pc
edit_exact parent_child _den_pc
if [ ${#_den_pc[@]} -gt 0 ]
then
IFS=${NO_WSP}
parent_child=( ${parent_child[@]} ${_den_pc[@]} )
IFS=${WSP_IFS}
fi
# Update list known_pair (Address and Name).
unique_lines _den_pair _den_pair
edit_exact known_pair _den_pair
if [ ${#_den_pair[@]} -gt 0 ] # Anything new?
then
IFS=${NO_WSP}
known_pair=( ${known_pair[@]} ${_den_pair[@]} )
IFS=${WSP_IFS}
fi
# Update list of reverse pairs.
unique_lines _den_rev _den_rev
edit_exact reverse_pair _den_rev
if [ ${#_den_rev[@]} -gt 0 ] # Anything new?
then
IFS=${NO_WSP}
reverse_pair=( ${reverse_pair[@]} ${_den_rev[@]} )
IFS=${WSP_IFS}
fi
# Check indirection limit -- give up if reached.
if ! _den_lmt=$(limit_chk ${1})
then
return 0
fi
# Execution engine is LIFO. Order of pend operations is important.
# Did we define any new addresses?
unique_lines _den_address _den_address # Scrub duplicates.
edit_exact known_address _den_address # Scrub already processed.
edit_exact un_address _den_address # Scrub already waiting.
if [ ${#_den_address[@]} -gt 0 ] # Anything new?
then
uc_address=( ${uc_address[@]} ${_den_address[@]} )
pend_func expand_input_address ${_den_lmt}
_trace_log[${#_trace_log[@]}]='# # # Added '${#_den_address[@]}' unchecked address(s). # # #'
fi
# Did we find any new names?
unique_lines _den_name _den_name # Scrub duplicates.
edit_exact known_name _den_name # Scrub already processed.
edit_exact uc_name _den_name # Scrub already waiting.
if [ ${#_den_name[@]} -gt 0 ] # Anything new?
then
uc_name=( ${uc_name[@]} ${_den_name[@]} )
pend_func expand_input_name ${_den_lmt}
_trace_log[${#_trace_log[@]}]='# # # Added '${#_den_name[@]}' unchecked name(s). # # #'
fi
return 0
}
# The parse-it-yourself delegation reply
# Input is the chk_address list.
# detail_each_address <indirection_limit>
detail_each_address() {
[ ${#chk_address[@]} -gt 0 ] || return 0
unique_lines chk_address chk_address
edit_exact known_address chk_address
if [ ${#chk_address[@]} -gt 0 ]
then
known_address=( ${known_address[@]} ${chk_address[@]} )
unset chk_address[@]
fi
return 0
}
# # # Application specific output functions # # #
# Pretty print the known pairs.
report_pairs() {
echo
echo 'Known network pairs.'
col_print known_pair 2 5 30
if [ ${#auth_chain[@]} -gt 0 ]
then
echo
echo 'Known chain of authority.'
col_print auth_chain 2 5 30 55
fi
if [ ${#reverse_pair[@]} -gt 0 ]
then
echo
echo 'Known reverse pairs.'
col_print reverse_pair 2 5 55
fi
return 0
}
# Check an address against the list of blacklist servers.
# A good place to capture for GraphViz: address->status(server(reports))
# check_lists <ip_address>
check_lists() {
[ $# -eq 1 ] || return 1
local -a _cl_fwd_addr
local -a _cl_rev_addr
local -a _cl_reply
local -i _cl_rc
local -i _ls_cnt
local _cl_dns_addr
local _cl_lkup
split_ip ${1} _cl_fwd_addr _cl_rev_addr
_cl_dns_addr=$(dot_array _cl_rev_addr)'.'
_ls_cnt=${#list_server[@]}
echo ' Checking address '${1}
for (( _cl = 0 ; _cl < _ls_cnt ; _cl++ ))
do
_cl_lkup=${_cl_dns_addr}${list_server[${_cl}]}
if short_text ${_cl_lkup} _cl_reply
then
if [ ${#_cl_reply[@]} -gt 0 ]
then
echo ' Records from '${list_server[${_cl}]}
address_hits[${#address_hits[@]}]=${1}' '${list_server[${_cl}]}
_hs_RC=2
for (( _clr = 0 ; _clr < ${#_cl_reply[@]} ; _clr++ ))
do
echo ' '${_cl_reply[${_clr}]}
done
fi
fi
done
return 0
}
# # # The usual application glue # # #
# Who did it?
credits() {
echo
echo 'Advanced Bash Scripting Guide: is_spammer.bash, v2, 2004-msz'
}
# How to use it?
# (See also, "Quickstart" at end of script.)
usage() {
cat <<-'_usage_statement_'
The script is_spammer.bash requires either one or two arguments.
arg 1) May be one of:
a) A domain name
b) An IPv4 address
c) The name of a file with any mix of names
and addresses, one per line.
arg 2) May be one of:
a) A Blacklist server domain name
b) The name of a file with Blacklist server
domain names, one per line.
c) If not present, a default list of (free)
Blacklist servers is used.
d) If a filename of an empty, readable, file
is given,
Blacklist server lookup is disabled.
All script output is written to stdout.
Return codes: 0 -> All OK, 1 -> Script failure,
2 -> Something is Blacklisted.
Requires the external program 'dig' from the 'bind-9'
set of DNS programs. See: http://www.isc.org
The domain name lookup depth limit defaults to 2 levels.
Set the environment variable SPAMMER_LIMIT to change.
SPAMMER_LIMIT=0 means 'unlimited'
Limit may also be set on the command line.
If arg#1 is an integer, the limit is set to that value
and then the above argument rules are applied.
Setting the environment variable 'SPAMMER_DATA' to a filename
will cause the script to write a GraphViz graphic file.
For the development version;
Setting the environment variable 'SPAMMER_TRACE' to a filename
will cause the execution engine to log a function call trace.
_usage_statement_
}
# The default list of Blacklist servers:
# Many choices, see: http://www.spews.org/lists.html
declare -a default_servers
# See: http://www.spamhaus.org (Conservative, well maintained)
default_servers[0]='sbl-xbl.spamhaus.org'
# See: http://ordb.org (Open mail relays)
default_servers[1]='relays.ordb.org'
# See: http://www.spamcop.net/ (You can report spammers here)
default_servers[2]='bl.spamcop.net'
# See: http://www.spews.org (An 'early detect' system)
default_servers[3]='l2.spews.dnsbl.sorbs.net'
# See: http://www.dnsbl.us.sorbs.net/using.shtml
default_servers[4]='dnsbl.sorbs.net'
# See: http://dsbl.org/usage (Various mail relay lists)
default_servers[5]='list.dsbl.org'
default_servers[6]='multihop.dsbl.org'
default_servers[7]='unconfirmed.dsbl.org'
# User input argument #1
setup_input() {
if [ -e ${1} ] && [ -r ${1} ] # Name of readable file
then
file_to_array ${1} uc_name
echo 'Using filename >'${1}'< as input.'
else
if is_address ${1} # IP address?
then
uc_address=( ${1} )
echo 'Starting with address >'${1}'<'
else # Must be a name.
uc_name=( ${1} )
echo 'Starting with domain name >'${1}'<'
fi
fi
return 0
}
# User input argument #2
setup_servers() {
if [ -e ${1} ] && [ -r ${1} ] # Name of a readable file
then
file_to_array ${1} list_server
echo 'Using filename >'${1}'< as blacklist server list.'
else
list_server=( ${1} )
echo 'Using blacklist server >'${1}'<'
fi
return 0
}
# User environment variable SPAMMER_TRACE
live_log_die() {
if [ ${SPAMMER_TRACE:=} ] # Wants trace log?
then
if [ ! -e ${SPAMMER_TRACE} ]
then
if ! touch ${SPAMMER_TRACE} 2>/dev/null
then
pend_func echo $(printf '%q/n' /
'Unable to create log file >'${SPAMMER_TRACE}'<')
pend_release
exit 1
fi
_log_file=${SPAMMER_TRACE}
_pend_hook_=trace_logger
_log_dump=dump_log
else
if [ ! -w ${SPAMMER_TRACE} ]
then
pend_func echo $(printf '%q/n' /
'Unable to write log file >'${SPAMMER_TRACE}'<')
pend_release
exit 1
fi
_log_file=${SPAMMER_TRACE}
echo '' > ${_log_file}
_pend_hook_=trace_logger
_log_dump=dump_log
fi
fi
return 0
}
# User environment variable SPAMMER_DATA
data_capture() {
if [ ${SPAMMER_DATA:=} ] # Wants a data dump?
then
if [ ! -e ${SPAMMER_DATA} ]
then
if ! touch ${SPAMMER_DATA} 2>/dev/null
then
pend_func echo $(printf '%q]n' /
'Unable to create data output file >'${SPAMMER_DATA}'<')
pend_release
exit 1
fi
_dot_file=${SPAMMER_DATA}
_dot_dump=dump_dot
else
if [ ! -w ${SPAMMER_DATA} ]
then
pend_func echo $(printf '%q/n' /
'Unable to write data output file >'${SPAMMER_DATA}'<')
pend_release
exit 1
fi
_dot_file=${SPAMMER_DATA}
_dot_dump=dump_dot
fi
fi
return 0
}
# Grope user specified arguments.
do_user_args() {
if [ $# -gt 0 ] && is_number $1
then
indirect=$1
shift
fi
case $# in # Did user treat us well?
1)
if ! setup_input $1 # Needs error checking.
then
pend_release
$_log_dump
exit 1
fi
list_server=( ${default_servers[@]} )
_list_cnt=${#list_server[@]}
echo 'Using default blacklist server list.'
echo 'Search depth limit: '${indirect}
;;
2)
if ! setup_input $1 # Needs error checking.
then
pend_release
$_log_dump
exit 1
fi
if ! setup_servers $2 # Needs error checking.
then
pend_release
$_log_dump
exit 1
fi
echo 'Search depth limit: '${indirect}
;;
*)
pend_func usage
pend_release
$_log_dump
exit 1
;;
esac
return 0
}
# A general purpose debug tool.
# list_array <array_name>
list_array() {
[ $# -eq 1 ] || return 1 # One argument required.
local -a _la_lines
set -f
local IFS=${NO_WSP}
eval _la_lines=/(/ /$/{$1/[@/]/}/ /)
echo
echo "Element count "${#_la_lines[@]}" array "${1}
local _ln_cnt=${#_la_lines[@]}
for (( _i = 0; _i < ${_ln_cnt}; _i++ ))
do
echo 'Element '$_i' >'${_la_lines[$_i]}'<'
done
set +f
return 0
}
# # # 'Hunt the Spammer' program code # # #
pend_init # Ready stack engine.
pend_func credits # Last thing to print.
# # # Deal with user # # #
live_log_die # Setup debug trace log.
data_capture # Setup data capture file.
echo
do_user_args $@
# # # Haven't exited yet - There is some hope # # #
# Discovery group - Execution engine is LIFO - pend
# in reverse order of execution.
_hs_RC=0 # Hunt the Spammer return code
pend_mark
pend_func report_pairs # Report name-address pairs.
# The two detail_* are mutually recursive functions.
# They also pend expand_* functions as required.
# These two (the last of ???) exit the recursion.
pend_func detail_each_address # Get all resources of addresses.
pend_func detail_each_name # Get all resources of names.
# The two expand_* are mutually recursive functions,
#+ which pend additional detail_* functions as required.
pend_func expand_input_address 1 # Expand input names by address.
pend_func expand_input_name 1 # #xpand input addresses by name.
# Start with a unique set of names and addresses.
pend_func unique_lines uc_address uc_address
pend_func unique_lines uc_name uc_name
# Separate mixed input of names and addresses.
pend_func split_input
pend_release
# # # Pairs reported -- Unique list of IP addresses found
echo
_ip_cnt=${#known_address[@]}
if [ ${#list_server[@]} -eq 0 ]
then
echo 'Blacklist server list empty, none checked.'
else
if [ ${_ip_cnt} -eq 0 ]
then
echo 'Known address list empty, none checked.'
else
_ip_cnt=${_ip_cnt}-1 # Start at top.
echo 'Checking Blacklist servers.'
for (( _ip = _ip_cnt ; _ip >= 0 ; _ip-- ))
do
pend_func check_lists $( printf '%q/n' ${known_address[$_ip]} )
done
fi
fi
pend_release
$_dot_dump # Graphics file dump
$_log_dump # Execution trace
echo
##############################
# Example output from script #
##############################
:<<-'_is_spammer_outputs_'
./is_spammer.bash 0 web4.alojamentos7.com
Starting with domain name >web4.alojamentos7.com<
Using default blacklist server list.
Search depth limit: 0
.:....::::...:::...:::.......::..::...:::.......::
Known network pairs.
66.98.208.97 web4.alojamentos7.com.
66.98.208.97 ns1.alojamentos7.com.
69.56.202.147 ns2.alojamentos.ws.
66.98.208.97 alojamentos7.com.
66.98.208.97 web.alojamentos7.com.
69.56.202.146 ns1.alojamentos.ws.
69.56.202.146 alojamentos.ws.
66.235.180.113 ns1.alojamentos.org.
66.235.181.192 ns2.alojamentos.org.
66.235.180.113 alojamentos.org.
66.235.180.113 web6.alojamentos.org.
216.234.234.30 ns1.theplanet.com.
12.96.160.115 ns2.theplanet.com.
216.185.111.52 mail1.theplanet.com.
69.56.141.4 spooling.theplanet.com.
216.185.111.40 theplanet.com.
216.185.111.40 www.theplanet.com.
216.185.111.52 mail.theplanet.com.
Checking Blacklist servers.
Checking address 66.98.208.97
Records from dnsbl.sorbs.net
"Spam Received See: http://www.dnsbl.sorbs.net/lookup.shtml?66.98.208.97"
Checking address 69.56.202.147
Checking address 69.56.202.146
Checking address 66.235.180.113
Checking address 66.235.181.192
Checking address 216.185.111.40
Checking address 216.234.234.30
Checking address 12.96.160.115
Checking address 216.185.111.52
Checking address 69.56.141.4
Advanced Bash Scripting Guide: is_spammer.bash, v2, 2004-msz
_is_spammer_outputs_
exit ${_hs_RC}
####################################################
# The script ignores everything from here on down #
#+ because of the 'exit' command, just above. #
####################################################
Quickstart
==========
Prerequisites
Bash version 2.05b or 3.00 (bash --version)
A version of Bash which supports arrays. Array
support is included by default Bash configurations.
'dig,' version 9.x.x (dig $HOSTNAME, see first line of output)
A version of dig which supports the +short options.
See: dig_wrappers.bash for details.
Optional Prerequisites
'named,' a local DNS caching program. Any flavor will do.
Do twice: dig $HOSTNAME
Check near bottom of output for: SERVER: 127.0.0.1#53
That means you have one running.
Optional Graphics Support
'date,' a standard *nix thing. (date -R)
dot Program to convert graphic description file to a
diagram. (dot -V)
A part of the Graph-Viz set of programs.
See: [http://www.research.att.com/sw/tools/graphviz||GraphViz]
'dotty,' a visual editor for graphic description files.
Also a part of the Graph-Viz set of programs.
Quick Start
In the same directory as the is_spammer.bash script;
Do: ./is_spammer.bash
Usage Details
1. Blacklist server choices.
(a) To use default, built-in list: Do nothing.
(b) To use your own list:
i. Create a file with a single Blacklist server
domain name per line.
ii. Provide that filename as the last argument to
the script.
(c) To use a single Blacklist server: Last argument
to the script.
(d) To disable Blacklist lookups:
i. Create an empty file (touch spammer.nul)
Your choice of filename.
ii. Provide the filename of that empty file as the
last argument to the script.
2. Search depth limit.
(a) To use the default value of 2: Do nothing.
(b) To set a different limit:
A limit of 0 means: no limit.
i. export SPAMMER_LIMIT=1
or whatever limit you want.
ii. OR provide the desired limit as the first
argument to the script.
3. Optional execution trace log.
(a) To use the default setting of no log output: Do nothing.
(b) To write an execution trace log:
export SPAMMER_TRACE=spammer.log
or whatever filename you want.
4. Optional graphic description file.
(a) To use the default setting of no graphic file: Do nothing.
(b) To write a Graph-Viz graphic description file:
export SPAMMER_DATA=spammer.dot
or whatever filename you want.
5. Where to start the search.
(a) Starting with a single domain name:
i. Without a command line search limit: First
argument to script.
ii. With a command line search limit: Second
argument to script.
(b) Starting with a single IP address:
i. Without a command line search limit: First
argument to script.
ii. With a command line search limit: Second
argument to script.
(c) Starting with (mixed) multiple name(s) and/or address(es):
Create a file with one name or address per line.
Your choice of filename.
i. Without a command line search limit: Filename as
first argument to script.
ii. With a command line search limit: Filename as
second argument to script.
6. What to do with the display output.
(a) To view display output on screen: Do nothing.
(b) To save display output to a file: Redirect stdout to a filename.
(c) To discard display output: Redirect stdout to /dev/null.
7. Temporary end of decision making.
press RETURN
wait (optionally, watch the dots and colons).
8. Optionally check the return code.
(a) Return code 0: All OK
(b) Return code 1: Script setup failure
(c) Return code 2: Something was blacklisted.
9. Where is my graph (diagram)?
The script does not directly produce a graph (diagram).
It only produces a graphic description file. You can
process the graphic descriptor file that was output
with the 'dot' program.
Until you edit that descriptor file, to describe the
relationships you want shown, all that you will get is
a bunch of labeled name and address nodes.
All of the script's discovered relationships are within
a comment block in the graphic descriptor file, each
with a descriptive heading.
The editing required to draw a line between a pair of
nodes from the information in the descriptor file may
be done with a text editor.
Given these lines somewhere in the descriptor file:
# Known domain name nodes
N0000 [label="guardproof.info."] ;
N0002 [label="third.guardproof.info."] ;
# Known address nodes
A0000 [label="61.141.32.197"] ;
/*
# Known name->address edges
NA0000 third.guardproof.info. 61.141.32.197
# Known parent->child edges
PC0000 guardproof.info. third.guardproof.info.
*/
Turn that into the following lines by substituting node
identifiers into the relationships:
# Known domain name nodes
N0000 [label="guardproof.info."] ;
N0002 [label="third.guardproof.info."] ;
# Known address nodes
A0000 [label="61.141.32.197"] ;
# PC0000 guardproof.info. third.guardproof.info.
N0000->N0002 ;
# NA0000 third.guardproof.info. 61.141.32.197
N0002->A0000 ;
/*
# Known name->address edges
NA0000 third.guardproof.info. 61.141.32.197
# Known parent->child edges
PC0000 guardproof.info. third.guardproof.info.
*/
Process that with the 'dot' program, and you have your
first network diagram.
In addition to the conventional graphic edges, the
descriptor file includes similar format pair-data that
describes services, zone records (sub-graphs?),
blacklisted addresses, and other things which might be
interesting to include in your graph. This additional
information could be displayed as different node
shapes, colors, line sizes, etc.
The descriptor file can also be read and edited by a
Bash script (of course). You should be able to find
most of the functions required within the
"is_spammer.bash" script.
# End Quickstart.
Additional Note
========== ====
Michael Zick points out that there is a "makeviz.bash" interactive
Web site at rediris.es. Can't give the full URL, since this is not
a publically accessible site.
%%%&&&is_spammer.bash&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@is-spammer.sh@@@!!!************************************************************************************
#! /bin/bash
# is-spammer.sh: 鉴别一个垃圾邮件域
# $Id: is-spammer, v 1.4 2004/09/01 19:37:52 mszick Exp $
# 上边这行是RCS ID信息.
#
# 这是附件中捐献脚本is_spammer.bash
#+ 的一个简单版本.
# is-spammer <domain.name>
# 使用外部程序: 'dig'
# 测试版本: 9.2.4rc5
# 使用函数.
# 使用IFS来分析分配在数组中的字符串.
# 做一些有用的事: 检查e-mail黑名单.
# 使用来自文本体中的domain.name:
# http://www.good_stuff.spammer.biz/just_ignore_everything_else
# ^^^^^^^^^^^
# 或者使用来自任意e-mail地址的domain.name:
# Really_Good_Offer@spammer.biz
#
# 并将其作为这个脚本的唯一参数.
#(另: 你的Inet连接应该保证连接好)
#
# 这样, 在上边两个实例中调用这个脚本:
# is-spammer.sh spammer.biz
# Whitespace == :Space:Tab:Line Feed:Carriage Return:
WSP_IFS=$'/x20'$'/x09'$'/x0A'$'/x0D'
# No Whitespace == Line Feed:Carriage Return
No_WSP=$'/x0A'$'/x0D'
# 域分隔符为点分10进制ip地址
ADR_IFS=${No_WSP}'.'
# 取得dns文本资源记录.
# get_txt <error_code> <list_query>
get_txt() {
# 分析在"."中分配的$1.
local -a dns
IFS=$ADR_IFS
dns=( $1 )
IFS=$WSP_IFS
if [ "${dns[0]}" == '127' ]
then
# 查看此处是否有原因.
echo $(dig +short $2 -t txt)
fi
}
# 取得dns地址资源纪录.
# chk_adr <rev_dns> <list_server>
chk_adr() {
local reply
local server
local reason
server=${1}${2}
reply=$( dig +short ${server} )
# 假设应答可能是一个错误码 . . .
if [ ${#reply} -gt 6 ]
then
reason=$(get_txt ${reply} ${server} )
reason=${reason:-${reply}}
fi
echo ${reason:-' not blacklisted.'}
}
# 需要从名字中取得 IP 地址.
echo 'Get address of: '$1
ip_adr=$(dig +short $1)
dns_reply=${ip_adr:-' no answer '}
echo ' Found address: '${dns_reply}
# 一个可用的应答至少是4个数字加上3个点.
if [ ${#ip_adr} -gt 6 ]
then
echo
declare query
# 通过点中的分配进行分析.
declare -a dns
IFS=$ADR_IFS
dns=( ${ip_adr} )
IFS=$WSP_IFS
# 用8进制表示法将dns查询循序记录起来.
rev_dns="${dns[3]}"'.'"${dns[2]}"'.'"${dns[1]}"'.'"${dns[0]}"'.'
# 查看: http://www.spamhaus.org (传统地址, 维护的很好)
echo -n 'spamhaus.org says: '
echo $(chk_adr ${rev_dns} 'sbl-xbl.spamhaus.org')
# 查看: http://ordb.org (开放转发Open mail relay)
echo -n ' ordb.org says: '
echo $(chk_adr ${rev_dns} 'relays.ordb.org')
# 查看: http://www.spamcop.net/ (你可以在这里报告spammer)
echo -n ' spamcop.net says: '
echo $(chk_adr ${rev_dns} 'bl.spamcop.net')
# # # 其他的黑名单操作 # # #
# 查看: http://cbl.abuseat.org.
echo -n ' abuseat.org says: '
echo $(chk_adr ${rev_dns} 'cbl.abuseat.org')
# 查看: http://dsbl.org/usage (不同的邮件转发mail relay)
echo
echo 'Distributed Server Listings'
echo -n ' list.dsbl.org says: '
echo $(chk_adr ${rev_dns} 'list.dsbl.org')
echo -n ' multihop.dsbl.org says: '
echo $(chk_adr ${rev_dns} 'multihop.dsbl.org')
echo -n 'unconfirmed.dsbl.org says: '
echo $(chk_adr ${rev_dns} 'unconfirmed.dsbl.org')
else
echo
echo 'Could not use that address.'
fi
exit 0
# 练习:
# -----
# 1) 检查脚本参数,
# 并且如果必要的话, 可以使用合适的错误消息退出.
# 2) 察看调用这个脚本的时候是否在线,
# 并且如果必要的话, 可以使用合适的错误消息退出.
# 3) 用一般变量来替换掉"硬编码"的BHL domain.
# 4) 通过对'dig'命令使用"+time="选项
# 来给这个脚本设置一个暂停.
%%%&&&is-spammer.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@keypress.sh@@@!!!************************************************************************************
#!/bin/bash
# keypress.sh: 检测用户按键("hot keys").
echo
old_tty_settings=$(stty -g) # 保存老的设置(为什么?).
stty -icanon
Keypress=$(head -c1) # 或者使用$(dd bs=1 count=1 2> /dev/null)
# 在非GNU系统上
echo
echo "Key pressed was /""$Keypress"/"."
echo
stty "$old_tty_settings" # 恢复老的设置.
# 感谢, Stephane Chazelas.
exit 0
%%%&&&keypress.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@kill-byname.sh@@@!!!************************************************************************************
#!/bin/bash
# kill-byname.sh: 通过名字kill进程.
# 与脚本kill-process.sh相比较.
# 例如,
#+ 试一下 "./kill-byname.sh xterm" --
#+ 并且查看你系统上的所有xterm都将消失.
# 警告:
# -----
# 这是一个非常危险的脚本.
# 运行它的时候一定要小心. (尤其是以root身份运行时)
#+ 因为运行这个脚本可能会引起数据丢失或产生其他一些不好的效果.
E_BADARGS=66
if test -z "$1" # 没有参数传递进来?
then
echo "Usage: `basename $0` Process(es)_to_kill"
exit $E_BADARGS
fi
PROCESS_NAME="$1"
ps ax | grep "$PROCESS_NAME" | awk '{print $1}' | xargs -i kill {} 2&>/dev/null
# ^^ ^^
# -----------------------------------------------------------
# 注意:
# -i 参数是xargs命令的"替换字符串"选项.
# 大括号对的地方就是替换点.
# 2&>/dev/null 将会丢弃不需要的错误消息.
# -----------------------------------------------------------
exit $?
# 在这个脚本中, "killall"命令具有相同的效果,
#+ 但是这么做就没有教育意义了.
%%%&&&kill-byname.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@kill-process.sh@@@!!!************************************************************************************
#!/bin/bash
# kill-process.sh
NOPROCESS=2
process=xxxyyyzzz # 使用不存在的进程.
# 只不过是为了演示...
# ... 并不想在这个脚本中杀掉任何真正的进程.
#
# 举个例子, 如果你想使用这个脚本来断线Internet,
# process=pppd
t=`pidof $process` # 取得$process的pid(进程id).
# 'kill'只能使用pid(不能用程序名)作为参数.
if [ -z "$t" ] # 如果没这个进程, 'pidof' 返回空.
then
echo "Process $process was not running."
echo "Nothing killed."
exit $NOPROCESS
fi
kill $t # 对于某些顽固的进程可能需要使用'kill -9'.
# 这里需要做一个检查, 看看进程是否允许自身被kill.
# 或许另一个 " t=`pidof $process` " 或许 ...
# 整个脚本都可以使用下边这句来替换:
# kill $(pidof -x process_name)
# 或者
# killall process_name
# 但是这就没有教育意义了.
exit 0
%%%&&&kill-process.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@length.sh@@@!!!************************************************************************************
#!/bin/bash
# length.sh
E_NO_ARGS=65
if [ $# -eq 0 ] # 这个演示脚本必须有命令行参数.
then
echo "Please invoke this script with one or more command-line arguments."
exit $E_NO_ARGS
fi
var01=abcdEFGH28ij
echo "var01 = ${var01}"
echo "Length of var01 = ${#var01}"
# 现在, 让我们试试在变量中嵌入一个空格.
var02="abcd EFGH28ij"
echo "var02 = ${var02}"
echo "Length of var02 = ${#var02}"
echo "Number of command-line arguments passed to script = ${#@}"
echo "Number of command-line arguments passed to script = ${#*}"
exit 0
%%%&&&length.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@letter-count2.sh@@@!!!************************************************************************************
#! /bin/sh
# letter-count2.sh: 在文本文件中计算字符的出现次数.
#
# 由nyal [nyal@voila.fr]编写.
# 授权使用.
# 本文作者重新注释.
# 版本 1.1: 经过修改可用于gawk 3.1.3.
# (也可用于awk的早期版本.)
INIT_TAB_AWK=""
# 初始化awk脚本的参数.
count_case=0
FILE_PARSE=$1
E_PARAMERR=65
usage()
{
echo "Usage: letter-count.sh file letters" 2>&1
# 比如: ./letter-count2.sh filename.txt a b c
exit $E_PARAMERR # 传递到脚本的参数个数不够.
}
if [ ! -f "$1" ] ; then
echo "$1: No such file." 2>&1
usage # 打印使用信息并退出.
fi
if [ -z "$2" ] ; then
echo "$2: No letters specified." 2>&1
usage
fi
shift # 指定的字符.
for letter in `echo $@` # for循环遍历 . . .
do
INIT_TAB_AWK="$INIT_TAB_AWK tab_search[${count_case}] = /"$letter/"; final_tab[${count_case}] = 0; "
# 作为参数传递到下边的awk脚本中.
count_case=`expr $count_case + 1`
done
# 调试:
# echo $INIT_TAB_AWK;
cat $FILE_PARSE |
# 将目标文件通过管道传递下边的awk脚本中.
# ----------------------------------------------------------------------------------
# 下边是本脚本的早期版本使用的方法:
# awk -v tab_search=0 -v final_tab=0 -v tab=0 -v nb_letter=0 -v chara=0 -v chara2=0 /
awk /
"BEGIN { $INIT_TAB_AWK } /
{ split(/$0, tab, /"/"); /
for (chara in tab) /
{ for (chara2 in tab_search) /
{ if (tab_search[chara2] == tab[chara]) { final_tab[chara2]++ } } } } /
END { for (chara in final_tab) /
{ print tab_search[chara] /" => /" final_tab[chara] } }"
# ----------------------------------------------------------------------------------
# 不是所有的都那么复杂, 只是 . . .
#+ for循环, if条件判断, 和几个指定函数而已.
exit $?
# 与脚本letter-count.sh相比较.
%%%&&&letter-count2.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
***@@@letter-count.sh@@@!!!************************************************************************************
#!/bin/bash
# letter-count.sh: 统计一个文本文件中某些字母出现的次数.
# 由Stefano Palmeri所编写.
# 经过授权可以使用在本书中.
# 本书作者做了少许修改.
MINARGS=2 # 本脚本至少需要2个参数.
E_BADARGS=65
FILE=$1
let LETTERS=$#-1 # 指定了多少个字母(作为命令行参数).
# (从命令行参数的个数中减1.)
show_help(){
echo
echo Usage: `basename $0` file letters
echo Note: `basename $0` arguments are case sensitive.
echo Example: `basename $0` foobar.txt G n U L i N U x.
echo
}
# 检查参数个数.
if [ $# -lt $MINARGS ]; then
echo
echo "Not enough arguments."
echo
show_help
exit $E_BADARGS
fi
# 检查文件是否存在.
if [ ! -f $FILE ]; then
echo "File /"$FILE/" does not exist."
exit $E_BADARGS
fi
# 统计字母出现的次数.
for n in `seq $LETTERS`; do
shift
if [[ `echo -n "$1" | wc -c` -eq 1 ]]; then # 检查参数.
echo "$1" -/> `cat $FILE | tr -cd "$1" | wc -c` # 统计.
else
echo "$1 is not a single char."
fi
done
exit $?
# 这个脚本在功能上与letter-count2.sh完全相同,
#+ 但是运行得更快.
# 为什么?
%%%&&&letter-count.sh&&&%%%>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>
- 330pics-shell scripts_fourth
- 330pics-shell scripts_third
- 330pics shell scripts_fifth
- 330pics shell scripts_sixth
- 330pics-shell scripts-first
- 330pics-shell scripts-second
- pics
- GSM pics
- SIM PICS
- MMS PICS
- vedio PICS
- 556 Instagram pics d
- WCDMA protocol PICS
- WCDMA RF PICS
- LTE Protocol PICS
- LTE RF PICS
- SIM PICS 2
- IMS volte PICS
- 330pics-shell scripts_third
- (转)基于API的录音机程序
- 解决 惠普HP CQ40 519TX 安装XP系统更新驱动后,有爆破音。
- Flex构建类Google下拉框的联想功能
- 基于开源软件做个和GTalk聊天的小命令行程序[ZT]
- 330pics-shell scripts_fourth
- C# 调用clicktell提供的SMS服务
- 330pics shell scripts_fifth
- java设计模式之Singleton
- 累
- 编译原理
- 当没有用 EXISTS 引入子查询时,在选择列表中只能指定一个表达式。
- ASP.NET中WEB用户控件和自定义控件
- test