Optimize shell and awk script
- by bryan
I am using a combination of a shell script, awk script and a find command to perform multiple text replacements in hundreds of files. The files sizes vary between a few hundred bytes and 20 kbytes. 
I am looking for a way to speed up this script. 
I am using cygwin.
The shell script - 
#!/bin/bash
if [ $# = 0 ]; then
 echo "Argument expected"
 exit 1
fi
while [ $# -ge 1 ]
do
   if [ ! -f $1 ]; then
     echo "No such file as $1"
     exit 1
   fi
  awk -f ~/scripts/parse.awk $1  > ${1}.$$
   if [ $? != 0 ]; then
      echo "Something went wrong with the script"
     rm ${1}.$$
      exit 1
   fi
mv ${1}.$$ $1
shift
done
The awk script (simplified) - 
#! /usr/bin/awk -f
/HHH.Web/{
    if ( index($0,"Email") == 0)  {
        sub(/HHH.Web/,"HHH.Web.Email");
    }
    printf("%s\r\n",$0); 
    next;
}
The command line 
find .  -type f  | xargs ~/scripts/run_parser.sh