The other day I was searching for a particular email file in an Maildir on one of our mail servers. Knowing only the From address, I thought I’d use
grep. Unfortunately it failed and gave me an error: /bin/grep: Argument list too long. Here is how to overcome this grep error…
/bin/grep: Argument list too long
“Argument list too long” indicates there are too many arguments into a single command which hits the ARG_MAX limit. How to grep through a large number of files? For example:
grep -r "example\.com" * -bash: /bin/grep: Argument list too longCode language: Bash (bash)
Well, this was no wonder, since there are over 190500 files in that particular Maildir:
ls -1 | wc -l 190516Code language: Bash (bash)
There are several ways of overcoming this error message. One, for instance is:
find . -type f | xargs grep "example\.com"Code language: Bash (bash)
A second option to overcome the error is: substitute the “
*” (asterisk) with a “
.” (dot), like:
grep -r "example\.com" .Code language: Bash (bash)
In newer versions of grep you can omit the “
.“, as the current directory is implied.
Did you know you can use grep on Windows? This comes in handy for forensic log parsing and analysis on Windows Server IIS.
Argument list too long when deleting files in Bash (Bonus!)
Resolve: /bin/rm: Argument list too long.
If you’re trying to delete a lot of files in Bash, this may also fail with an “Argument list too long” error message:
~/rm-test$ rm * -bash: /bin/rm: Argument list too longCode language: Bash (bash)
“Argument list too long” indicates when a user feeds too many arguments into a single command which hits the
ARG_MAX limit. The
ARG_MAX defines the maximum length of arguments to the exec function. Other commands subject to this limitation are commands such as
mv, and so on. (thanks Hayden James).
There are a couple of commands you can use to resolve this error and to remove those many thousands files. In my following examples, I have a dir
~/rm-test with 200.000 files, that I’ll use (and I’ll
time the commands for speed measurement).
time find . -type f -delete real 0m2.484s user 0m0.120s sys 0m2.341sCode language: Bash (bash)
time ls -1 | xargs rm -f real 0m2.988s user 0m0.724s sys 0m2.375sCode language: Bash (bash)
mkdir empty time rsync -a --delete empty/ rm-test/ real 0m2.880s user 0m0.170s sys 0m2.642sCode language: Bash (bash)
Or remove the entire directory:
time rm -r rm-test/ real 0m2.776s user 0m0.220s sys 0m2.510sCode language: Bash (bash)
In Linux, you can recreate that directory with the
mkdir command if you still need it. Create 200.000 small test file with:
dd if=/dev/urandom bs=1024 count=200000 | split -a 4 -b 1k - file.
As you can see, the timing results don’t differ that much. In this test, using
-delete is the fastest solution, followed by
rm -Rf. The slowest is using
ls-1 | xargs rm -f. The downside of
rsync is that 1) it is CPU heavy, and 2) you need to create an empty directory first.
I had been struggling with this for quite some time. This helped. Thanks a lot.
Great to hear my post helped you with this Mohini, you’re welcome!