While using mv or rm command I get command line length error (Argument list too long error). How do I find out current running shell command line length limitations? How do I overcomes these limitations while writing UNIX / BSD / Linux shell utilities?
All shell have / has a limit for the command line length. UNIX / Linux / BSD system has a limit on how many bytes can be used for the command line argument and environment variables. When you start a new process or type a command these limitations are applied and you will see an error message as follows on screen:
Argument list too long
How do I find out current command line length limitations?
Type the following command (works under Linux / UNIX / BSD operating systems):
$ getconf ARG_MAX
Sample output:
262144
BSD operating system also supports following command:
$ sysctl kern.argmax
Sample output:
kern.argmax=262144
To get accurate picture about limitation type the following command (hat tip to Jeff):
$ echo $(( $(getconf ARG_MAX) - $(env | wc -c) ))
Output:
261129
How do overcome shell command line length?
You have following option to get around these limitations:
- Use find or xargs command
- Use shell for / while loop
find command example to get rid of “argument list too long” error
$ find /nas/data/accounting/ -type f -exec ls -l {} \;
$ find /nas/data/accounting/ -type f -exec /bin/rm -f {} \;
xargs command example to get rid of “argument list too long” error
$ echo /nas/data/accounting/* | xargs ls -l
$ echo /nas/data/accounting/* | xargs /bin/rm -f
while loop example to get rid of “argument list too long” error
ls -1 /nas/data/accounting/ | while read file; do mv /nas/data/accounting/$file /local/disk/ ; done
Alternatively, you can combine above methods:
find /nas/data/accounting/ -type f | while read file do mv /nas/data/accounting/$file /local/disk/ done
time command – give resource usage
Use time command to find out exact system resource usage for each command:
$ time find blah blah
$ time ls -1 blah | while read file; do #blah on $file; done
Further readings:
- Your shell documentation
- man pages ksh, bash, getconf, sysconf, sysctl, find, and xargs
🐧 Get the latest tutorials on Linux, Open Source & DevOps via:
- RSS feed or Weekly email newsletter
- Share on Twitter • Facebook • 9 comments... add one ↓
Category | List of Unix and Linux commands |
---|---|
File Management | cat |
Firewall | Alpine Awall • CentOS 8 • OpenSUSE • RHEL 8 • Ubuntu 16.04 • Ubuntu 18.04 • Ubuntu 20.04 |
Network Utilities | dig • host • ip • nmap |
OpenVPN | CentOS 7 • CentOS 8 • Debian 10 • Debian 8/9 • Ubuntu 18.04 • Ubuntu 20.04 |
Package Manager | apk • apt |
Processes Management | bg • chroot • cron • disown • fg • jobs • killall • kill • pidof • pstree • pwdx • time |
Searching | grep • whereis • which |
User Information | groups • id • lastcomm • last • lid/libuser-lid • logname • members • users • whoami • who • w |
WireGuard VPN | Alpine • CentOS 8 • Debian 10 • Firewall • Ubuntu 20.04 |
When using ‘+’ as end for find(1), you emulate xargs behavior, in that it will only pass the arguments to the command when the argument list is saturated, getting rid of numerous forks.
Brief & precise, very helpful and just what I was looking for. Thanks.
It is a small world… I just googled something unrelated and found this page. Then I noticed that command and realized you linked my website. Thanks!
Hi, thanks for your post. But I guess the “echo * | xargs cmd” would fail because the asterisk will be replaced with all file names and then the echo command will be faced with an “argument list too long” error. Again the find command helps: find -type f | xargs rm -f
:)
Another way of staying under the limit is to use your resources wisely. Don’t forget that “$ /bin/rm /nas/data/accounting/*” will, after globbing, prepend every filename with “/nas/data/accounting/”, wasting 21 characters per filename, dramatically increasing the length of your command line. Instead, “$ pushd /nas/data/accounting/; /bin/rm *; popd”. Also, I cannot stress enough how important it is to actually test your examples. In addition to the problem with your xargs example, neither of the while loop samples will work, because you specified “/nas/data/accounting” in both the find/ls and the rm. This will duplicate the path in the rm and cause the rm to fail, if you’re lucky. If you’re unlucky it will succeed, and delete the wrong file.
With Linux 2.6.23, ARG_MAX is not hardcoded anymore. It is limited to a 1/4-th of the stack size (ulimit -s), which ensures that the program still can run at all.
getconf ARG_MAX might still report the former limit (being careful about applications or glibc not catching up, but especially because the kernel still defines it)
reference: http://www.in-ulm.de/~mascheck/various/argmax/
Thanks for the information. Also try Long Path Tool. It helped me with Error 1320 in Win 7. :)
For the common case, GNU `find` (but not BSD `find`) also offers a `-delete` parameter/command. So your
$ find /nas/data/accounting/ -type f -exec /bin/rm -f {} \;
example can become
$ find /nas/data/accounting/ -type f -delete
I think most OS are limited to that number unless the software you are using can go around that. I’ve been using GS RichCopy 360 in dealing with long path file errors and it works very well. Any other suggestions that can be tested out?