I'm using the following to count the number of files in a directory, and its subdirectories:
find . -type f | wc -l
But I have half a million files in there, and the count takes a long time.
Is there a faster way to get a count of the number of files, that doesn't involve piping a huge amount of text to something that counts lines? It seems like an inefficient way to do things.
If you have this on a dedicated file-system, or you have a steady number of files overhead, you may be able to get a rough enough count of the number of files by looking at the number of inodes in the file-system via "df -i":
On my test box above I have 75,885 inodes allocated. However, these inodes are not just files, they are also directories. For example:
NOTE: Not all file-systems maintain inode counts the same way. ext2/3/4 will all work, however btrfs always reports 0.
If you have to differentiate files from directories, you're going to have to walk the file-system and "stat" each one to see if it's a file, directory, sym-link, etc... The biggest issue here is not the piping of all the text to "wc", but seeking around among all the inodes and directory entries to put that data together.
Other than the inode table as shown by "df -i", there really is no database of how many files there are under a given directory. However, if this information is important to you, you could create and maintain such a database by having your programs increment a number when they create a file in this directory and decrement it when deleted. If you don't control the programs that create them, this isn't an option.
I wrote a custom file-counting program for this StackOverflow question: https://stackoverflow.com/questions/1427032/fast-linux-file-count-for-a-large-number-of-files
You can find the GitHub repo here if you'd like to browse, download, or contribute: https://github.com/ChristopherSchultz/fast-file-count
If you want to count recursively the number of files in a directory the locate command is the fastet one I know, assumed you have an up-to-date database (sudo update database .. made per default via chron job every day). However, you can speed up the command if you avoid the grep pipe.
See man locate:
So the fastest command is:
I would also try:
And then process the output, reducing into a count for the directories.
This is especially useful if you anticipate the directory structure.
if you have locate installed you can use
more on locate, you can play with
or to get a result filesystem-wide
It will be much much faster than find if you got many files.
the only drawback is, it also counts directories
And I recommend using plocate https://plocate.sesse.net/
Parallelize it. Run a separate
find
command for each subdirectory and run them at the same time. Can automate this usingxargs
.Try this handy little Python script to see if its any faster.
Andrew