http://qs321.pair.com?node_id=11115417


in reply to glob() and dot files

This glob thing can be a problem. A long time ago I got tripped up with the 3 versions of glob that were in use at that time in the ActiveState version of Perl that I was using. I changed my code to use readdir() and that solved the problem.

Nowadays, Perl glob is a lot more uniform and well behaved. This prints all simple files, but skips directories.

my @files = glob ('*.*'); print "",join("\n",@files),"\n"
For what you want, I would consider File::Find.
Consider this code also.
use strict; use warnings; opendir (my $dir, ".") or die "unable to open current directory $_"; my @files; my @directories; foreach my $file (grep{ ($_ ne '.') and ($_ ne '..')} readdir $dir) { if (-f $file ) {push @files, $file;} elsif (-d $file) {push @directories, $file;} else { die "weird beast found! $file"} } print "@files\n"; print "@directories\n";
I think in Unix there can be special things that are not simple files or directories. I would use a file test to see what this name actually means.
Note that if this is not the current directory, you need to spec the full path name for file tests.

Update: File operations like "open file" or "open directory" are "expensive" in terms of performance. I would expect my code to run faster than the OP's code, but I did not benchmark this in any serious way. If the directories are small and this is not done that often, I don't think that will make any difference at all. Also be aware that there is a special variable for repeated file tests, "_". like  elsif (-d _) {do something{ That tests the structure returned by the previous file test operation for a different flag.

Overall, unless there is a performance or other problem (special kinds of files), I see no problem with the OP's code.