App-Fetchware
view release on metacpan or search on metacpan
t/Test-Fetchware.t view on Meta::CPAN
#!perl
# Test-Fetchware.t tests Test::Fetchware's utility subroutines, which
# provied helper functions such as logging and file & dirlist downloading.
use strict;
use warnings;
use 5.010001;
# Set a umask of 022 just like bin/fetchware does. Not all fetchware tests load
# bin/fetchware, and so all fetchware tests must set a umask of 0022 to ensure
# that any files fetchware creates during testing pass fetchware's safe_open()
# security checks.
umask 0022;
# Test::More version 0.98 is needed for proper subtest support.
use Test::More 0.98 tests => '11'; #Update if this changes.
use File::Spec::Functions qw(splitpath catfile rel2abs tmpdir catdir);
use Path::Class;
use URI::Split 'uri_split';
use Cwd 'cwd';
use File::Temp qw(tempdir tempfile);
use File::Path qw(remove_tree make_path);
use App::Fetchware::Config ':CONFIG';
# Set PATH to a known good value.
$ENV{PATH} = '/usr/local/bin:/usr/bin:/bin';
# Delete *bad* elements from environment to make it safer as recommended by
# perlsec.
delete @ENV{qw(IFS CDPATH ENV BASH_ENV)};
# Test if I can load the module "inside a BEGIN block so its functions are exported
# and compile-time, and prototypes are properly honored."
# There is no ':OVERRIDE_START' to bother importing.
BEGIN { use_ok('Test::Fetchware', ':TESTING'); }
# Print the subroutines that App::Fetchware imported by default when I used it.
note("App::Fetchware's default imports [@Test::Fetchware::EXPORT_OK]");
# make_test_dist()'s test need access to bin/fetchware's cmd_install() to
# actually fully test it, so import it as bin/fetchware's own test suite does.
BEGIN {
my $fetchware = 'fetchware';
use lib 'bin';
require $fetchware;
fetchware->import(':TESTING');
ok(defined $INC{$fetchware}, 'checked bin/fetchware loading and import')
}
###BUGALERT### Add tests for :TESTING subs that have no tests!!!
subtest 'TESTING export what they should' => sub {
my @expected_testing_exports = qw(
eval_ok
print_ok
fork_ok
fork_not_ok
skip_all_unless_release_testing
make_clean
make_test_dist
md5sum_file
expected_filename_listing
verbose_on
export_ok
end_ok
add_prefix_if_nonroot
create_test_fetchwarefile
rmdashr_ok
);
# sort them to make the testing their equality very easy.
@expected_testing_exports = sort @expected_testing_exports;
my @sorted_testing_tag = sort @{$Test::Fetchware::EXPORT_TAGS{TESTING}};
is_deeply(\@sorted_testing_tag, \@expected_testing_exports,
'checked for correct exports.');
};
subtest 'test print_ok()' => sub {
# Can't easily test the exceptions print_ok() throws, because they're if
# open()ing a scalar ref fails, and if calling close() actually failes,
# which can't easily be forced to fail.
# Test print_ok() string message.
my $test_message = 'A test message';
print_ok(sub {print $test_message},
$test_message, 'checked print_ok() string message success');
# Test print_ok() regex.
print_ok(sub {print $test_message},
qr/$test_message/, 'checked print_ok() regex message success');
print_ok(sub {print $test_message},
sub {return 1 if $_[0] eq $test_message; return;},
'checked print_ok() simple coderef success');
};
subtest 'test make_test_dist()' => sub {
###HOWTOTEST### How do I test for mkdir() failure, open() failure, and
#Archive::Tar->create_archive() failure?
my $file_name = 'test-dist';
my $ver_num = '1.00';
my $retval = make_test_dist(file_name => $file_name, ver_num => $ver_num);
is(file($retval)->basename(), "$file_name-$ver_num.fpkg",
'check make_test_dist() success.');
ok(-e $retval, 'check make_test_dist() existence.');
ok(unlink $retval, 'checked make_test_dist() cleanup');
# Test more than one call as used in t/bin-fetchware-upgrade-all.t
my @filenames = qw(test-dist another-dist);
my @retvals;
for my $filename (@filenames) {
my $retval = make_test_dist(file_name => $file_name, ver_num => $ver_num);
is(file($retval)->basename(), "$file_name-$ver_num.fpkg",
t/Test-Fetchware.t view on Meta::CPAN
subtest 'test verbose_on()' => sub {
# turn on verbose.
verbose_on();
# Test if $fetchware::verbose has been set to true.
ok($fetchware::verbose,
'checked verbose_on() success.');
};
subtest 'test add_prefix_if_nonroot() success' => sub {
# Skip all of add_prefix_if_nonroot()'s tests if run as nonroot, because
# this subtest only tests for correct output when run as nonroot. When run
# as root add_prefix_if_nonroot() returns undef, which the test does not
# account for.
plan(skip_all => q{Only test add_prefix_if_nonroot() if we're nonroot})
if $> == 0;
# Clear out any other use of config().
__clear_CONFIG();
my $prefix = add_prefix_if_nonroot();
ok(-e (config('prefix')),
'checked add_prefix_if_nonroot() tempfile creation.');
ok(-e $prefix,
'checked add_prefix_if_nonroot() prefix existence.');
# Clear prefix between test runs.
__clear_CONFIG();
$prefix = add_prefix_if_nonroot(sub {
$prefix = tempdir("fetchware-test-$$-XXXXXXXXXX",
TMPDIR => 1, CLEANUP => 1);
config(prefix => $prefix);
return $prefix;
}
);
ok(-e (config('prefix')),
'checked add_prefix_if_nonroot() tempfile creation.');
ok(-e $prefix,
'checked add_prefix_if_nonroot() prefix existence.');
};
subtest 'test fork_ok()' => sub {
fork_ok(sub {ok(1, 'successful fork_ok() test.')},
'checked fork_ok() success.');
# Abuse a TODO block to test fork_ok() failure by turning that failure into
# success. When this test fails it succeeds, because it is testing failure.
TODO: {
todo_skip 'Turn failure into success.', 1;
fork_ok(sub { return 0 },
'checked fork_ok() failure.');
}
};
subtest 'test fork_not_ok()' => sub {
fork_not_ok(sub {ok(0, 'successful fork_not_ok() test.')},
'checked fork_not_ok() success.');
# Abuse a TODO block to test fork_not_ok() failure by turning that failure into
# success. When this test fails it succeeds, because it is testing failure.
TODO: {
todo_skip 'Turn failure into success.', 1;
fork_not_ok(sub { return 1 },
'checked fork_not_ok() failure.');
}
};
subtest 'test rmdashr_ok()' => sub {
# rmdashr_ok() calls Test::More functions for me, so I can skip them here.
# Perhaps Test::Module testing stuff should be used for this instead?
my ($fh, $filename) = tempfile('fetchware-test-XXXXXXXXX', TMPDIR => 1);
close $filename; # Don't actually need $filname open.
ok(-e $filename, 'checked rmdashr_ok() test file existence.');
rmdashr_ok($filename, 'checked rmdashr_ok() test file unlink.');
ok((not -e $filename), 'checked rmdashr_ok() test file unlinked successfully.');
my $tempdir = tempdir('fetchware-test-XXXXXXXXXXXX', TMPDIR => 1);
ok(-e $tempdir, 'checked rmdashr_ok() test directory existence.');
rmdashr_ok($tempdir, 'checked rmdashr_ok() test directory delete.');
ok((not -e $tempdir), 'checked rmdashr_ok() test directory deleted successfully.');
# Test rmdashr on some "recursive" directories.
$tempdir = tempdir('fetchware-test-XXXXXXXXXXXX', DIR => tmpdir());
my @test_dirs = make_path(catdir($tempdir, qw(1 2 3 4 5 6 7 8 9 0)));
my @extra_test_dirs;
push @extra_test_dirs, make_path(catdir($_, qw(a b c d e )))
for @test_dirs;
my @extra_test_files;
for my $dir (@test_dirs) {
my $testfile = catdir($dir, 'testfile');
push @extra_test_files, $testfile;
open my $fh, '>', $testfile
or fail("Failed to create testdir [$testfile]: $!");
print $fh "Something instead of nothing\n";
close $fh;
}
rmdashr_ok($tempdir, 'checked rmdashr_ok() recursive delete success.');
# A simple noe -e $_bizarrely fails, so just try opening it with open
# failing being the actual "success" we're looking for.
ok((not -e $_), "Checked rmdashr_ok() recursive delete [$_]")
for @test_dirs, @extra_test_dirs, @extra_test_files;
#TODOAdd a test to cmd_look() using this!!!!!
# Abuse a TODO block to test rmdashr_ok() failure by turning that failure into
# success. When this test fails it succeeds, because it is testing failure.
TODO: {
todo_skip 'Turn failure into success.', 1;
rmdashr_ok('Nonexistantfile-' . int(rand(238393890293)));
}
};
# Remove this or comment it out, and specify the number of tests, because doing
# so is more robust than using this, but this is better than no_plan.
#done_testing();
( run in 0.851 second using v1.01-cache-2.11-cpan-39bf76dae61 )