DXR is a code search and navigation tool aimed at making sense of large projects. It supports full-text and regex searches as well as structural queries.

Mercurial (27a812186ff4)

VCS Links

Line Code
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117
#!/usr/bin/perl
#
# This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.

# This is a modified version of Chris Hofmann's <chofmann@netscape.com>
# infamous "browser buster" test harness. It's a bit simpler (CGI
# instead of using cookies; IFRAME instead of FRAMESET), and has some
# extra parameters that make it a bit easier to test with, but it's
# pretty faithful otherwise.
#
# It accepts a couple of parameters, including
#
#   file=<filename> Set this to the name of the file containing
#     the URLs that you want the buster to cycle through. This
#     might be a security hole, so don't run this script on a
#     server with s3kret stuff on it, mmkay?
#
#   page=<number> This is used to maintain state, and is the line
#     number in the file that the buster will pull up in the
#     IFRAME. Set if by hand if you need to for some reason.
#
#   last=<number> The buster will run until it's exhausted all
#     the URLs in the file, or until it reaches this line in the
#     file; e.g., setting it to "5" will load five URLs.
#
#   refresh=<number> The timeout (in seconds) to wait before doing
#     a page refresh, and thus loading the next URL. Defaults to
#     thirty.

use CGI;

# Find the page'th URL in the file with the specified name
sub FindURL($$)
{
    my ($file, $page) = @_;

    open URLS, $file
        || die("can't open $::File");

    LINE: while (<URLS>) {
        next LINE if /^#/;
        last LINE unless --$page;
    }

    close URLS;

    chomp;
    return $_;
}

# Scrape parameters
$::Query = new CGI;

$::File = $::Query->param("file");
$::File = "top100.txt" unless $::File;

$::Page = $::Query->param("page");
$::Page = 0 unless $::Page;
$::URL = FindURL($::File, ++$::Page);

$::Last = $::Query->param("last");
$::Last = -1 unless $::Last;

$::Refresh = $::Query->param("refresh");
$::Refresh = 30 unless $::Refresh;

# Header
print qq{Content-type: text/html

<html>
<head>
};

# Meat
if ($::URL && ($::Page <= $::Last || $::Last == -1)) {
    # Make a web page that'll load $::URL in an IFRAME, with
    # a meta-refresh that'll reload us again in short order.
    print qq{<meta http-equiv="Pragma" content="no-cache">
<meta http-equiv="refresh" content="$::Refresh;url=buster.cgi?file=$::File&page=$::Page&last=$::Last&refresh=$::Refresh">
<title>BrowserBuster II: $::URL</title>
<style type="text/css">
body {
  overflow: hidden;
  border: 0;
  margin: 0;
}
</style>
</head>
<script>
dump("+++ loading $::URL\\n");
</script>
<body>
};
    print "$::File: $::URL";
    if ($::Last != -1) {
        print " ($::Page of $::Last)<br>";
    }
    print qq{
<iframe width="100%" height="100%" src="$::URL">
};
}
else {
    # Make a web page that'll close the current browser
    # window, terminating the test app.
    print qq{<head>
<title>BrowserBuster II: Done!</title>
<body onload="window.close();">
All done!
};
}

# Footer
print qq{</body>
</html>
};