$ wget http://scripts.sil.org/OFL
--2018-02-13 13:42:50--  http://scripts.sil.org/OFL
Resolving scripts.sil.org (scripts.sil.org)... 209.12.63.143
Connecting to scripts.sil.org (scripts.sil.org)|209.12.63.143|:80...
connected.
HTTP request sent, awaiting response... 302 Found
Location: http://scripts.sil.org/cms/scripts/page.php?site_id=nrsi&id=OFL
[following]
--2018-02-13 13:42:52--
http://scripts.sil.org/cms/scripts/page.php?site_id=nrsi&id=OFL
Reusing existing connection to scripts.sil.org:80.
HTTP request sent, awaiting response... 302 Moved Temporarily
Location: /cms/scripts/page.php?site_id=nrsi&id=OFL&_sc=1 [following]
--2018-02-13 13:42:52--
http://scripts.sil.org/cms/scripts/page.php?site_id=nrsi&id=OFL&_sc=1
Reusing existing connection to scripts.sil.org:80.
HTTP request sent, awaiting response... 302 Moved Temporarily
Location: /cms/scripts/page.php?site_id=nrsi&id=OFL [following]
--2018-02-13 13:42:53--
http://scripts.sil.org/cms/scripts/page.php?site_id=nrsi&id=OFL
Reusing existing connection to scripts.sil.org:80.
HTTP request sent, awaiting response... 200 OK
Length: unspecified [text/html]
Saving to: ‘OFL’

    [
<=>
] 37,439      59.6KB/s   in 0.6s

2018-02-13 13:42:55 (59.6 KB/s) - ‘OFL’ saved [37439]

so it may not be following the 302s. I'm not sure you're using the correct
tool here.  A little more straight forward

andy@wiwmb-md-afb-mint:~/spam$ wget http://scripts.sil.org/robots.txt
--2018-02-13 13:47:27--  http://scripts.sil.org/robots.txt
Resolving scripts.sil.org (scripts.sil.org)... 209.12.63.143
Connecting to scripts.sil.org (scripts.sil.org)|209.12.63.143|:80...
connected.
HTTP request sent, awaiting response... 200 OK
Length: 36 [text/plain]
Saving to: ‘robots.txt’

100%[========================================================================================>]
36          --.-K/s   in 0s

2018-02-13 13:47:27 (2.99 MB/s) - ‘robots.txt’ saved [36/36]

but
$ is_it_up.pl
http://scripts.sil.org/robots.txt is DOWN!!!!

You might look at more LWP tools:
$ lwp-dump https://www.sil.org
HTTP/1.1 403 Forbidden
Cache-Control: max-age=10
Connection: keep-alive
Date: Tue, 13 Feb 2018 19:49:47 GMT
Server: cloudflare
Content-Type: text/html; charset=UTF-8
Expires: Tue, 13 Feb 2018 19:49:57 GMT
CF-RAY: 3eca501a5d569895-LAX
Expect-CT: max-age=604800, report-uri="
https://report-uri.cloudflare.com/cdn-cgi/beacon/expect-ct";
Set-Cookie: __cfduid=dd8038f4f2c995fa4b4c7fa8beb2b42f31518551387;
expires=Wed, 13-Feb-19 19:49:47 GMT; path=/; domain=.sil.org; HttpOnly
X-Frame-Options: SAMEORIGIN

<!DOCTYPE html>
<!--[if lt IE 7]> <html class="no-js ie6 oldie" lang="en-US"> <![endif]-->
<!--[if IE 7]>    <html class="no-js ie7 oldie" lang="en-US"> <![endif]-->
<!--[if IE 8]>    <html class="no-js ie8 oldie" lang="en-US"> <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en-US"> <!--<![endif]-->
<head>
<title>Access denied | www.sil.org used Cloudflare to restrict
access</title>
<meta charset="UTF-8" />
<meta http-equiv="Content-Type" content="text/html; charset=UTF-8" />
<meta http-equiv=...
(+ 2770 more bytes not shown)

so it's up, but "forbidden" probably as the user agent isn't set or some
other policy reason.


On Tue, Feb 13, 2018 at 11:33 AM, Manfred Lotz <ml_n...@posteo.de> wrote:

> On Tue, 13 Feb 2018 10:47:42 -0600
> Andy Bach <afb...@gmail.com> wrote:
>
> > The site doesn't like 'head' requests? get works
> > #!/usr/bin/perl
> >
> > use strict;
> > use warnings;
> >
> > use LWP::Simple;
> > #  my $url="https://shlomif.github.io/";;
> > my $url="http://www.notabug.org/";;
> > print "$url is ", (
> >                 (! get($url)) ?  "DOWN!!!!"
> >                                 : "up"
> >                 ), "\n";
> >
> > $ is_it_up.pl
> > http://www.notabug.org/ is up
> >
>
> You are right.
>
> But am afraid this is not all of it. If I test
> http://scripts.sil.org/OFL then I get an error but it is fine in
> firefox.
>
> Very strange.
>
> --
> Manfred
>
>
>
> >
> > On Tue, Feb 13, 2018 at 5:25 AM, Manfred Lotz <ml_n...@posteo.de>
> > wrote:
> >
> > > Hi there,
> > > Somewhere I found an example how to check if a website is up.
> > >
> > > Here my sample:
> > >
> > > #! /usr/bin/perl
> > >
> > > use strict;
> > >
> > > use LWP::Simple;
> > > my $url="https://notabug.org";;
> > > if (! head($url)) {
> > >     die "$url is DOWN!!!!"
> > > }
> > >
> > > Running above code I get
> > >   https://notabug.org is DOWN!!!! at ./check_url.pl line 8.
> > >
> > >
> > > However, firefox shows the site works ok.
> > >
> > >
> > > What am I doing wrong?
> > >
> > >
> > > --
> > > Thanks,
> > > Manfred
> > >
> > > --
> > > To unsubscribe, e-mail: beginners-unsubscr...@perl.org
> > > For additional commands, e-mail: beginners-h...@perl.org
> > > http://learn.perl.org/
> > >
> > >
> > >
> >
> >
>
> --
> To unsubscribe, e-mail: beginners-unsubscr...@perl.org
> For additional commands, e-mail: beginners-h...@perl.org
> http://learn.perl.org/
>
>
>


-- 

a

Andy Bach,
afb...@gmail.com
608 658-1890 cell
608 261-5738 wk

Reply via email to