URL Minder, URL Watcher

URL Minder, URL Watcher

Post by J » Sat, 22 Dec 2001 00:23:32



Does anyone know of a script I can use that will email a notification
whenever the contents change on any given URL?  I know there are some
"URL Minder" services freely available, but we're behind a corporate
firewall, so they won't work.  I'd like to run a local version.
Ideally, this should recurse any filesystems/directories without
having to explicitly set it up for individual URLs.

Thanks.
JR

 
 
 

URL Minder, URL Watcher

Post by Jeremiah DeWitt Weine » Sat, 22 Dec 2001 01:21:17



> Does anyone know of a script I can use that will email a notification
> whenever the contents change on any given URL?  I know there are some
> "URL Minder" services freely available, but we're behind a corporate
> firewall, so they won't work.  I'd like to run a local version.
> Ideally, this should recurse any filesystems/directories without
> having to explicitly set it up for individual URLs.

        I don't know of any, but it should be pretty easy to gin one up.
Use wget or 'lynx -dump' or similar to grab the page(s) and save them.
Use wget or whatever to grab the pages again and save them to a different
location.
Use diff to compare the pages, and if they're different, send email.

        It's a little trickier if you've got multiple pages, so you have to
recurse through them, but not much.

--
If mail to me bounces, try removing the "+STRING" part of the address.

 
 
 

URL Minder, URL Watcher

Post by Chris F.A. Johnso » Sat, 22 Dec 2001 07:04:49




> > Does anyone know of a script I can use that will email a notification
> > whenever the contents change on any given URL?  I know there are some
> > "URL Minder" services freely available, but we're behind a corporate
> > firewall, so they won't work.  I'd like to run a local version.
> > Ideally, this should recurse any filesystems/directories without
> > having to explicitly set it up for individual URLs.

>    I don't know of any, but it should be pretty easy to gin one up.
> Use wget or 'lynx -dump' or similar to grab the page(s) and save them.
> Use wget or whatever to grab the pages again and save them to a different
> location.
> Use diff to compare the pages, and if they're different, send email.

>    It's a little trickier if you've got multiple pages, so you have to
> recurse through them, but not much.

If you don't want to keep local copies of the files, download each
site with 'wget -r', and store the date of last modification of each
file in a table, then erase the download.

To compare, repeat the process and diff the two tables.
Replace the old table with the new one.

--
    Chris F.A. Johnson                        http://cfaj.freeshell.org
    ===================================================================
    My code (if any) in this post is copyright 2001, Chris F.A. Johnson
    and may be copied under the terms of the GNU General Public License

 
 
 

URL Minder, URL Watcher

Post by Billy N » Sun, 23 Dec 2001 02:48:29


Will this way only work on static pages?

Billy





> > > Does anyone know of a script I can use that will email a notification
> > > whenever the contents change on any given URL?  I know there are some
> > > "URL Minder" services freely available, but we're behind a corporate
> > > firewall, so they won't work.  I'd like to run a local version.
> > > Ideally, this should recurse any filesystems/directories without
> > > having to explicitly set it up for individual URLs.

> > I don't know of any, but it should be pretty easy to gin one up.
> > Use wget or 'lynx -dump' or similar to grab the page(s) and save them.
> > Use wget or whatever to grab the pages again and save them to a
different
> > location.
> > Use diff to compare the pages, and if they're different, send email.

> > It's a little trickier if you've got multiple pages, so you have to
> > recurse through them, but not much.

> If you don't want to keep local copies of the files, download each
> site with 'wget -r', and store the date of last modification of each
> file in a table, then erase the download.

> To compare, repeat the process and diff the two tables.
> Replace the old table with the new one.

> --
>     Chris F.A. Johnson                        http://cfaj.freeshell.org
>     ===================================================================
>     My code (if any) in this post is copyright 2001, Chris F.A. Johnson
>     and may be copied under the terms of the GNU General Public License

 
 
 

URL Minder, URL Watcher

Post by Chris F.A. Johnso » Sun, 23 Dec 2001 03:37:21







> > > > Does anyone know of a script I can use that will email a notification
> > > > whenever the contents change on any given URL?  I know there are some
> > > > "URL Minder" services freely available, but we're behind a corporate
> > > > firewall, so they won't work.  I'd like to run a local version.
> > > > Ideally, this should recurse any filesystems/directories without
> > > > having to explicitly set it up for individual URLs.

> > > I don't know of any, but it should be pretty easy to gin one up.
> > > Use wget or 'lynx -dump' or similar to grab the page(s) and save them.
> > > Use wget or whatever to grab the pages again and save them to a
> different
> > > location.
> > > Use diff to compare the pages, and if they're different, send email.

> > > It's a little trickier if you've got multiple pages, so you have to
> > > recurse through them, but not much.

> > If you don't want to keep local copies of the files, download each
> > site with 'wget -r', and store the date of last modification of each
> > file in a table, then erase the download.

> > To compare, repeat the process and diff the two tables.
> > Replace the old table with the new one.

[please don't top post]

Quote:> Will this way only work on static pages?

Presumably.

To include CGIs and other dynamic pages, you could create a table of
checksums instead of dates.

Pipe "lynx -dump URL" to a script that ouputs the URL, the number of
characters, and a checksum, and use that for the table.

--
    Chris F.A. Johnson                        http://cfaj.freeshell.org
    ===================================================================
    My code (if any) in this post is copyright 2001, Chris F.A. Johnson
    and may be copied under the terms of the GNU General Public License

 
 
 

1. Local Script Like "URL-Minder"??



|       Creating the lists and all that is a piece of cake, but does
| anyone know how to "query" a server, in C or Perl, in such a way that
| all that is return is on of the HEAD fields, Content-Length in particular?
| If I can compare Content-Length.last to Content-Length.current, to see if
| there are any changes, the rest is easy.

The easiest way without going into programming it, is to use Lynx FM and
the command:

lynx -head -dump http://www.....etc/path/

 Paul.

---------------------------------------------------------------------------
         1 line sigs are _so_ cool. Unfortunately mine isn't one.

Phone: +44 (0)1849-460860   +44 (0)378-980-570 (M)  Fax: +44 (0)1849-465567
           Eight out of every five people are math illiterates.

2. Printing, again

3. Netscape Commerce Server (Redirecting URLs and URL Aliases)

4. Help...need access to a machine running Solaris

5. Redirecting Virtual Host URL to real URL - HOW?

6. RH Linux and Air-H

7. mod_rewrite URL-->URL

8. g++, Motif and Solaris 2.4

9. Apache mod_rewrite question, url rewrite, url forward

10. avoid warning of redirection to non-secure url from secure url

11. rewrite rule for http URL -> https URL on specified directory

12. Replace Browser Short-URL with Full-Name-URL

13. Need a _text_ utility to fetch an URL and save the URL's source code as text