Site Capture & a Misc Note on RegSeeker and EzClearner

  • Thread starter Thread starter Bob S.
  • Start date Start date
B

Bob S.

Looking for a program to capture info (graphics and text) from a web site
that spans across multiple pages. For instance, a "How To" article will
include drawings plus text and then you need to click on Next to see the
next section.

I've tried v6.22 SnagIt but all I capture are some of the graphics and not
the entire article.

Thank you,

Bob S.

Misc Note:

Have used Toni Arts, EzCleaner (v1.7f) in the past and it worked great -
never had it create a problem and I always thought it did a thorough job of
cleaning under Win98. Moved up to WinXP Pro, installed the new version 2.0
of EzCleaner and it worked fine and has some excellent features. Then I saw
a post here about RegSeeker and gave it a try...

First did a pass with EzCleaner and it found 27 items.
Followed that with a pass from RegSeeker and it found 477 additional items
to be cleaned...!

Considerable difference. Did the cleaning (saving everything of course) and
ran just about every major program (from AutoCad to WeatherWatcher) to test
that there were no adverse reactions. So far, no major problems but it did
break the association on a .est extension which is used with MS Streets and
Maps. That was easily fixed and everything is running good... Thank you
for putting me on to that utility.
 
Bob S. said:
First did a pass with EzCleaner and it found 27 items.
Followed that with a pass from RegSeeker and it found 477 additional items
to be cleaned...!

These two programs, IMO, are not even in the same class. Amateur v Pro.

Btw, Bob, thanks for posting your experience on running these...
Considerable difference. Did the cleaning (saving everything of course) and
ran just about every major program (from AutoCad to WeatherWatcher) to test
that there were no adverse reactions. So far, no major problems but it did
break the association on a .est extension which is used with MS Streets and
Maps. That was easily fixed and everything is running good... Thank you
for putting me on to that utility.

I just looked, and do not see any kind of option in Regseeker for ignoring
references to the CD drive, which would have prevented it getting confused
about the MS Streets and Maps file references. I'd propose that the author
should consider putting the option as a new feature for next release.
 
omega said:
These two programs, IMO, are not even in the same class. Amateur v
Pro.

Maybe so. But is Bob's PC any better off with those 477 additional registry
items removed?
 
That is a hard question to answer since it would involve many variables
which are beyond my capabilities to explain. I've only worked on computers
for 30+ years but I'm a relative newbie to WinXP and all it has to offer -
good, bad and/or indifferent...

The short answer is - Yes. It did make an apparent difference in overall
performance but I don't have any hard numbers (bench tests) that I can use
to prove it. It didn't appear to do any harm and if it cleaned out a little
of the garbage and made the load lighter, then I guess it was well worth the
price. That aside, RegSeeker also has some other nice features that may
interest you such as Startup entries, some tweaks, clean out history's,
manage Favorites, and some others.

Bob S.
 
Bob S. said:
Looking for a program to capture info (graphics and text) from a web site
that spans across multiple pages. For instance, a "How To" article will
include drawings plus text and then you need to click on Next to see the
next section.

I've tried v6.22 SnagIt but all I capture are some of the graphics and not
the entire article.

SnagIt is a payware, so many of us won't be familiar with what it does. I
did go ahead and put its name into Google. The excerpts indicates that it
is a screen capture utility. From your description of the function you
need - following links - you might be best served by a robot, a spidering
program.

Httrack is the most sophisticated. And I believe that the default settings
with its filters work the way you'd want (given normal site structure), ie
like this...

given
www.domain.com\030524\subject\article222\

yes (lower down)
www.domain.com\030524\subject\article222\page1.htm
www.domain.com\030524\subject\article222\page2.htm

no (higher up)
www.domain.com\030524\subject3\article134\page1.htm

yes (default filter= +.jpg +.gif +.png +.bmp)
www.domain.com\images\clear3.gif
www.domain.com\images\screenshot5.png
(images linked on the article pages)


Your hard part will be if the page you want does its links by javascript.
Httrack is faraway the most advanced of the freeware spiderbots, so it's
the one that has the best shot at success. Same time, some js can be just
too convoluted and tricky to parse.


http://www.pricelessware.org/PL2003INTERNET.htm#DownloadWebSites
http://www.webattack.com/Freeware/downloader/fwoffline.html
 
omega,

Thank you for taking the time to not only research my question but to also
explain what I need to do. That was very kind of you and greatly
appreciated.

Bob S.
 
MLC said:
Il 10/dic/2003 Bob S. ha scritto:

It's been there since 11 October :)

Yes, but today is the first day it will run on Bob's machine. I am hoping
HTTrack will put on one of its finer performances for him. :)
 
Il 10/dic/2003 omega ha scritto:
Yes, but today is the first day it will run on Bob's machine. I am hoping
HTTrack will put on one of its finer performances for him. :)

I wish that to him, me too.
However, the subject in his post made me think there was an update of the
program. I thought it was worth clarifying it.
:)
 
And run it did !

Karen, it worked perfectly and accomplished exactly what I needed it to do -
the first time out. I will follow-up with a note of thanks to the author
and although it's freeware, I'm sure he would accept a donation to help keep
his site on the air.

Your help was key and sincerely appreciated.

Thanks,

Bob S.
 
Bob S. said:
And run it did !

Karen, it worked perfectly and accomplished exactly what I needed it to do -
the first time out. I will follow-up with a note of thanks to the author
and although it's freeware, I'm sure he would accept a donation to help keep
his site on the air.

Neat, I'd think he would like that..
Your help was key and sincerely appreciated.

Thank you for being so gracious. I'm happy to recommend a good program.
I am also very glad to hear that its default settings worked right out of
the box. The truth is, that's been an evolution. In the early days of
HTTrack, it could be tricky to work it out. Even at present, it takes a
sincere computer user, in approaching it for the first time, and getting
the right sense of it. Congrats on the success.
 
Looking for a program to capture info (graphics and text) from a web site
that spans across multiple pages. For instance, a "How To" article will
include drawings plus text and then you need to click on Next to see the
next section.

I've tried v6.22 SnagIt but all I capture are some of the graphics and not
the entire article.

It depends. Is the web site static (doesn't change) or dynamic (i.e.
pages are served based on cgi scripts)? Wget works pretty well (DOS
and unix) and so does WinHttrack (windows).

I don't know how these web site grabbers work with dynamic sites but
they work pretty well for static sites. I have used both of these and
wget has a better facility for getting only changed/new pages (uses
less bandwidth) and for fooling the server that it's not a website
grabber. (They don't like you sucking up that much bandwidth.)

--
Freezone Freeware: 1200+ applications
http://chuckr.freezone.org (zip file of file list only)
http://freezone.darksoft.co.nz
http://home.att.net/~chuckr30/index.html (zip file of file list only)
http://chuckr.bravepages.com
 
Back
Top