Wget or Curl are quite popular, but there are various front ends for
them as well as standalone tools like Java Websuck (delightful name),
Ebot, Larbin and Heretrix.
Zotero is also useful, but not easily scriptable (although arguably one
could pull some Python or Ruby based scripting that dumps into a similar
compact SQLite style DB schema.
On 10/5/2010 3:48 PM, Will Lowe wrote:
> For selective mirroring of arbitrary websites, try wget - http://www.gnu.org/software/wget/ Googling 'graphical wget' retrieves any number of front-ends for it on Mac and Windows.
>
> Will
>
>
>
> On 5 Oct 2010, at 22:46, Paul Gronke wrote:
>
>>> Colleagues
>>>
>>> I have a student who is writing a thesis on the rise of the tea party movement. He is proposing a comparative case study of a number of Senate races. I have suggested that beyond examining candidate statements, candidate ads, and newspaper stories, that he attempt to capture the content placed on the candidate websites on a regular basis up to the election.
>>>
>>> My question is this: does anyone know of an easy to use program whereby you can provide it with a list of websites, some sort of script, and the program goes out and creates some sort of time stamped copy of a website?
>>>
>>> If emails come to me personally, I'll try to summarize for the list.
>>>
>>> Thanks in advance
>>> Paul G.
>>> ---
>>> Paul Gronke Ph: 503-517-7393
>>> Fax: 734-661-0801
>>>
>>> Professor, Reed College
>>> Director, Early Voting Information Center 3203 SE Woodstock Blvd.
>>> Portland OR 97202
>>>
>>> EVIC: http://earlyvoting.ne
**********************************************************
Political Methodology E-Mail List
Editors: Diana O'Brien <[log in to unmask]>
Jon C. Rogowski <[log in to unmask]>
**********************************************************
Send messages to [log in to unmask]
To join the list, cancel your subscription, or modify
your subscription settings visit:
http://polmeth.wustl.edu/polmeth.php
**********************************************************
|