Angelfire

From Archiveteam
Revision as of 21:22, 23 August 2018 by Adinbied (talk | contribs) (Add URL discovery script and output, and change status to upcoming)
Jump to navigation Jump to search
Angelfire
Angelfire- Welcome to Angelfire 1303510943179.png
URL http://www.angelfire.lycos.com/
Status Online!
Archiving status Upcoming...
Archiving type Unknown
IRC channel #archiveteam-bs (on hackint)

Angelfire is a web hosting service since 1996, containing big chunks of early WWW history (which people love to mock at).

It is not expected that the Angelfire archive can ever be truly complete, as Angelfire, like other free hosts such as Homestead, has or had a policy of deleting "inactive" accounts. As there is no known mirror of many of these former accounts and associated web pages, there may be no way to recover such deleted websites.

Angelfire underwent some changes in 2010, apparently not disruptive but requiring users to pay for some options like the old Web Shell tool; we do not know whether this caused some older websites to become unaccessible for their owners and whether that could cause inactivity and hence deletion. The Alexa rank of the property seems in constant fall, from better than 2000th position in early 2012 to worse than 3400th in early 2014.

It's not clear how bad Lycos is. A quick search for Lycos shutdowns only points to their (independently operated) Lycos Europe liquidation, which gave less than a month for the users to save their emails before deletion. Lycos Tripod on the other hand, which was in 2003 Europe's largest homepage building community (with special Google alliance), found a last minute buyer for its European wing but then suddenly went down in July 2013 (it was around 60,000th Alexa position in 2012 and fell well below 100,000 in early 2013).

Status

Archivebot gave it a try, http://archive.fart.website/archivebot/viewer/job/9yhap

Schbirid has some ugly Bash scripts: https://github.com/SpiritQuaddicted/angelfire (ask before you use, they are probably out of date)

A full discovery of all of the URLs is currently underway (using a modified version of Schbirid's script which can be found here: https://github.com/adinbied/angelfire-items). A partial output for the above script has been uploaded here: https://archive.org/details/angelfireURLS_0x00-0x14, where each user has their own .urls file in the directory that their site was found in the sitemap indexes. In each sitemap folder, there is a blogusers.todo file with the users that have infinite calendars (which wget stalls out on as explained below) - that still has to be figured out.

Discovery & Downloading

First grab all the sitemap indexes:

curl http://www.angelfire.com/robots.txt | grep -Eo 'http.*gz' > sitemap-index-urls

http://www.angelfire.com/sitemap-index-00.xml.gz
http://www.angelfire.com/sitemap-index-01.xml.gz
http://www.angelfire.com/sitemap-index-02.xml.gz
...
http://www.angelfire.com/sitemap-index-ff.xml.gz


Use that to grab all the sitemaps:

wget -i sitemap-index-urls

Inside you will see the users' sitemaps URLs

<sitemap><loc>http://www.angelfire.com/punk4/jori_loves_jackass/sitemap.xml</loc><lastmod>2012-04-10</lastmod></sitemap>
<sitemap><loc>http://www.angelfire.com/vevayaqo/sitemap.xml</loc><lastmod>2012-04-10</lastmod></sitemap>
<sitemap><loc>http://www.angelfire.com/planet/dumbass123/sitemap.xml</loc><lastmod>2012-04-10</lastmod></sitemap>
...


Extract the user sitemap URLs:

zgrep -hEo 'http:.*xml' sitemap-index-*.xml.gz > sitemap-urls

http://www.angelfire.com/punk4/jori_loves_jackass/sitemap.xml
http://www.angelfire.com/vevayaqo/sitemap.xml
http://www.angelfire.com/planet/dumbass123/sitemap.xml
...

Extract the webpage URLs:

grep -Eo '<loc>.*</loc>' www.angelfire.com/"${user}"/sitemap.xml | sed 's#<loc>##' | sed 's#</loc>##' > "${user}.urls"

http://www.angelfire.com/ab7/pledgecry/band.html
http://www.angelfire.com/ab7/pledgecry/biography.html
http://www.angelfire.com/ab7/pledgecry/ernst.html
http://www.angelfire.com/ab7/pledgecry/header.html
...

Grab them with options like: -m --no-parent --no-cookies -e robots=off --page-requisites --domains=angelfire.com,lycos.com


As of 2015-05-08 there are 3895290 users

You will want --no-cookies because angelfire wants to set them everywhere.

Reject http://www.angelfire.lycos.com/doc/images/track/ot_noscript.gif.* and reject http://www.angelfire.com/adm/ad/ (ads) --> --reject-regex='(www.angelfire.com\/adm\/ad\/|www.angelfire.com\/doc\/images\/track\/ot_noscript\.gif)'

Some images are hosted on http://www.angelfire.lycos.com --> --domains=angelfire.com,lycos.com


Guestbooks have been killed in 2012, e.g. http://htmlgear.lycos.com/guest/control.guest?u=gosanson&i=2&a=view

Some users have blogs with infinite calendars, like this in the sitemap: http://filesha.angelfire.com/blog/index.blog . Wget will run infinitely on those, better skip them for now.

Many users have no URLs in their sitemaps. Not sure what to do with those.

External links