Angelfire

From Archiveteam
Revision as of 11:52, 17 January 2017 by Megalanya1 (talk | contribs) (MOTHERFUCKER ! ! !)
Jump to navigation Jump to search
The printable version is no longer supported and may have rendering errors. Please update your browser bookmarks and please use the default browser print function instead.
Angelfire
Angelfire- Welcome to Angelfire 1303510943179.png
URL http://www.angelfire.lycos.com/
Status Online!
Archiving status Not saved yet
Archiving type Unknown
IRC channel #archiveteam-bs (on hackint)

Angelfire is a web hosting service since 1996, containing big chunks of early WWW history (which people love to mock at).

It is not expected that the Angelfire archive can ever be truly complete, as Angelfire, like other free hosts such as Homestead, has or had a policy of deleting "inactive" accounts. As there is no known mirror of many of these former accounts and associated web pages, there may be no way to recover such deleted websites.

Angelfire underwent some changes in 2010, apparently not disruptive but requiring users to pay for some options like the old Web Shell tool; we do not know whether this caused some older websites to become unaccessible for their owners and whether that could cause inactivity and hence deletion. The Alexa rank of the property seems in constant fall, from better than 2000th position in early 2012 to worse than 3400th in early 2014.

It's not clear how bad Lycos is. A quick seach for Lycos shutdowns only points to their (independently operated) Lycos Europe liquidation, which gave less than a month for the users to save their emails before deletion. Lycos Tripod on the other hand, which was in 2003 Europe's largest homepage building community (with special Google alliance), found a last minute buyer for its European wing but then suddenly went down in July 2013 (it was around 60,000th Alexa position in 2012 and fell well below 100,000 in early 2013).

MOTHERFUCKER ! ! !

MOTHERFUCKER ! ! !

MOTHERFUCKER ! ! !

Discovery & Downloading

First grab all the sitemap indexes:

curl http://www.angelfire.com/robots.txt | grep -Eo 'http.*gz' > sitemap-index-urls

http://www.angelfire.com/sitemap-index-00.xml.gz
http://www.angelfire.com/sitemap-index-01.xml.gz
http://www.angelfire.com/sitemap-index-02.xml.gz
...
http://www.angelfire.com/sitemap-index-ff.xml.gz


Use that to grab all the sitemaps:

wget -i sitemap-index-urls

Inside you will see the users' sitemaps URLs

<sitemap><loc>http://www.angelfire.com/punk4/jori_loves_jackass/sitemap.xml</loc><lastmod>2012-04-10</lastmod></sitemap>
<sitemap><loc>http://www.angelfire.com/vevayaqo/sitemap.xml</loc><lastmod>2012-04-10</lastmod></sitemap>
<sitemap><loc>http://www.angelfire.com/planet/dumbass123/sitemap.xml</loc><lastmod>2012-04-10</lastmod></sitemap>
...


Extract the user sitemap URLs:

zgrep -hEo 'http:.*xml' sitemap-index-*.xml.gz > sitemap-urls

http://www.angelfire.com/punk4/jori_loves_jackass/sitemap.xml
http://www.angelfire.com/vevayaqo/sitemap.xml
http://www.angelfire.com/planet/dumbass123/sitemap.xml
...

Extract the webpage URLs:

grep -Eo '<loc>.*</loc>' www.angelfire.com/"${user}"/sitemap.xml | sed 's#<loc>##' | sed 's#</loc>##' > "${user}.urls"

http://www.angelfire.com/ab7/pledgecry/band.html
http://www.angelfire.com/ab7/pledgecry/biography.html
http://www.angelfire.com/ab7/pledgecry/ernst.html
http://www.angelfire.com/ab7/pledgecry/header.html
...

Grab them with options like: -m --no-parent --no-cookies -e robots=off --page-requisites --domains=angelfire.com,lycos.com


As of 2015-05-08 there are 3895290 users

You will want --no-cookies because angelfire wants to set them everywhere.

Reject http://www.angelfire.lycos.com/doc/images/track/ot_noscript.gif.* and reject http://www.angelfire.com/adm/ad/ (ads) --> --reject-regex='(www.angelfire.com\/adm\/ad\/|www.angelfire.com\/doc\/images\/track\/ot_noscript\.gif)'

Some images are hosted on http://www.angelfire.lycos.com --> --domains=angelfire.com,lycos.com


Guestbooks have been killed in 2012, eg http://htmlgear.lycos.com/guest/control.guest?u=gosanson&i=2&a=view

Some users have blogs with infinite calendars, like this in the sitemap: http://filesha.angelfire.com/blog/index.blog . Wget will run infinitely on those, better skip them for now.

Many users have no URLs in their sitemaps. Not sure what to do with those.

External links