|Archiving status||Not saved yet|
|IRC channel||(on hackint)|
Twitter is a microblogging service. With each "entry" being 140 characters or less, the ease with which you can track the tiniest details of your life is amazing. The site has become very popular as a result.
The site is becoming so popular, in fact, that many people are deserting or cutting back on their weblogs to just use the Twitter service for what their weblogging used to fulfill; and with that comes rampant centralization, and with that, greater risk. Back up your tweets!
There are currently a few archives (but only partially):
- Cheng-Caverlee-Lee September 2009 - January 2010 Twitter Scrape: almost 10 millon tweets
- The May 2011 Calufa Twitter Scrape: 90+ million tweets from more than 6 million users
The Twitter search API seemingly returns only the latest 7 days worth of tweets.
- Twitter enables you to request an archive of all of your tweets from the main settings page, which includes every tweet of yours (therefore bypassing the normal 3200 tweet API limit). This is then emailed to the account linked with the account.
- Tweetscan Data downloads your Twitter archive from 12/2007 onward in CSV format (requires Twitter account login/password)
- t by sferik is a command-line interface for Twitter using the API via an application you create on your account. Not only does it allow easy CSV/JSON export of your own data, but it allows you to scrape others tweets. API limits apply but this tool is very powerful
Twitter automatically resizes uploaded images. To get image in its original resolution, append :orig after the url, e.g.:
When using ArchiveBot, the following arguments are helpful:
--phantomjs --ignore-sets twitter
It is also important to add a trailing slash to the URL, so it gets each tweet individually, rather than only trying to download the whole timeline.
A simple way to archive many Tweets from accounts and searches is ChromeBot.
A common way to archive entire Twitter accounts is to put the tweets into an URL list using snscrape, uploading them to https://transfer.sh/ or https://transfer.notkiska.pw/ and feeding them into ArchiveBot using the
!ao < file command.
- The Python Twitter API by sixohsix has some pretty easy to use scripts for archiving Twitter accounts to a TXT file for people who aren't as technically inclined. It can only save the last 3K or so tweets due to inbuilt Twitter limits, though. (Note: the "-o" flag is pretty much required to archive accounts.)
- Here's a Gist with a step-by-step guide to getting a long list of a user's tweet status URL's, using a Python program called Tweep.
See Site exploration for details.
Very stable, probably not going anywhere too soon without warning.
- Twitter watchlist – Accounts with volatile content.
- List of lost Twitter accounts – Twitter accounts that went private, disabled, suspended or deleted a significant amount of content.
Library of Congress
The U.S. Library of Congress announced in April 2010, via its official Twitter account that it will be acquiring the entire archive of Twitter messages back through March 2006. As of 2016-02-23, this archive is still not available, and when/if it does become accessible it will likely be restricted to researchers, rather than the general public. In January 2017, it was announced that the Library of Congress will no longer archive all tweets, just ones from major news stories.