https://wiki.archiveteam.org/api.php?action=feedcontributions&user=Antiufo&feedformat=atomArchiveteam - User contributions [en]2024-03-28T08:54:21ZUser contributionsMediaWiki 1.37.1https://wiki.archiveteam.org/index.php?title=INTERNETARCHIVE.BAK&diff=45053INTERNETARCHIVE.BAK2020-07-05T11:42:47Z<p>Antiufo: Add link to iabak-sharp implementation</p>
<hr />
<div>[[Image:Ohreally.gif|right|frame|Backing up the Internet Archive]]<br />
The wonder of the [[Internet Archive]]'s petabytes of stored data is that they're a world treasure, providing access to a mass of information and stored culture, gathered from decades of history (in some case, centuries), and available in a (relatively) easy-to-find fashion. And as media and popular websites begin to cover the Archive's mission in earnest, the audience is growing notably. <br />
<br />
In each wave of interest, questions come forward out of the accolades: <br />
<br />
* Why is the Internet Archive only in one (actually, limited) physical space? <br />
* What is the disaster recovery plan? <br />
* What steps are being taken to ensure integrity and access?<br />
<br />
Some of the answers are essays in themselves, but regarding the protection of the Internet Archive's storehouses of data, a project has been launched to add some real-world information to the theoretical consideration of data protection.<br />
<br />
The '''INTERNETARCHIVE.BAK project''' (also known as '''IA.BAK''' or '''IABAK''') is a combined experiment and research project to back up the Internet Archive's data stores, utilizing zero infrastructure of the Archive itself (save for bandwidth used in download) and, along the way, gain real-world knowledge of what issues and considerations are involved with such a project. Started in April 2015, the project already has dozens of contributors and partners, and has resulted in a fairly robust environment backing up terabytes of the Archive in multiple locations around the world. <br />
<br />
''Visit http://iabak.archiveteam.org to see the current status of the project's data storage and to learn how you can contribute disk space and bandwidth.''.<br />
<br />
'''IA.BAK has been broken and unmaintained since about December 2016. The above status page is not accurate as of December 2019.'''<br />
<br />
== THE PHILOSOPHY ==<br />
<br />
The project's main goals include:<br />
<br />
* Maintaining multiple, verified, geographically-disparate copies of Internet Archive data<br />
* No use of Internet Archive infrastructure, that is, machines or drives (bandwidth just for download)<br />
* Keep at least 3 external copies in three different physical locations from the Archive.<br />
* Conduct bi-weekly verification that the data is secure in the external locations.<br />
* Expire or remove data that has not been verified after 30 days, replacing it with functional space.<br />
<br />
The project's secondary goals include:<br />
<br />
* Add ease of use to the end-user clients so the maximum amount of drive space contributors can participate.<br />
* Offer optional peer-to-peer data provision for both the main site's items and to speed synchronization.<br />
* Issue useful conclusions as to how the project is conducted.<br />
* Ultimately provide encrypted versions of some data so that internal/system data can be backed up too.<br />
<br />
The project rests on several assumptions:<br />
<br />
* Given an opportunity and good feedback, there are petabytes of volunteer disk space available<br />
* The Internet Archive will gain more awareness by the general public by this project<br />
* A non-homogeneous environment for the data makes the data that much stronger and disaster-resistant<br />
* The project is an educational workshop as well as a good-faith effort, providing hard answers to theoretical questions.<br />
<br />
== WHAT WE HAVE LEARNED SO FAR (ABOUT THE DATA) ==<br />
<br />
To understand what might be involved in backing up a large set of data, information about that dataset needed to be researched.<br />
<br />
The project, known as the ''[[Internet_Archive_Census|Internet Archive Census]]'' resulted in some solid numbers, as well as a set of facts that helped nail down the scope of the project. They include:<br />
<br />
* The Internet Archive has roughly 21 petabytes of unique data at this juncture. (Growing daily.)<br />
* Some is not available for direct download by the general public, so IA.BAK is not addressing it (yet).<br />
* Some data is, by its nature, more "critical" or "important" than others. Unique acquired data versus, say, universally available datasets or mirrors of prominent and accessible-from-many-locations music videos.<br />
* A lot of the data — the vast, vast majority — is extremely static, and meant to live forever.<br />
* A lot of the files are "derives", and marked as such - files that were derived from other files.<br />
* The Internet Archive reserves the right to delete uploaded content.<br />
<br />
The census classified the total store of public-accessible, unique data as very roughly 14 petabytes. Of that data, there is some level of redundancy and there are collections that might be called "crown jewels" or treasures (the Prelinger Archives, the Bali Culture collection, collections of scanned books from partner libraries) while others are more "nice to have" (universally accessible podcasts or video blogs, multiple revisions of the same website backup).<br />
<br />
== WHAT WE HAVE DONE SO FAR (ABOUT THE IMPLEMENTATION) ==<br />
<br />
[[Image:gitannex.png|right]]<br />
<br />
There is a channel, '''#internetarchive.bak''', on EFNet, for discussion about the project as well as bot-driven updates about code changes and new contributors. It is occasionally very active and mostly quiet in between new issues being handled. It is not required to use as a contributor of disk space, but is an excellent resource for questions.<br />
<br />
After multiple discussions, the back-end for this phase of this project is using the [http://git-annex.branchable.com/ git-annex] suite of tools. Reasons include the maturity of the project and the willing contribution of time by the git-annex code designer and maintainer, Joey Hess (also a co-founder of Archive Team). ''Other possible software suites are listed at the end of this Wiki entry and might play a part in future revisions of the project.''<br />
<br />
To read the git-annex overview and proposal, go to: '''[[INTERNETARCHIVE.BAK/git-annex_implementation]]'''<br />
<br />
To read the git-annex implementation documentation, go to: '''[https://git-annex.branchable.com/design/iabackup/ https://git-annex.branchable.com/design/iabackup/]'''<br />
<br />
== CHOOSING WHICH DATA AT THE ARCHIVE TO BACK UP USING IA.BAK ==<br />
<br />
As space is relatively limited to provide to IA.BAK, the project has had a second task line of choosing which public-facing data at the Internet Archive to push into IA.BAK. Generally, the group is looking for:<br />
<br />
* Data unique enough that it is predominantly at the Internet Archive<br />
* Data that does not flood the capacity of the project, i.e. hundreds of Terabytes<br />
* Data that does not generally limit which countries can host it<br />
* Data that is generally smaller in size, yet still unique<br />
<br />
There is now a page to nominate sets of data at the Internet Archive that should be added to future shards:<br />
<br />
* [[INTERNETARCHIVE.BAK/nominations]]<br />
<br />
The project status page has methods for seeing what collections are now in the project, although it lacks simple search (yet).<br />
<br />
== KEEPING TRACK OF THE DOWNLOAD AND BACKUP PROGRESS ==<br />
[[File:iabakpage.png|300px|right]]<br />
<br />
The IA.BAK project page, http://iabak.archiveteam.org/, provides a graphical overview of the progress of the project. Updated frequently, it shows total amount of space backed up, the status (red to green) of amount of backups of a given set of data, and timelines of how data is being acquired from the original sources. It also includes maps to show the geographical disparity of the clients, as well as how the clients are performing.<br />
<br />
It is a continual challenge, and an ongoing one, to make navigation of these pages intuitive and informative. As the dataset grows in size and complexity, major changes to the layout of the pages will continue.<br />
<br />
As with everything else, the infrastructure to maintain this page is completely separate from Internet Archive machines and resources.<br />
<br />
== CONCERNS ==<br />
<br />
Since the beginning, there have been a number of concerns voiced about the project, and while most can't be glibly dismissed, it is good to at least acknowledge them (as they tend to take the first minutes of a conversation about the project).<br />
<br />
* '''Bad Actors''' will always be a continued concern, as person or persons determined to prevent the backing up of data could execute endless actions to slow down, corrupt, or ruin their contributions of space.<br />
<br />
* '''The Great Restore''' is the time when data from IA.BAK must return to its home to handle a physical or software-based failure of the Internet Archive's data stores. (This could be anything from an earthquake, fibre line cut, or to the dropping of an EMP). A restoring would definitely involve the leasing of a large amount of upload bandwidth and some amount of physical transfer.<br />
<br />
== See Also and Additional Information ==<br />
<br />
Alternate proposed software suites for this project besides git-annex:<br />
<br />
* [[INTERNETARCHIVE.BAK/torrents_implementation]]<br />
* [[INTERNETARCHIVE.BAK/ipfs_implementation]]<br />
* [[INTERNETARCHIVE.BAK/iabak-sharp_implementation]]<br />
<br />
* [[Valhalla]] - Valhalla was a discussion about the "ultimate home" for uploaded Archive Team data, be it the Internet Archive, elsewhere, or both. IA.BAK is an example of a potential shared home, although Archive Team projects have not generally been backed up using IA.BAK, yet.<br />
<br />
__NOTOC__<br />
<br />
{{Navigation box}}</div>Antiufohttps://wiki.archiveteam.org/index.php?title=INTERNETARCHIVE.BAK/iabak-sharp_implementation&diff=45054INTERNETARCHIVE.BAK/iabak-sharp implementation2020-07-05T11:29:15Z<p>Antiufo: Add screenshot, fix typos</p>
<hr />
<div>'''[https://github.com/antiufo/iabak-sharp iabak-sharp]''' is an experimental implementation of [[INTERNETARCHIVE.BAK]].<br />
<br />
It is available for Windows and Linux, and is a command line tool that can be left running in the background.<br />
<br />
A central server takes care of coordinating the Internet Archive items that each client should back up.<br />
Each item can be given a priority score (currently, these priorities are assigned based on size and "uniqueness"<ref>For example, "warc-example1.com" has higher priority than all the "warc-example2-20200623", "warc-example2-20200624", "warc-example2-20200625" etc.</ref> of the item type).<br />
<br />
[[File:Iabak-sharp.png|thumb|right|iabak-sharp running on Windows 10.]]<br />
<br />
=== Currently implemented features ===<br />
<br />
* User registration (optional)<br />
* Retrieval of items from [[Internet Archive|IA]]<br />
* Hash consistency checks<br />
* Disk space checks<br />
* Coordination server and job assignment<br />
* Self-update<br />
* Download resume (file granularity)<br />
* Run on startup (Windows only)<br />
<br />
=== GitHub repository ===<br />
More info on the github page: [https://github.com/antiufo/iabak-sharp iabak-sharp]<br />
<br />
=== Comparison with git-annex implementation ===<br />
* Written in a more maintainable language (as opposed to bash)<br />
* No concept of shards: because we're not constrained by git repository size limits, each client only has to worry about the metadata of the files that they're actually storing on their drive. The server only stores a minimal amount of metadata (identifier, total size, and users having that item).<br />
* We're free to implement features that don't perfectly match the git use cases (eg. remote verification/challanges, encryption support, alternate distribution mechanisms eg. ipfs)<br />
* Supports Windows (in addition to Linux)<br />
* Single binary, no external dependencies.<br />
<br />
=== Notes===</div>Antiufohttps://wiki.archiveteam.org/index.php?title=INTERNETARCHIVE.BAK/iabak-sharp_implementation&diff=44986INTERNETARCHIVE.BAK/iabak-sharp implementation2020-07-03T21:48:30Z<p>Antiufo: Add iabak-sharp implementation for IA.BAK</p>
<hr />
<div>'''[https://github.com/antiufo/iabak-sharp iabak-sharp]''' is an experimental implementation of [[INTERNETARCHIVE.BAK]].<br />
<br />
It is available for Windows and Linux, and is a command line tool that can be left running in the background.<br />
<br />
A central server takes care of coordinating the Internet Archive items that each client should back up.<br />
Each item can be given a priority score (currently, these priority are assigned based on size and "uniqueness"<ref>For example, "warc-example1.com" has more priority than all the "warc-example2-20200623", "warc-example2-20200624", "warc-example2-20200625" etc.</ref> of the item type).<br />
<br />
=== Currently implemented features ===<br />
<br />
* User registration (optional)<br />
* Retrieval of items from IA<br />
* Hash consistency checks<br />
* Disk space checks<br />
* Coordination server and job assignment<br />
* Self-update<br />
* Download resume (file granularity)<br />
* Run on startup (Windows only)<br />
<br />
=== GitHub repository ===<br />
More info on the github page: [https://github.com/antiufo/iabak-sharp iabak-sharp]<br />
<br />
=== Comparison with git-annex implementation ===<br />
* Written in a more maintainable language (as opposed to bash)<br />
* No concept of shards: because we're not constrained by git repository size limits, each client only has to worry about the metadata of the files that they're actually storing on their drive. The server only stores a minimal amount of metadata (identifier, total size, and users having that item).<br />
* We're free to implement features that don't perfectly match the git use cases (eg. remote verification/challanges, encryption support, alternate distribution mechanisms eg. ipfs)<br />
* Supports Windows (in addition to Linux)<br />
<br />
=== Notes===</div>Antiufohttps://wiki.archiveteam.org/index.php?title=File:Iabak-sharp.png&diff=44985File:Iabak-sharp.png2020-07-03T21:23:06Z<p>Antiufo: </p>
<hr />
<div></div>Antiufohttps://wiki.archiveteam.org/index.php?title=Facebook&diff=41060Facebook2019-07-31T17:53:35Z<p>Antiufo: Abandoned project</p>
<hr />
<div>{{Infobox project<br />
| title = Facebook<br />
| image = Facebooklogo.png<br />
| description = Facebook Logo<br />
| URL = http://facebook.com<br />
| project_status = {{online}}<br />
| archiving_status = {{nosavedyet}}<br />
}}<br />
<br />
'''Facebook''' is a social networking site whose popularity has exploded in recent years. As of January 2014, there are more than ''1.2 billion'' active users of the site.<ref>http://thenextweb.com/facebook/2014/01/29/facebook-passes-1-23-billion-monthly-active-users-945-million-mobile-users-757-million-daily-users/</ref> Facebook hosts untold billions of users' photos, videos, thoughts, conversations, and other content. The total size of Facebook's databases is easily in the ''hundreds of petabytes''.<ref>https://www.facebook.com/notes/facebook-engineering/presto-interacting-with-petabytes-of-data-at-facebook/10151786197628920</ref><ref>https://research.facebook.com/blog/1522692927972019/facebook-s-top-open-data-problems/</ref><br />
<br />
The judicious user will have a well-designed backup plan for all that content that they retain full control over, but it is a reasonable assumption that the majority of users rely totally on Facebook to safeguard their data. '''This is a mistake.''' <br />
<br />
It might seem completely unthinkable that a site as massive and as popular as Facebook could ever disappear, taking your data with it. The reality is that websites, even hugely popular ones, can decline in popularity over time and eventually go away, taking your data with them with little or no warning. We've seen it happen.<br />
<br />
While Facebook may not be in any immediate danger, you should consider that the data you put on Facebook may be immensely important to you in 10 or 20 years, similar to your family's photo albums. Facebook could be long dead by then. Start planning for this eventuality right now. <br />
<br />
== Download Your Data From Facebook ==<br />
<br />
Facebook has created a tool to download an incomplete archive of your Facebook account. This includes your own photos and videos, chat conversations, status updates and wall posts. Please bear in mind that the archive is INCOMPLETE. It does NOT include any content you posted on any Facebook group. It does NOT include any comments from your wall posts. It also does NOT include photos and videos belonging to other people even if you are tagged in them.<br />
<br />
To create your archive, click the little down arrow next to your name in the upper right area of the page and go to "account settings". You should then see a screen like the one below: <br />
[[File:Fbdownload.png | border | center]]<br />
<br />
The next screen will explain what's going on. Press "Start my Archive" and you will be presented with a popup telling you that this will take some time - around one hour is not unheard of. Press Start again and Facebook will generate the file for you. This may indeed take several minutes. In the mean time you can continue using Facebook as usual. They will email you when the archive is ready for download. <br />
<br />
Your email will contain a link to download your archive. Follow that link and enter your Facebook password to continue. The next page presents you with a download button and an estimate of the archive's size. Download that somewhere convenient for you. '''This file contains highly personal and potentially sensitive information''' so keep it safe! You may wish to encrypt it with a password with a free tool like [http://www.axantum.com/axcrypt/ Axcrypt]. The easiest way to browse the information is to extract the contents of the zip file, and then open the index.html file with your browser of choice. From there you can look at your profile, your wall posts, photos and videos, and private messages. <br />
<br />
Note that as of April, 2012, this download tool seems to have some bugs -- in my tests it failed to completely back up all of my conversations, for example. It's better than nothing but for now at least I don't trust that it is perfect. <br />
<br />
=== Unofficial Backup tools ===<br />
<br />
* [http://www.friendstogmail.com/ friendstogmail] exports your contacts to CSV files, ostensibly for import to Gmail. Claims to work only in-browser. Retrieves your friends' birthdays, locations, bios, work history, and hometowns.<br />
<br />
* [http://www.vincentcheung.ca/facedown/ Facedown] downloads photo albums from Facebook. Official distribution of the software has stopped. <br />
<br />
* [http://photograbber.org/ Photograbber] is an open-source photo album downloader, capable of retrieving ragged photos, albums, tags, comments and likes. It can also download the albums of your friends.<br />
<br />
* [http://givememydata.com Give Me My Data (gmmd)] a Facebook application that helps users export their data out of Facebook for reuse in visualizations, archives, or any possible method of digital storytelling. Data can be exported in common formats like CSV, XML, and JSON as well as customized network graph formats.<br />
<br />
== Public pages and groups ==<br />
While Facebook contains personal profiles and Facebook versions of many blogs/news sites, it also contains blogs that only exist on Facebook. It could be nice to archive these pages, which are the most endangered:<br />
* Facebook is known to have banned totally reasonable pages in the past. They usually come back with a new profile, but the old content is gone.<br />
* Admins could delete pages and groups.<br />
* Pages can be bought and repurposed for commercial/spammy purposes.<br />
* Archival of pages and groups can be difficult (especially for general-purpose archivers), because Facebook is based on a (quite messy) implementation of AJAX. APIs are tightly locked and it is no longer possible to use them for interacting with/reading posts from groups.<ref>http://stackoverflow.com/questions/31490435/facebook-api-2-4-user-group-permission</ref><br />
* Alternatively, one could disable JavaScript and download web pages from the mobile website, [https://m.facebook.com/ m.facebook.com]. Archival of pages and groups using this method is much easier.<br />
<br />
=== Proposal ===<br />
We could start archiving the most popular public pages that 1) are not backed by an external website and 2) are not simply used for boring promotional purposes (eg. pages belonging to brands).<br />
<br />
TODO: We need to build a list of the most important pages matching the above criteria.<br />
<br />
== Vital Signs ==<br />
<br />
Currently stable.<br />
<br />
== References ==<br />
<references /><br />
<br />
== External links ==<br />
<br />
* http://facebook.com<br />
<br />
{{Navigation box}}<br />
[[Category:Social networks]]</div>Antiufohttps://wiki.archiveteam.org/index.php?title=MuseScore&diff=34309MuseScore2019-01-24T18:47:20Z<p>Antiufo: Created page with "{{Infobox project | title = MuseScore | URL = http://musescore.com | project_status = {{online}} | archiving_status = {{notsavedyet}} }} '''MuseScore''' is a scorewriting soft..."</p>
<hr />
<div>{{Infobox project<br />
| title = MuseScore<br />
| URL = http://musescore.com<br />
| project_status = {{online}}<br />
| archiving_status = {{notsavedyet}}<br />
}}<br />
'''MuseScore''' is a scorewriting software, and a community where users can upload and share music scores. Each free user can upload up to 5 scores.<br />
Scores are available in open formats (MID, PDF, MXL and MSCZ), but they are only available to authenticated users (making it harder for archival tools to preserve the site's content).</div>Antiufohttps://wiki.archiveteam.org/index.php?title=Mozilla_Addons&diff=29809Mozilla Addons2017-10-16T11:24:08Z<p>Antiufo: </p>
<hr />
<div>{{Infobox project<br />
| title = Mozilla Addons<br />
| URL = http://addons.mozilla.org/<br />
| project_status = {{closing}}<br />
| archiving_status = {{notsaved}}<br />
}}<br />
<br />
'''Mozilla Addons''' will remove all legacy (XPI-based) addons from its website.<ref>https://blog.mozilla.org/addons/2017/10/03/legacy-add-on-support-on-firefox-esr/#comment-224382</ref><br />
<br />
Mozilla has recently decided to deprecate all XPI-based addons. Firefox 57 is the first version that only supports WebExtensions, the new addon format.<br />
<br />
The total number of addons should be approximately 20,000.<ref>Based on pagination and results per page: https://addons.mozilla.org/en-US/firefox/extensions/?sort=created</ref><br />
<br />
== References ==<br />
<references/></div>Antiufohttps://wiki.archiveteam.org/index.php?title=Mozilla_Addons&diff=29808Mozilla Addons2017-10-16T11:17:40Z<p>Antiufo: Created page with "{{Infobox project | title = Mozilla Addons | URL = http://addons.mozilla.org/ | project_status = {{closing}} | archiving_status = {{notsaved}} }} '''Mozilla Addons''' will re..."</p>
<hr />
<div>{{Infobox project<br />
| title = Mozilla Addons<br />
| URL = http://addons.mozilla.org/<br />
| project_status = {{closing}}<br />
| archiving_status = {{notsaved}}<br />
}}<br />
<br />
'''Mozilla Addons''' will remove all legacy (XPI-based) addons from its website.<ref>https://blog.mozilla.org/addons/2017/10/03/legacy-add-on-support-on-firefox-esr/#comment-224382</ref><br />
<br />
Mozilla has recently decided to deprecate all XPI-based addons. Firefox 57 is the first version that only supports WebExtensions, the new addon format.<br />
<br />
== References ==<br />
<references/></div>Antiufohttps://wiki.archiveteam.org/index.php?title=Facebook&diff=29299Facebook2017-03-24T00:48:32Z<p>Antiufo: Archival of Facebook public pages</p>
<hr />
<div>{{Infobox project<br />
| title = Facebook<br />
| image = Facebooklogo.png<br />
| description = Facebook Logo<br />
| URL = http://facebook.com<br />
| project_status = {{online}}<br />
| archiving_status = {{nosavedyet}}<br />
}}<br />
<br />
'''Facebook''' is a social networking site whose popularity has exploded in recent years. As of January 2014, there are more than ''1.2 billion'' active users of the site.<ref>http://thenextweb.com/facebook/2014/01/29/facebook-passes-1-23-billion-monthly-active-users-945-million-mobile-users-757-million-daily-users/</ref> Facebook hosts untold billions of users' photos, videos, thoughts, conversations, and other content. The total size of Facebook's databases is easily in the ''hundreds of petabytes''.<ref>https://www.facebook.com/notes/facebook-engineering/presto-interacting-with-petabytes-of-data-at-facebook/10151786197628920</ref><ref>https://research.facebook.com/blog/1522692927972019/facebook-s-top-open-data-problems/</ref><br />
<br />
The judicious user will have a well-designed backup plan for all that content that they retain full control over, but it is a reasonable assumption that the majority of users rely totally on Facebook to safeguard their data. '''This is a mistake.''' <br />
<br />
It might seem completely unthinkable that a site as massive and as popular as Facebook could ever disappear, taking your data with it. The reality is that websites, even hugely popular ones, can decline in popularity over time and eventually go away, taking your data with them with little or no warning. We've seen it happen.<br />
<br />
While Facebook may not be in any immediate danger, you should consider that the data you put on Facebook may be immensely important to you in 10 or 20 years, similar to your family's photo albums. Facebook could be long dead by then. Start planning for this eventuality right now. <br />
<br />
== Download Your Data From Facebook ==<br />
<br />
Facebook has created a tool to download an entire archive of your Facebook account. This includes all of your own photos and videos, chat conversations, messages, status updates and wall posts. It does NOT include photos and videos belonging to other people even if you are tagged in them, so do keep that in mind. <br />
<br />
To create your archive, click the little down arrow next to your name in the upper right area of the page and go to "account settings". You should then see a screen like the one below: <br />
[[File:Fbdownload.png | border | center]]<br />
<br />
The next screen will explain what's going on. Press "Start my Archive" and you will be presented with a popup telling you that this will take some time - around one hour is not unheard of. Press Start again and Facebook will generate the file for you. This may indeed take several minutes. In the mean time you can continue using Facebook as usual. They will email you when the archive is ready for download. <br />
<br />
Your email will contain a link to download your archive. Follow that link and enter your Facebook password to continue. The next page presents you with a download button and an estimate of the archive's size. Download that somewhere convenient for you. '''This file contains highly personal and potentially sensitive information''' so keep it safe! You may wish to encrypt it with a password with a free tool like [http://www.axantum.com/axcrypt/ Axcrypt]. The easiest way to browse the information is to extract the contents of the zip file, and then open the index.html file with your browser of choice. From there you can look at your profile, your wall posts, photos and videos, and private messages. <br />
<br />
Note that as of April, 2012, this download tool seems to have some bugs -- in my tests it failed to completely back up all of my conversations, for example. It's better than nothing but for now at least I don't trust that it is perfect. <br />
<br />
=== Unofficial Backup tools ===<br />
<br />
* [http://www.friendstogmail.com/ friendstogmail] exports your contacts to CSV files, ostensibly for import to Gmail. Claims to work only in-browser. Retrieves your friends' birthdays, locations, bios, work history, and hometowns.<br />
<br />
* [http://www.vincentcheung.ca/facedown/ Facedown] downloads photo albums from Facebook. Official distribution of the software has stopped. <br />
<br />
* [http://photograbber.org/ Photograbber] is an open-source photo album downloader, capable of retrieving ragged photos, albums, tags, comments and likes. It can also download the albums of your friends.<br />
<br />
* [http://givememydata.com Give Me My Data (gmmd)] a Facebook application that helps users export their data out of Facebook for reuse in visualizations, archives, or any possible method of digital storytelling. Data can be exported in common formats like CSV, XML, and JSON as well as customized network graph formats.<br />
<br />
== Public pages ==<br />
While Facebook contains personal profiles and Facebook versions of many blogs/news sites, it also contains blogs that only exist on Facebook. It could be nice to archive these pages, which are the most endangered:<br />
* Facebook is known to have banned totally reasonable pages in the past. They usually come back with a new profile, but the old content is gone.<br />
* Pages can be bought and repurposed for commercial/spammy purposes.<br />
* Archival of pages and groups can be difficult (especially for general-purpose archivers), because Facebook is based on a (quite messy) implementation of AJAX. APIs are tightly locked and it is no longer possible to use them for interacting with/reading posts from groups.<ref>http://stackoverflow.com/questions/31490435/facebook-api-2-4-user-group-permission</ref><br />
<br />
=== Proposal ===<br />
We could start archiving the most popular public pages that 1) are not backed by an external website and 2) are not simply used for boring promotional purposes (eg. pages belonging to brands).<br />
<br />
[[User:Antiufo|Antiufo]] has started archiving some pages with a custom tool that will soon be posted on GitHub.<br />
<br />
Most pages are in the 50MB-150MB range (posts + normal-sized images + top comments)<br />
<br />
TODO: We need to build a list of the most important pages matching the above criteria.<br />
<br />
== Vital Signs ==<br />
<br />
Currently stable.<br />
<br />
== References ==<br />
<references /><br />
<br />
== External links ==<br />
<br />
* http://facebook.com<br />
<br />
{{Navigation box}}<br />
[[Category:Social networks]]</div>Antiufo