https://wiki.archiveteam.org/api.php?action=feedcontributions&user=Systwi&feedformat=atomArchiveteam - User contributions [en]2024-04-10T06:08:12ZUser contributionsMediaWiki 1.37.1https://wiki.archiveteam.org/index.php?title=TikTok&diff=51895TikTok2024-03-17T21:57:36Z<p>Systwi: /* Notes */ Grammar tweak.</p>
<hr />
<div>{{Infobox project<br />
| title = TikTok<br />
| logo = TikTok logo.png<br />
| image = TikTok home page 2019-11-09.png<br />
| description = tiktok home page as seen on November 9, 2019<br />
| URL = https://www.tiktok.com/<br />
| project_status = {{online}} but possibly {{endangered}}<br />
| archiving_status = {{nosavedyet}}<br />
| irc = TikOff<br />
}}<br />
<br />
'''TikTok''' is a social media app and Website for short, user generated videos developed by the Chinese company ByteDance. The app was launched internationally in September 2017 and merged with musical.ly on 9 November 2017. It is currently under intense scrutiny for its collection and use of user data, primarily in its relationship with China.<ref>https://en.wikipedia.org/wiki/TikTok#User_privacy_concerns</ref><br />
<br />
== Douyin ==<br />
<br />
Douyin(抖音) is the China-only version of TikTok.<br />
<br />
Sites:<br />
* https://www.douyin.com | Main page, where videos are<br />
* https://sso.douyin.com | Login to Douyin<br />
* https://renzheng.douyin.com | Douyin for enterprise<br />
* https://open.douyin.com | API and other developer tools<br />
* https://musician.douyin.com | Douyin for musician<br />
* https://creator.douyin.com | Douyin for creator<br />
* https://www.iesdouyin.com | Redirects to main page<br />
<br />
If anyone who can read Chinese would like to help out, that'd be great.<br />
<br />
== Archiving tools ==<br />
<br />
[https://github.com/yt-dlp/yt-dlp yt-dlp] is able to grab videos straight from the site<br />
<br />
A tool to [https://github.com/carcabot/tiktok-signature generate the signature] TikTok uses to sign all their requests exists and is actively updated<br />
<br />
=== TikTokTake ===<br />
<br />
Work is being done to create a tool that can get all of a users videos and WARC them up, that's being implemented in a cli tool and irc bot [https://github.com/JackDallas/TikTokTake TikTokTake]<br />
<br />
=== TikUp ===<br />
<br />
TikUp is a Python package that gets all videos from a user or a hashtag, and uploads them to the Internet Archive under `tiktok-<video id>`. [https://github.com/Coloradohusky/TikUp Github link]<br />
<br />
== Vital Signs ==<br />
<br />
* The first request for archival was in [https://old.reddit.com/r/Archiveteam/comments/dazw5l/tiktok/ October 2019] due to removal of pro-LGBTQ content, even in countries where homosexuality has never been illegal. <ref>[https://www.theguardian.com/technology/2019/sep/26/tiktoks-local-moderation-guidelines-ban-pro-lgbt-content "TikTok's local moderation guidelines ban pro-LGBT content"] - Article on ''The Guardian''</ref><br />
* Videos from Douyin (Chinese TikTok) can not be seen outside of China. <ref>[https://www.wired.com/story/tiktok-is-the-latest-window-into-chinas-police-state/ "TikTok - Yes, TikTok - Is the Latest Window Into China’s Police State"] - Article on ''Wired''</ref><br />
* In August 2020 Trump said "As far as TikTok is concerned, we're banning them from the United States," <ref>[https://web.archive.org/web/20200801075923/https://edition.cnn.com/2020/07/31/tech/tiktok-trump-bytedance-sale/index.html "Trump says he will ban TikTok"] - Article on CNN</ref> <br />
* In 2020-09-18, the Trump administrator will ban downloads of the TikTok app starting 2020-09-20, and access will be banned on November 12.<ref>[https://www.abc.net.au/news/2020-09-19/us-bans-wechat-and-tik-tok-over-national-security-concerns-china/12680516]</ref><br />
<br />
== Archival Locations ==<br />
<br />
[https://archive.org/details/tiktoks Tiktok Video Bin] on the [[Internet Archive]] is where the TikToks uploaded with TikUp and all videos uploaded with the identifier 'tiktok-<video id>' go. The Internet Archive deindexed this collection some time in 2020.<ref>https://archive.fart.website/bin/irclogger_log/archiveteam-ot?date=2021-04-13,Tue</ref><br />
<br />
== Notes ==<br />
<br />
* https://m.tiktok.com/v/<id> is a better way to archive videos, versus https://www.tiktok.com/@<username>/video/<id>. (because it requires the ID only, not the username and the ID)<br />
* Videos without sound can't be viewed directly on desktop (it just says 'Video currently unavailable', however viewing them via user page will work), and their links don't work, but they can be viewed and downloaded on mobile.<br />
* If we wanted to make a Warrior, we could use https://vm.tiktok.com/*******/, I think it's alphanumeric, but I'm not sure if these links work for every TikTok.<br />
** A minor caution regarding using https://vm.tiktok.com/*******/ URLs; they redirect to the full video URLs mentioned above, but also include telemetry parameters (<code>_t=...</code> and <code>_r=...</code>). It is believed that the former parameter equates to a unique identifier TikTok has assigned to both authenticated and non-authenticated users.<br />
<br />
{{Navigation box}}<br />
[[Category:Video hosting|Video hosting websites]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=IRC&diff=51105IRC2023-11-06T20:27:51Z<p>Systwi: Added some IRC server URLs, sorted the list descending by subdomain and made a few other minor adjustments.</p>
<hr />
<div>:''Are you looking for '''[[Archiveteam:IRC|ArchiveTeam's IRC]]'''?''<br />
<br />
'''Internet Relay Chat (IRC)''' is an old chat protocol, being designed in 1988. IRC clients are simple and available for every platform imaginable, including your old Nintendo DS. IRC is fully decentralized, with anybody being able to run a server. There are numerous large networks with their own histories. With the advent of modern chat platforms that are more mobile-friendly among other advantages, IRC usage has been on the decline, and lately (2019) it is being abandoned even by techie folks.<br />
<br />
== IRC network aggregators ==<br />
There are a few IRC network aggregators, but the biggest one is probably [http://irc.netsplit.de/networks/top100.php irc.netsplit.de], which also scrapes channels, their historical topics and number of users.<br />
<br />
* [http://irc.netsplit.de/networks/top100.php irc.netsplit.de] ({{job|2emzfqe3gkdnus4wsdozh1mo5}} on 2019-04-30)<br />
<br />
== IRC networks ==<br />
<br />
On 2019-04-26, Mozilla announced the plan to deprecate and decommission its IRC network in the "next small number of months"[http://exple.tive.org/blarg/2019/04/26/synchronous-text/], with the Rust community already making the move to [[Discord]][https://blog.rust-lang.org/2019/04/26/Mozilla-IRC-Sunset-and-the-Rust-Channel.html].<br />
<br />
== IRC quotes ==<br />
There are a lot of sites providing [[IRC_Quotes|quotes from IRC channels]].<br />
<br />
== IRC logs ==<br />
IRC servers typically do not provide logs. Clients usually keep local logs of chats, however, unless run on a reliable server running 24/7 those logs will contain blind spots.<br />
<br />
There is an implicit agreement between IRC participants to not publish IRC logs. However, many channels have a notice about public logs in their topic and some log servers are still available online. The [[IRC/Logs|IRC Logs project]] aims to capture them.<br />
<br />
== (Incomplete) List of IRC Server URLs ==<br />
<br />
* irc://chat.freenode.net/<br />
* ircs://irc.badnik.net/<br />
* irc://irc.dal.net/<br />
* irc://irc.efnet.org/<br />
* ircs://irc.hackint.org/<br />
* ircs://irc.hak5.org/<br />
* ircs://irc.irc-nerds.net/<br />
* ircs://irc.irc4fun.net/<br />
* ircs://irc.libera.chat/<br />
* ircs://irc.lunarirc.net/<br />
* ircs://irc.oftc.net/<br />
* ircs://irc.r-type.ca/<br />
* ircs://irc.retroit.org/<br />
* ircs://irc.rizon.net/<br />
* ircs://irc.sdf.org/<br />
* ircs://irc.snoonet.org/<br />
* ircs://irc.spotchat.org/<br />
* irc://irc.undernet.org/<br />
* ircs://irc.zenet.org/<br />
<br />
<br />
{{Template:Instant messengers}}</div>Systwihttps://wiki.archiveteam.org/index.php?title=Site_exploration&diff=50594Site exploration2023-08-27T02:39:16Z<p>Systwi: /* Twitter */ Mentioned Nitter and Twitter's false 200s.</p>
<hr />
<div>This page contains some '''tips and tricks for exploring soon-to-be-dead websites''', to find URLs to feed into the Archive Team crawlers.<br />
<br />
== Wayback Machine ==<br />
The Wayback Machine's CDX server can be used to find pages already in the WBM (with caveats, see [[User:OrIdow6/Info]]). An easy way to get all pages is (for instance with [[Egloos]]):<br />
<br />
<code>curl 'https://web.archive.org/cdx/search?url=*.egloos.com&showNumPages=true'</code><br />
<br />
(this gives 749 pages)<br />
<br />
<code>seq 0 708 | awk '{ printf("https://web.archive.org/cdx/search?url=*.egloos.com&page=%s&fl=urlkey\n", $1) }' | wget --input-file - --retry-on-http-error=429</code><br />
<br />
To enumerate the subdomains of a domain that is in the WBM:<br />
<br />
<code><br />
domain=egloos.com<br />
pages=$(curl -s "https://web.archive.org/cdx/search/cdx?url=*.$domain&collapse=original&fl=original&showNumPages=1)<br />
seq 0 $pages |<br />
xargs -d '\n' printf "https://web.archive.org/cdx/search/cdx?url=*.$domain&collapse=original&fl=original&page=%s\n" |<br />
wget --input-file - --retry-on-http-error=429 |<br />
grep -oP "^https?://[^/]*$domain(:[0-9]+)?/" |<br />
sed 's/:[0-9]\+//' |<br />
sort -u</code><br />
<br />
==Open Directory Project data==<br />
The Open Directory Project offers [http://www.dmoz.org/rdf.html machine-readable downloads] of its data. You want the "content.rdf.u8.gz" from there.<br />
<br />
wget <nowiki>http://rdf.dmoz.org/rdf/content.rdf.u8.gz</nowiki><br />
gunzip content.rdf.u8.gz<br />
<br />
Quick-and-dirty shell parsing for the not-too-fussy:<br />
<br />
<nowiki>grep '<link r:resource=.*dyingsite\.com' content.rdf.u8 | sed 's/.*<link r\:resource="\([^"]*\).*".*/\1/' | sort | uniq > odp-sitelist.txt</nowiki><br />
<br />
==MediaWiki wikis==<br />
MediaWiki wikis, especially the very large ones operated by the Wikimedia Foundation, often return a large number of important sites hosted with a service.<br />
<br />
[https://github.com/lewiscollard/mwlinkscrape mwlinkscrape.py] is a tool by an Archive Team patriot which extracts a machine-readable list from a number of wikis (it actually uses the text of [[List of major MediaWiki wikis with the LinkSearch extension|this page]] to get a list of wikis to scrape). <br />
<br />
./mwlinkscrape.py "*.dyingsite.com" > mw-sitelist.txt<br />
<br />
== Search Engines ==<br />
<br />
There exists tools such as [https://github.com/NikolaiT/GoogleScraper GoogleScraper] which will scrape various search engines using a web browser instead of an API.<br />
<br />
Other tools such as [https://github.com/oksiquatzel/googler googler] or [https://github.com/jarun/ddgr ddgr] download and parse search results HTML pages.<br />
<br />
Below lists some specific helpful tips.<br />
<br />
=== Google ===<br />
Google doesn't let its search results be scraped by automated tools. One must do it manually, but there are some tools and tips that still let you do a good discovery quite quickly.<br />
<br />
To find results under a domain, let your search term be <code>site:dyingsite.com</code>.<br />
<br />
To find results under subdomains of a domain, let your search term be <code>site:*.dyingsite.com</code>.<br />
<br />
If you want more than 10 results per page, add the <code>num</code> parameter to the URL like this:<br />
<br />
<code>https://www.google.com/search?q=site:dyingsite.com&amp;num=100</code><br />
<br />
To go to the next page of the results, don't use the "Next" link on the bottom; that would give you the next ''ten'' results. Instead, use the <code>start</code> parameter in the URL:<br />
<br />
<code>https://www.google.com/search?q=site:dyingsite.com&amp;num=100&amp;start=100</code><br />
<br />
You can go up to 300, that is the 301–400th results. Google doesn't let you browse more than the first 400 results. However, there are some good news:<br />
<br />
* The estimated number of results shown is usually like ten or hundred times more than the actual number of results you'll be presented. So don't panic.<br />
* Should the number of results be indeed more than 400, the easiest workaround is clicking on "Search tools", then on "Any time", and selecting "Custom range". Setting a specific range, you can reduce the number of results in one search, and going, say, year by year, you'll be presented with all the results. (Hopefully.)<br />
<br />
For exporting the results efficiently, there must be several tools around. One of them is ''SEOquake'', a Firefox extension. (In fact, exporting search results is just one feature of it.) After installing the extension and restarting the browser, Google search results will have buttons to export (save or append) the results in CSV format. (It is recommended to disable – in SEOquake options – all those analysis appearing in the search results and on the toolbar, they are just slowing things down and occupy a lot of space in the CSV.) – After some repetitive but easy work, you'll have the URL list in your CSV(s). If SEOquake analysis things are turned off, it will be just the URLs embraced with quotation marks. Replace "s with nothing in a text editor, or for Linux terminal geeks, <code> cut -d'"' -f 2</code> is the way.<br />
<br />
Should Google stand in your way with a captcha, fill it, then you can proceed. (Cookies must be enabled for it to work.)<br />
<br />
===Bing API===<br />
Microsoft, bless their Redmondish hearts, have an [https://datamarket.azure.com/dataset/bing/searchweb API for fetching Bing search engine results], which has a free tier of 5000 queries per month (this will cover you for about 250 sets of 1000 results). However, it only returns the first 1000 results for any query, so you can't just search "site:dyingsite.com" and get all the things on a site. You'll need to get a bit creative with the search terms.<br />
<br />
Grab [https://web.archive.org/web/20140725221029/http://paste.archivingyoursh.it/raw/fehelaqoyu this Python script] (look for "BING_API_KEY" and replace it with your "Primary Account Key"), and then:<br />
<br />
python bingscrape.py "site:dyingsite.com" >> bing-sitelist.txt<br />
python bingscrape.py "about me site:dyingsite.com" >> bing-sitelist.txt<br />
python bingscrape.py "gallery site:dyingsite.com" >> bing-sitelist.txt<br />
python bingscrape.py "in memoriam site:dyingsite.com" >> bing-sitelist.txt<br />
python bingscrape.py "diary site:dyingsite.com" >> bing-sitelist.txt<br />
python bingscrape.py "bob site:dyingsite.com" >> bing-sitelist.txt<br />
<br />
And so on.<br />
<br />
To find subdomains use e.g. <code>site:example.com+</code>.<br />
<br />
==Common Crawl Index==<br />
<br />
The [http://archive.org/details/2013_common_crawl_index_urls Common Crawl index] is a very big (21 gigabytes compressed) list of URLs in the Common Crawl corpus. Grepping this list may well reveal plenty of URLs to archive. The list is in an odd format; along the lines of <code>com.deadsite.www/subdirectory/subsubdirectory:http</code> so you'll need to some filtering of the results. The results can [https://github.com/trivio/common_crawl_index/issues/12 sometimes be ambiguous].<br />
<br />
grep '^com\.dyingsite[/\.]' zfqwbPRW.txt > commoncrawl-sitelist.txt<br />
<br />
Our [[User:Ivan|Ivan]] wrote a [http://paste.archivingyoursh.it/taguqoxora.py Python script] ([http://pastebin.com/MDB5NWHm Mirror]) which will take your list of URLs on standard input and print out a list of normally-formed URLs on standard output.<br />
<br />
You can also use the [http://urlsearch.commoncrawl.org/ Common Crawl URL search] and get the results as a JSON file. Quick-and-dirty grep/sed parsing:<br />
<br />
grep -F '"url":' locations.json | sed 's/.*url": "\([^"]*\).*/\1/' | sort | uniq > commoncrawl-sitelist.txt<br />
<br />
<br />
During the [[Imgur]] project a group tried to scrape not just the Common Crawl index but the entire Common Crawl dataset, which led to complaints from them directed at us.<br />
<br />
== Twitter ==<br />
<br />
* Twitter's search API doesn't offer historical results. However, their web search does a complete index now<ref>https://blog.twitter.com/2014/building-a-complete-tweet-index</ref> including a searching expanded URLs.<br />
<br />
* A tool like [https://github.com/chfoo/litterapi Litterapi] will scrape their web search and build a fake API.<br />
<br />
* [https://github.com/sferik/t t by sferik] is a command-line interface for Twitter using the API via an application you create on your account. Not only does it allow easy CSV/JSON export of your own data, but it allows you to scrape others tweets. API limits apply but this tool is <b>very</b> powerful<br />
<br />
* [http://topsy.com Topsy] offers a competing search service with an API of all Tweets. However, it is not free (but perhaps you can borrow their API key) and does not search expanded URLs.<br />
<br />
* [https://github.com/zedeus/nitter Nitter] is an alternative Twitter front-end that, when working (as Twitter appear to actively resist its use), can have significantly greater chances of successful data archival using tools such as [[ArchiveBot]]. On a similar note, the [https://www.twitter.com/ vanilla Twitter web site] is known to return 200s on pages in which their access has been denied.<br />
<br />
== Subdomain enumeration ==<br />
<br />
[[Finding subdomains|Locating subdomains]] is important to get complete scrapes of websites and is particularly critical for many sites offering to host user content.<br />
<br />
== See Also ==<br />
<br />
* [[Keywords]]<br />
* [[Collecting items randomly]]<br />
<br />
== References == <br />
<references/><br />
<br />
[[Category:Archive Team]]<br />
<br />
{{Navigation box}}</div>Systwihttps://wiki.archiveteam.org/index.php?title=Twitter&diff=50593Twitter2023-08-27T02:27:12Z<p>Systwi: /* Vital Signs */ Added meme and serious caption.</p>
<hr />
<div>{{Infobox project<br />
| title = Twitter<br />
| image = Twitter_account_timeline.png<br />
| description = <br />
| URL = https://twitter.com/<br />
| project_status = {{endangered}}<br />
| archiving_status = {{notsavedyet}}<br />
}}<br />
'''Twitter''' is a microblogging service. With each "entry" being 280 (formerly 140) characters or less, the ease with which you can track the tiniest details of your life is amazing. The site has become very popular as a result.<br />
<br />
The site is becoming so popular, in fact, that many people are deserting or cutting back on their weblogs to just use the Twitter service for what their weblogging used to fulfill; and with that comes rampant centralization, and with ''that'', greater risk. Back up your tweets!<br />
<br />
== Archives ==<br />
There are currently a few archives (but only partially):<br />
* [https://archive.org/details/twitter_cikm_2010 Cheng-Caverlee-Lee September 2009 - January 2010 Twitter Scrape]: almost 10 millon tweets<br />
* [https://archive.org/details/2011-05-calufa-twitter-sql The May 2011 Calufa Twitter Scrape]: 90+ million tweets from more than 6 million users<br />
* {{IA item|twitterstream}}<br />
<br />
The Twitter search API seemingly returns only the latest 7 days worth of tweets.<br />
<br />
== Backup Tools ==<br />
<br />
* Twitter enables you to [https://twitter.com/settings/your_twitter_data request an archive of all of your tweets from the main settings page], which includes every tweet of yours (therefore bypassing the normal 3200 tweet API limit). This is then emailed to the address linked with the account.<br />
<br />
* [https://github.com/JustAnotherArchivist/snscrape snscrape]<br />
<br />
* Tweets can be archived easily through socialbot (an IRC bot that utilizes snscrape, [[ArchiveBot]]. Alternatively, run snscrape locally, upload the list of URLs to https://transfer.archivete.am/, and feed them into ArchiveBot using the <code>!ao < <i>file</i></code> command.<br />
<br />
* [https://github.com/sferik/t t by sferik] is a command-line interface for Twitter using the API via an application you create on your account. Not only does it allow easy CSV/JSON export of your own data, but it allows you to scrape others tweets. API limits apply but this tool is <b>very</b> powerful<br />
<br />
Twitter automatically resizes uploaded images. To get image in its original resolution, append :orig after the url, e.g.:<br />
https://pbs.twimg.com/media/CBAoaU1UwAIUPIc.jpg:orig<br />
<br />
* [https://github.com/sixohsix/twitter The Python Twitter API by sixohsix] has some pretty easy to use scripts for archiving Twitter accounts to a TXT file for people who aren't as technically inclined. It can only save the last 3K or so tweets due to inbuilt Twitter limits, though. (Note: the "-o" flag is pretty much required to archive accounts.)<br />
<br />
* [https://github.com/DocNow/twarc twarc] - for example, you can archive a user with <code>snscrape -f "{id}" twitter-user USER | twarc hydrate > data.txt</code>.<br />
<br />
* [https://gist.github.com/Asparagirl/e3ee274e4df49230875c880255819d95 Here's a Gist with a step-by-step guide] to getting a long list of a user's tweet status URL's, using a Python program called Tweep.<br />
<br />
* [https://web.archive.org/web/20140105210020/http://www.tweetscan.com:80/data.php Tweetscan Data] (died sometime after 2014) downloaded your Twitter archive from 12/2007 onward in CSV format (requires Twitter account login/password)<br />
<br />
* [https://github.com/mikf/gallery-dl gallery-dl]'s Twitter extractor still seems to work as of 2023-07-17<br />
<br />
=== Scraping ===<br />
<br />
See [[Site exploration#Twitter|Site exploration]] for details.<br />
<br />
== Vital Signs == <br />
[[File:Duck Hunt (World)-0--twitter-5.png|frame|right|Many changes to Twitter since Musk's acquisition render this service's longevity (and archivability) rather questionable.]]<br />
In 2017, a subsidiary forum, "translate.twitter.com", was discontinued.<ref>https://archive.fo/https://translate.twitter.com/forum%2A</ref><br />
<br />
In 2022, Elon Musk bought Twitter.<ref>{{URL|https://www.hindustantimes.com/world-news/end-of-free-twitter-elon-musk-hints-he-may-charge-commercial-govt-users-101651624937957.html}}</ref> This may lock up accessible tweets behind paywall, login wall or identity verification wall.<br />
<br />
The main site is still online, although Elon Musk fired half of the Twitter staff in November 2022<ref>{{URL|https://www.theguardian.com/technology/2022/nov/04/twitter-layoffs-elon-musk-revenue-drop}}</ref> so it's possible this could ultimately put the site in jeopardy. He also reportedly warned employees that the site is at risk of going bankrupt in 2023<ref>{{URL|https://finance.yahoo.com/news/musk-says-twitter-could-face-bankruptcy-next-year-report-221322212.html}}</ref>.<br />
<br />
As of November 17, 2022, many former Twitter employees are saying that many vital Twitter services are now being maintained by a single-digit number of technicians. This could ultimately result in Twitter losing its services.<!--citation needed--><br />
<br />
On December 9, 2022, Elon Musk announced that accounts with no Tweets that have no activity for an extended period of time would be deleted to make room for new accounts. As an alternative interpretation of the statement, this could refer to accounts that have not Tweeted in a long time.<ref>{{URL|https://twitter.com/elonmusk/status/1601124219009409024}}</ref><br />
<br />
On February 2, 2023, Twitter announced free API access will be discontinued and replaced with a paid subscription tier. <ref>{{URL|https://twitter.com/TwitterDev/status/1621026986784337922}}</ref><br />
<br />
During March 2023, Twemoji's maintainers were fired. Its usage in Twitter's mobile apps has since been dropped, with the native emojis now being used. A fork of Twemoji has been created by the ex-maintainers and is currently used by [[Discord]].<ref>{{URL|https://twitter.com/lexikiq/status/1641186153742123009}}</ref><br />
<br />
During April 2023, Twitter Inc. was folded into Elon Musk's new company X Corp. As a result, it's possible that Twitter could inevitably be rebranded into "X" to align with Elon's visions of an "everything app".<ref>{{URL|https://www.independent.co.uk/tech/twitter-no-longer-exists-x-corp-b2318916.html}}</ref> This rebrand would eventually be initiated in July 2023, with the Twitter branding and logo being phased out.<ref>{{URL|https://edition.cnn.com/2023/07/24/tech/twitter-rebrands-x-elon-musk-hnk-intl/index.html}}</ref><br />
<br />
In mid-April 2023, access to Twitter's search function was locked behind a login prompt, requiring users to have an account in order to search for tweets. This makes it more difficult to scrape tweets. <ref>{{URL|https://www.neowin.net/news/the-login-prompt-for-twitter-search-has-returned-just-months-after-elon-musk-wanted-it-gone/}}</ref><ref>{{URL|https://mashable.com/article/twitter-search-registered}}</ref><br />
<br />
On 8 May 2023, Elon Musk announced that Twitter is "purging accounts that have had no activity at all for several years".<ref>{{URL|https://twitter.com/elonmusk/status/1655608985058267139}}</ref> He later clarified that "the accounts will be archived".<ref>{{URL|https://twitter.com/elonmusk/status/1655720120440823809}}</ref> Twitter's "inactive account policy" defines inactive accounts as those that have not been logged in for at least 30 days, and says that accounts "may be permanently removed due to prolonged inactivity".<ref>{{URL|https://help.twitter.com/en/rules-and-policies/inactive-twitter-accounts}}</ref> (Before May 2023, the time needed for an account to become "inactive" was 6 months.)<ref>https://web.archive.org/web/20230419143120/https://help.twitter.com/en/rules-and-policies/inactive-twitter-accounts</ref><br />
<br />
On 30 June 2023, Twitter made it impossible to view tweets without logging into an account.<ref>{{URL|https://www.engadget.com/twitter-isnt-showing-tweets-unless-youre-logged-in-165254006.html}}</ref> This was likely done in response to decreasing user activity levels.<ref>{{URL|https://www.zdnet.com/article/twitter-seeing-record-user-engagement-the-data-tells-a-different-story/}}</ref> Elon Musk later wrote on Twitter: "This will be unlocked shortly. Per my earlier post, drastic & immediate action was necessary due to EXTREME levels of data scraping."<ref>{{URL|https://twitter.com/elonmusk/status/1674942336583757825|Elon Musk on Twitter}} (via [https://news.ycombinator.com/item?id=36540957 Hacker News])</ref> Logged-in users are also subject to a per-day post rate limit of 10000 (for Twitter Blue subscribers), 1000 (for normal accounts), or 500 (for new, not-subscribed accounts).<ref>{{URL|https://twitter.com/elonmusk/status/1675260424109928449}}</ref> On 5 July, Twitter silently removed the login requirement to view individual tweets, although user feeds and tweet threads still require logging in.<ref>{{URL|https://techcrunch.com/2023/07/05/twitter-silently-removes-login-requirement-for-viewing-tweets/}}</ref><br />
<br />
== Also see ==<br />
* [[Twitter watchlist]] – Accounts with volatile content.<br />
* [[List of lost Twitter accounts]] – Twitter accounts that went private, disabled, suspended or deleted a significant amount of content.<br />
<br />
== Library of Congress ==<br />
<br />
The U.S. Library of Congress announced in April 2010, via its official Twitter account that it will be acquiring the entire archive of Twitter messages back through March 2006.<ref>{{URL|http://www.readwriteweb.com/archives/twitters_entire_archive_headed_to_the_library_of_c.php}}</ref> As of 2016-02-23, this archive is still not available, and when/if it does become accessible it will likely be restricted to researchers, rather than the general public.<ref>{{URL|https://www.politico.com/story/2015/07/library-of-congress-twitter-archive-119698.html}}</ref> In January 2017, it was announced that the Library of Congress will no longer archive all tweets, just ones from major news stories.<ref>{{URL|https://www.npr.org/sections/thetwo-way/2017/12/26/573609499/library-of-congress-will-no-longer-archive-every-tweet}}</ref><br />
<br />
== Lists ==<br />
<br />
* [[List of Twitter hashtags by country and territory]]<br />
* [[List of Twitter hashtags by language]]<br />
<br />
== External links ==<br />
* {{URL|https://twitter.com/}}<br />
* {{URL|https://github.com/MartinKBeck/TwitterScraper}}<br />
<br />
{{Navigation box}}<br />
<br />
[[Category:Twitter| ]]<br />
[[Category:Microblogging services]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=File:Duck_Hunt_(World)-0--twitter-5.png&diff=50592File:Duck Hunt (World)-0--twitter-5.png2023-08-27T02:19:29Z<p>Systwi: Mr. Peepers holding the Twitter bird, signifying the (gradual) death of Twitter.</p>
<hr />
<div>== Summary ==<br />
Mr. Peepers holding the Twitter bird, signifying the (gradual) death of Twitter.</div>Systwihttps://wiki.archiveteam.org/index.php?title=Topics_of_Archiving_Interest&diff=50032Topics of Archiving Interest2023-06-27T08:37:17Z<p>Systwi: /* Celebrities/Famous Individuals */ Added 9 entries.</p>
<hr />
<div>This page lists any general topics that may have little to no archiving coverage. The goal is to look further into these topics/related subtopics and further expand our archival scope in those areas. It is, by all means, <i>not</i> all-encompassing.<br />
<br />
The topics here are listed in no particular order.<br />
<br />
== Art/Entertainment ==<br />
* [https://en.wikipedia.org/wiki/Beadwork beadwork]<br />
** [https://web.archive.org/web/20220630043840id_/https://i.redd.it/iqyw4xictzd21.jpg Perler bead art]<br />
* [https://en.wikipedia.org/wiki/Board_game board games]<br />
* [https://en.wikipedia.org/wiki/Cross-stitch cross-stitching]<br />
* [https://en.wikipedia.org/wiki/Magic_(illusion) magic acts]<br />
<br />
== Businesses ==<br />
* [https://en.wikipedia.org/wiki/Broderbund Broderbund]<br />
* [https://en.wikipedia.org/wiki/Cranium,_Inc. Cranium, Inc.]<br />
* [https://en.wikipedia.org/wiki/Studio_Ghibli Studio Ghibli] ([https://en.wikipedia.org/wiki/List_of_Studio_Ghibli_works related])<br />
<br />
== Celebrities/Famous Individuals ==<br />
* [https://en.wikipedia.org/wiki/Lida_Abdul Lida Abdul]<br />
* [https://en.wikipedia.org/wiki/Paula_Abdul Paula Abdul]<br />
* [https://en.wikipedia.org/wiki/Gary_Baseman Gary Baseman]<br />
* [https://en.wikipedia.org/wiki/Reggie_Fils-Aim%C3%A9 Reggie Fils-Aimé (<i>colloq.</i> Reggie)]<br />
* [https://web.archive.org/web/20230627083412/https://bulbapedia.bulbagarden.net/wiki/G%C5%8D_Ichinose Gō Ichinose] (sometimes written as Go Ichinose)<br />
* [https://en.wikipedia.org/wiki/Tappi_Iwase Tappi Iwase]<br />
* [https://en.wikipedia.org/wiki/Janet_Jackson Janet Jackson]<br />
* [https://en.wikipedia.org/wiki/Hajime_Katoki Hajime Katoki]<br />
* [https://en.wikipedia.org/wiki/Hideo_Kojima Hideo Kojima]<br />
* [https://en.wikipedia.org/wiki/Junichi_Masuda Junichi Masuda]<br />
* [https://en.wikipedia.org/wiki/Rosa_Parks Rosa Parks]<br />
* [https://en.wikipedia.org/wiki/Alan_Parsons Alan Parsons]<br />
* [https://en.wikipedia.org/wiki/Yoji_Shinkawa Yoji Shinkawa]<br />
* [https://en.wikipedia.org/wiki/Sylvester_Stallone Sylvester Stallone]<br />
* [https://en.wikipedia.org/wiki/John_Stamos John Stamos]<br />
* [https://en.wikipedia.org/wiki/Barbra_Streisand Barbra Streisand] ([https://en.wikipedia.org/wiki/Streisand_effect related])<br />
* [https://en.wikipedia.org/wiki/Kiefer_Sutherland Kiefer Sutherland]<br />
* [https://en.wikipedia.org/wiki/Risa_Tabata Risa Tabata]<br />
* [https://en.wikipedia.org/wiki/Akira_Toriyama Akira Toriyama]<br />
* [https://en.wikipedia.org/wiki/Toro_y_Moi Toro y Moi (Chaz Bear)]<br />
<br />
== Leisure ==<br />
* [https://en.wikipedia.org/wiki/Sport competitive sports] (casual or championship/professional)<br />
** [https://en.wikipedia.org/wiki/Badminton badminton]<br />
** [https://en.wikipedia.org/wiki/Bowling bowling]<br />
** [https://en.wikipedia.org/wiki/Fishing fishing]<br />
** [https://en.wikipedia.org/wiki/Golf golfing]<br />
** [https://en.wikipedia.org/wiki/Skeet_shooting skeet shooting]<br />
** [https://en.wikipedia.org/wiki/Table_football table football (foosball)]<br />
** [https://en.wikipedia.org/wiki/Tennis tennis]<br />
*** [https://en.wikipedia.org/wiki/Table_tennis table tennis]<br />
* [https://en.wikipedia.org/wiki/Yoga yoga]<br />
** [https://en.wikipedia.org/wiki/Meditation meditation]<br />
<br />
== Multimedia Franchises ==<br />
* [https://en.wikipedia.org/wiki/Adiboo Adibou (Adiboo)]<br />
* [https://en.wikipedia.org/wiki/Animal_Crossing Animal Crossing]<br />
* [https://en.wikipedia.org/wiki/Benjamin_the_Elephant Benjamin Blümchen (Benjamin the Elephant)] <br />
* [https://en.wikipedia.org/wiki/Bibi_Blocksberg Bibi Blocksberg]<br />
* [https://en.wikipedia.org/wiki/Club_Kidsoft Club Kidsoft]<br />
* [https://en.wikipedia.org/wiki/Crayon_Shin-chan Crayon Shin-chan (Shin-chan)]<br />
* [https://en.wikipedia.org/wiki/Doraemon Doraemon]<br />
* [https://en.wikipedia.org/wiki/Felix_the_Cat Felix the Cat]<br />
* [https://en.wikipedia.org/wiki/Harry_Potter Harry Potter]<br />
* [https://en.wikipedia.org/wiki/Keitai_Denj%C5%AB_Telefang Keitai Denjū Telefang (Telefang)]<br />
* [https://en.wikipedia.org/wiki/Kingdom_Hearts Kingdom Hearts]<br />
* [https://en.wikipedia.org/wiki/Living_Books Living Books]<br />
* [https://en.wikipedia.org/wiki/Medabots Medabots (Medarot)]<br />
* [https://en.wikipedia.org/wiki/Pikmin Pikmin]<br />
* [https://en.wikipedia.org/wiki/Pok%C3%A9mon Pokémon (Pocket Monsters)]<br />
* [https://en.wikipedia.org/wiki/SD_Gundam SD Gundam]<br />
* [https://en.wikipedia.org/wiki/The_Smurfs The Smurfs (Les Schtroumpfs; De Smurfen)]<br />
* [https://en.wikipedia.org/wiki/SpongeBob_SquarePants SpongeBob SquarePants (SpongeBob)]<br />
* [https://en.wikipedia.org/wiki/Yo-kai_Watch Yo-kai Watch]<br />
<br />
== Work/Occupations ==<br />
* [https://en.wikipedia.org/wiki/Animal_training animal training]<br />
* [https://en.wikipedia.org/wiki/Cooking cooking (food preparation)]<br />
** [https://en.wikipedia.org/wiki/Candy_making candy making]<br />
* [https://en.wikipedia.org/wiki/Gardening gardening]<br />
* [https://en.wikipedia.org/wiki/Lifeguard lifeguard]<br />
* [https://en.wikipedia.org/wiki/Marine_biology marine biology]<br />
* [https://en.wikipedia.org/wiki/Soldering soldering]<br />
* [https://en.wikipedia.org/wiki/Woodworking woodwork]<br />
** [https://en.wikipedia.org/wiki/Pyrography wood burning]<br />
<br />
<!-- == Miscellaneous == --><br />
<br />
----<br />
<br />
== Finding Information on These Topics ==<br />
There are numerous places and search queries one can use to find websites, videos, software et al. related to these topics. Below are a few places and search query "templates" you can tweak and try yourself:<br />
<br />
* social media ([[Social_network#List_of_social_networks|see also]])<br />
** [[Facebook]] (usually requires login)<br />
** [[Instagram]] (almost always requires login)<br />
** [[Reddit]]<br />
<!-- *** [https://old.reddit.com/r/all/search?q=searchQuery&nsfw=on basic search] --><br />
** [[TikTok]]<br />
** [[Tumblr]]<br />
** [[Twitter]] (login prompt can be bypassed with a [https://web.archive.org/web/20220713041322/https://old.reddit.com/r/uBlockOrigin/comments/rnay51/twitter_log_in_nag_screen/hpxxcb2/ workaround])<br />
*** [https://twitter.com/explore basic search]<br />
*** [https://twitter.com/search-advanced advanced search]<br />
*** hashtag search: <code>https://twitter.com/search?q=%23<b>hashtagQuery</b></code><br />
** [[VKontakte|VK]]<br />
** [[YouTube]] (often blocks content access behind captchas)<br />
<br />
* search engines commonly utilize special syntax for fine-tuning searches, such as:<br />
** [https://en.wikipedia.org/wiki/DuckDuckGo DuckDuckGo]: [https://help.duckduckgo.com/duckduckgo-help-pages/results/syntax/ search syntax guide]<br />
** [[Google]]: [https://support.google.com/websearch/answer/2466433 search syntax guide]<br />
<br />
* search query "templates," such as:<br />
** <i><b>topic</b> blogs</i><br />
** <i><b>topic</b> blogspot</i><br />
** <i><b>topic</b> wordpress</i><br />
** <i><b>topic</b> for beginners</i><br />
** <i>all about <b>topic</b></i><br />
** <i>intro to <b>topic</b></i><br />
** <i><b>topic</b> [[Reddit|subreddits]]</i><br />
** <i><b>topic</b> videos</i><br />
** <i><b>topic</b> [https://en.wikipedia.org/wiki/Convention_(meeting) conventions]</i><br />
** <i><b>topic</b> [[Tumblr|tumblr]] blogs</i><br />
** <i><b>topic</b> tutorials</i></div>Systwihttps://wiki.archiveteam.org/index.php?title=Topics_of_Archiving_Interest&diff=50031Topics of Archiving Interest2023-06-27T08:23:19Z<p>Systwi: /* Leisure */ Added 1 entry.</p>
<hr />
<div>This page lists any general topics that may have little to no archiving coverage. The goal is to look further into these topics/related subtopics and further expand our archival scope in those areas. It is, by all means, <i>not</i> all-encompassing.<br />
<br />
The topics here are listed in no particular order.<br />
<br />
== Art/Entertainment ==<br />
* [https://en.wikipedia.org/wiki/Beadwork beadwork]<br />
** [https://web.archive.org/web/20220630043840id_/https://i.redd.it/iqyw4xictzd21.jpg Perler bead art]<br />
* [https://en.wikipedia.org/wiki/Board_game board games]<br />
* [https://en.wikipedia.org/wiki/Cross-stitch cross-stitching]<br />
* [https://en.wikipedia.org/wiki/Magic_(illusion) magic acts]<br />
<br />
== Businesses ==<br />
* [https://en.wikipedia.org/wiki/Broderbund Broderbund]<br />
* [https://en.wikipedia.org/wiki/Cranium,_Inc. Cranium, Inc.]<br />
* [https://en.wikipedia.org/wiki/Studio_Ghibli Studio Ghibli] ([https://en.wikipedia.org/wiki/List_of_Studio_Ghibli_works related])<br />
<br />
== Celebrities/Famous Individuals ==<br />
* [https://en.wikipedia.org/wiki/Lida_Abdul Lida Abdul]<br />
* [https://en.wikipedia.org/wiki/Gary_Baseman Gary Baseman]<br />
* [https://en.wikipedia.org/wiki/Reggie_Fils-Aim%C3%A9 Reggie Fils-Aimé (<i>colloq.</i> Reggie)]<br />
* [https://en.wikipedia.org/wiki/Rosa_Parks Rosa Parks]<br />
* [https://en.wikipedia.org/wiki/Alan_Parsons Alan Parsons]<br />
* [https://en.wikipedia.org/wiki/Sylvester_Stallone Sylvester Stallone]<br />
* [https://en.wikipedia.org/wiki/John_Stamos John Stamos]<br />
* [https://en.wikipedia.org/wiki/Barbra_Streisand Barbra Streisand] ([https://en.wikipedia.org/wiki/Streisand_effect related])<br />
* [https://en.wikipedia.org/wiki/Kiefer_Sutherland Kiefer Sutherland]<br />
* [https://en.wikipedia.org/wiki/Risa_Tabata Risa Tabata]<br />
* [https://en.wikipedia.org/wiki/Toro_y_Moi Toro y Moi (Chaz Bear)]<br />
<br />
== Leisure ==<br />
* [https://en.wikipedia.org/wiki/Sport competitive sports] (casual or championship/professional)<br />
** [https://en.wikipedia.org/wiki/Badminton badminton]<br />
** [https://en.wikipedia.org/wiki/Bowling bowling]<br />
** [https://en.wikipedia.org/wiki/Fishing fishing]<br />
** [https://en.wikipedia.org/wiki/Golf golfing]<br />
** [https://en.wikipedia.org/wiki/Skeet_shooting skeet shooting]<br />
** [https://en.wikipedia.org/wiki/Table_football table football (foosball)]<br />
** [https://en.wikipedia.org/wiki/Tennis tennis]<br />
*** [https://en.wikipedia.org/wiki/Table_tennis table tennis]<br />
* [https://en.wikipedia.org/wiki/Yoga yoga]<br />
** [https://en.wikipedia.org/wiki/Meditation meditation]<br />
<br />
== Multimedia Franchises ==<br />
* [https://en.wikipedia.org/wiki/Adiboo Adibou (Adiboo)]<br />
* [https://en.wikipedia.org/wiki/Animal_Crossing Animal Crossing]<br />
* [https://en.wikipedia.org/wiki/Benjamin_the_Elephant Benjamin Blümchen (Benjamin the Elephant)] <br />
* [https://en.wikipedia.org/wiki/Bibi_Blocksberg Bibi Blocksberg]<br />
* [https://en.wikipedia.org/wiki/Club_Kidsoft Club Kidsoft]<br />
* [https://en.wikipedia.org/wiki/Crayon_Shin-chan Crayon Shin-chan (Shin-chan)]<br />
* [https://en.wikipedia.org/wiki/Doraemon Doraemon]<br />
* [https://en.wikipedia.org/wiki/Felix_the_Cat Felix the Cat]<br />
* [https://en.wikipedia.org/wiki/Harry_Potter Harry Potter]<br />
* [https://en.wikipedia.org/wiki/Keitai_Denj%C5%AB_Telefang Keitai Denjū Telefang (Telefang)]<br />
* [https://en.wikipedia.org/wiki/Kingdom_Hearts Kingdom Hearts]<br />
* [https://en.wikipedia.org/wiki/Living_Books Living Books]<br />
* [https://en.wikipedia.org/wiki/Medabots Medabots (Medarot)]<br />
* [https://en.wikipedia.org/wiki/Pikmin Pikmin]<br />
* [https://en.wikipedia.org/wiki/Pok%C3%A9mon Pokémon (Pocket Monsters)]<br />
* [https://en.wikipedia.org/wiki/SD_Gundam SD Gundam]<br />
* [https://en.wikipedia.org/wiki/The_Smurfs The Smurfs (Les Schtroumpfs; De Smurfen)]<br />
* [https://en.wikipedia.org/wiki/SpongeBob_SquarePants SpongeBob SquarePants (SpongeBob)]<br />
* [https://en.wikipedia.org/wiki/Yo-kai_Watch Yo-kai Watch]<br />
<br />
== Work/Occupations ==<br />
* [https://en.wikipedia.org/wiki/Animal_training animal training]<br />
* [https://en.wikipedia.org/wiki/Cooking cooking (food preparation)]<br />
** [https://en.wikipedia.org/wiki/Candy_making candy making]<br />
* [https://en.wikipedia.org/wiki/Gardening gardening]<br />
* [https://en.wikipedia.org/wiki/Lifeguard lifeguard]<br />
* [https://en.wikipedia.org/wiki/Marine_biology marine biology]<br />
* [https://en.wikipedia.org/wiki/Soldering soldering]<br />
* [https://en.wikipedia.org/wiki/Woodworking woodwork]<br />
** [https://en.wikipedia.org/wiki/Pyrography wood burning]<br />
<br />
<!-- == Miscellaneous == --><br />
<br />
----<br />
<br />
== Finding Information on These Topics ==<br />
There are numerous places and search queries one can use to find websites, videos, software et al. related to these topics. Below are a few places and search query "templates" you can tweak and try yourself:<br />
<br />
* social media ([[Social_network#List_of_social_networks|see also]])<br />
** [[Facebook]] (usually requires login)<br />
** [[Instagram]] (almost always requires login)<br />
** [[Reddit]]<br />
<!-- *** [https://old.reddit.com/r/all/search?q=searchQuery&nsfw=on basic search] --><br />
** [[TikTok]]<br />
** [[Tumblr]]<br />
** [[Twitter]] (login prompt can be bypassed with a [https://web.archive.org/web/20220713041322/https://old.reddit.com/r/uBlockOrigin/comments/rnay51/twitter_log_in_nag_screen/hpxxcb2/ workaround])<br />
*** [https://twitter.com/explore basic search]<br />
*** [https://twitter.com/search-advanced advanced search]<br />
*** hashtag search: <code>https://twitter.com/search?q=%23<b>hashtagQuery</b></code><br />
** [[VKontakte|VK]]<br />
** [[YouTube]] (often blocks content access behind captchas)<br />
<br />
* search engines commonly utilize special syntax for fine-tuning searches, such as:<br />
** [https://en.wikipedia.org/wiki/DuckDuckGo DuckDuckGo]: [https://help.duckduckgo.com/duckduckgo-help-pages/results/syntax/ search syntax guide]<br />
** [[Google]]: [https://support.google.com/websearch/answer/2466433 search syntax guide]<br />
<br />
* search query "templates," such as:<br />
** <i><b>topic</b> blogs</i><br />
** <i><b>topic</b> blogspot</i><br />
** <i><b>topic</b> wordpress</i><br />
** <i><b>topic</b> for beginners</i><br />
** <i>all about <b>topic</b></i><br />
** <i>intro to <b>topic</b></i><br />
** <i><b>topic</b> [[Reddit|subreddits]]</i><br />
** <i><b>topic</b> videos</i><br />
** <i><b>topic</b> [https://en.wikipedia.org/wiki/Convention_(meeting) conventions]</i><br />
** <i><b>topic</b> [[Tumblr|tumblr]] blogs</i><br />
** <i><b>topic</b> tutorials</i></div>Systwihttps://wiki.archiveteam.org/index.php?title=Topics_of_Archiving_Interest&diff=50030Topics of Archiving Interest2023-06-27T08:20:54Z<p>Systwi: /* Celebrities/Famous Individuals */ Added 1 entry.</p>
<hr />
<div>This page lists any general topics that may have little to no archiving coverage. The goal is to look further into these topics/related subtopics and further expand our archival scope in those areas. It is, by all means, <i>not</i> all-encompassing.<br />
<br />
The topics here are listed in no particular order.<br />
<br />
== Art/Entertainment ==<br />
* [https://en.wikipedia.org/wiki/Beadwork beadwork]<br />
** [https://web.archive.org/web/20220630043840id_/https://i.redd.it/iqyw4xictzd21.jpg Perler bead art]<br />
* [https://en.wikipedia.org/wiki/Board_game board games]<br />
* [https://en.wikipedia.org/wiki/Cross-stitch cross-stitching]<br />
* [https://en.wikipedia.org/wiki/Magic_(illusion) magic acts]<br />
<br />
== Businesses ==<br />
* [https://en.wikipedia.org/wiki/Broderbund Broderbund]<br />
* [https://en.wikipedia.org/wiki/Cranium,_Inc. Cranium, Inc.]<br />
* [https://en.wikipedia.org/wiki/Studio_Ghibli Studio Ghibli] ([https://en.wikipedia.org/wiki/List_of_Studio_Ghibli_works related])<br />
<br />
== Celebrities/Famous Individuals ==<br />
* [https://en.wikipedia.org/wiki/Lida_Abdul Lida Abdul]<br />
* [https://en.wikipedia.org/wiki/Gary_Baseman Gary Baseman]<br />
* [https://en.wikipedia.org/wiki/Reggie_Fils-Aim%C3%A9 Reggie Fils-Aimé (<i>colloq.</i> Reggie)]<br />
* [https://en.wikipedia.org/wiki/Rosa_Parks Rosa Parks]<br />
* [https://en.wikipedia.org/wiki/Alan_Parsons Alan Parsons]<br />
* [https://en.wikipedia.org/wiki/Sylvester_Stallone Sylvester Stallone]<br />
* [https://en.wikipedia.org/wiki/John_Stamos John Stamos]<br />
* [https://en.wikipedia.org/wiki/Barbra_Streisand Barbra Streisand] ([https://en.wikipedia.org/wiki/Streisand_effect related])<br />
* [https://en.wikipedia.org/wiki/Kiefer_Sutherland Kiefer Sutherland]<br />
* [https://en.wikipedia.org/wiki/Risa_Tabata Risa Tabata]<br />
* [https://en.wikipedia.org/wiki/Toro_y_Moi Toro y Moi (Chaz Bear)]<br />
<br />
== Leisure ==<br />
* [https://en.wikipedia.org/wiki/Sport competitive sports] (casual or championship/professional)<br />
** [https://en.wikipedia.org/wiki/Badminton badminton]<br />
** [https://en.wikipedia.org/wiki/Bowling bowling]<br />
** [https://en.wikipedia.org/wiki/Fishing fishing]<br />
** [https://en.wikipedia.org/wiki/Golf golfing]<br />
** [https://en.wikipedia.org/wiki/Table_football table football (foosball)]<br />
** [https://en.wikipedia.org/wiki/Tennis tennis]<br />
*** [https://en.wikipedia.org/wiki/Table_tennis table tennis]<br />
* [https://en.wikipedia.org/wiki/Yoga yoga]<br />
** [https://en.wikipedia.org/wiki/Meditation meditation]<br />
<br />
== Multimedia Franchises ==<br />
* [https://en.wikipedia.org/wiki/Adiboo Adibou (Adiboo)]<br />
* [https://en.wikipedia.org/wiki/Animal_Crossing Animal Crossing]<br />
* [https://en.wikipedia.org/wiki/Benjamin_the_Elephant Benjamin Blümchen (Benjamin the Elephant)] <br />
* [https://en.wikipedia.org/wiki/Bibi_Blocksberg Bibi Blocksberg]<br />
* [https://en.wikipedia.org/wiki/Club_Kidsoft Club Kidsoft]<br />
* [https://en.wikipedia.org/wiki/Crayon_Shin-chan Crayon Shin-chan (Shin-chan)]<br />
* [https://en.wikipedia.org/wiki/Doraemon Doraemon]<br />
* [https://en.wikipedia.org/wiki/Felix_the_Cat Felix the Cat]<br />
* [https://en.wikipedia.org/wiki/Harry_Potter Harry Potter]<br />
* [https://en.wikipedia.org/wiki/Keitai_Denj%C5%AB_Telefang Keitai Denjū Telefang (Telefang)]<br />
* [https://en.wikipedia.org/wiki/Kingdom_Hearts Kingdom Hearts]<br />
* [https://en.wikipedia.org/wiki/Living_Books Living Books]<br />
* [https://en.wikipedia.org/wiki/Medabots Medabots (Medarot)]<br />
* [https://en.wikipedia.org/wiki/Pikmin Pikmin]<br />
* [https://en.wikipedia.org/wiki/Pok%C3%A9mon Pokémon (Pocket Monsters)]<br />
* [https://en.wikipedia.org/wiki/SD_Gundam SD Gundam]<br />
* [https://en.wikipedia.org/wiki/The_Smurfs The Smurfs (Les Schtroumpfs; De Smurfen)]<br />
* [https://en.wikipedia.org/wiki/SpongeBob_SquarePants SpongeBob SquarePants (SpongeBob)]<br />
* [https://en.wikipedia.org/wiki/Yo-kai_Watch Yo-kai Watch]<br />
<br />
== Work/Occupations ==<br />
* [https://en.wikipedia.org/wiki/Animal_training animal training]<br />
* [https://en.wikipedia.org/wiki/Cooking cooking (food preparation)]<br />
** [https://en.wikipedia.org/wiki/Candy_making candy making]<br />
* [https://en.wikipedia.org/wiki/Gardening gardening]<br />
* [https://en.wikipedia.org/wiki/Lifeguard lifeguard]<br />
* [https://en.wikipedia.org/wiki/Marine_biology marine biology]<br />
* [https://en.wikipedia.org/wiki/Soldering soldering]<br />
* [https://en.wikipedia.org/wiki/Woodworking woodwork]<br />
** [https://en.wikipedia.org/wiki/Pyrography wood burning]<br />
<br />
<!-- == Miscellaneous == --><br />
<br />
----<br />
<br />
== Finding Information on These Topics ==<br />
There are numerous places and search queries one can use to find websites, videos, software et al. related to these topics. Below are a few places and search query "templates" you can tweak and try yourself:<br />
<br />
* social media ([[Social_network#List_of_social_networks|see also]])<br />
** [[Facebook]] (usually requires login)<br />
** [[Instagram]] (almost always requires login)<br />
** [[Reddit]]<br />
<!-- *** [https://old.reddit.com/r/all/search?q=searchQuery&nsfw=on basic search] --><br />
** [[TikTok]]<br />
** [[Tumblr]]<br />
** [[Twitter]] (login prompt can be bypassed with a [https://web.archive.org/web/20220713041322/https://old.reddit.com/r/uBlockOrigin/comments/rnay51/twitter_log_in_nag_screen/hpxxcb2/ workaround])<br />
*** [https://twitter.com/explore basic search]<br />
*** [https://twitter.com/search-advanced advanced search]<br />
*** hashtag search: <code>https://twitter.com/search?q=%23<b>hashtagQuery</b></code><br />
** [[VKontakte|VK]]<br />
** [[YouTube]] (often blocks content access behind captchas)<br />
<br />
* search engines commonly utilize special syntax for fine-tuning searches, such as:<br />
** [https://en.wikipedia.org/wiki/DuckDuckGo DuckDuckGo]: [https://help.duckduckgo.com/duckduckgo-help-pages/results/syntax/ search syntax guide]<br />
** [[Google]]: [https://support.google.com/websearch/answer/2466433 search syntax guide]<br />
<br />
* search query "templates," such as:<br />
** <i><b>topic</b> blogs</i><br />
** <i><b>topic</b> blogspot</i><br />
** <i><b>topic</b> wordpress</i><br />
** <i><b>topic</b> for beginners</i><br />
** <i>all about <b>topic</b></i><br />
** <i>intro to <b>topic</b></i><br />
** <i><b>topic</b> [[Reddit|subreddits]]</i><br />
** <i><b>topic</b> videos</i><br />
** <i><b>topic</b> [https://en.wikipedia.org/wiki/Convention_(meeting) conventions]</i><br />
** <i><b>topic</b> [[Tumblr|tumblr]] blogs</i><br />
** <i><b>topic</b> tutorials</i></div>Systwihttps://wiki.archiveteam.org/index.php?title=Topics_of_Archiving_Interest&diff=50029Topics of Archiving Interest2023-06-27T08:19:06Z<p>Systwi: /* Work/Occupations */ Added 1 entry.</p>
<hr />
<div>This page lists any general topics that may have little to no archiving coverage. The goal is to look further into these topics/related subtopics and further expand our archival scope in those areas. It is, by all means, <i>not</i> all-encompassing.<br />
<br />
The topics here are listed in no particular order.<br />
<br />
== Art/Entertainment ==<br />
* [https://en.wikipedia.org/wiki/Beadwork beadwork]<br />
** [https://web.archive.org/web/20220630043840id_/https://i.redd.it/iqyw4xictzd21.jpg Perler bead art]<br />
* [https://en.wikipedia.org/wiki/Board_game board games]<br />
* [https://en.wikipedia.org/wiki/Cross-stitch cross-stitching]<br />
* [https://en.wikipedia.org/wiki/Magic_(illusion) magic acts]<br />
<br />
== Businesses ==<br />
* [https://en.wikipedia.org/wiki/Broderbund Broderbund]<br />
* [https://en.wikipedia.org/wiki/Cranium,_Inc. Cranium, Inc.]<br />
* [https://en.wikipedia.org/wiki/Studio_Ghibli Studio Ghibli] ([https://en.wikipedia.org/wiki/List_of_Studio_Ghibli_works related])<br />
<br />
== Celebrities/Famous Individuals ==<br />
* [https://en.wikipedia.org/wiki/Lida_Abdul Lida Abdul]<br />
* [https://en.wikipedia.org/wiki/Gary_Baseman Gary Baseman]<br />
* [https://en.wikipedia.org/wiki/Reggie_Fils-Aim%C3%A9 Reggie Fils-Aimé (<i>colloq.</i> Reggie)]<br />
* [https://en.wikipedia.org/wiki/Rosa_Parks Rosa Parks]<br />
* [https://en.wikipedia.org/wiki/Sylvester_Stallone Sylvester Stallone]<br />
* [https://en.wikipedia.org/wiki/John_Stamos John Stamos]<br />
* [https://en.wikipedia.org/wiki/Barbra_Streisand Barbra Streisand] ([https://en.wikipedia.org/wiki/Streisand_effect related])<br />
* [https://en.wikipedia.org/wiki/Kiefer_Sutherland Kiefer Sutherland]<br />
* [https://en.wikipedia.org/wiki/Risa_Tabata Risa Tabata]<br />
* [https://en.wikipedia.org/wiki/Toro_y_Moi Toro y Moi (Chaz Bear)]<br />
<br />
== Leisure ==<br />
* [https://en.wikipedia.org/wiki/Sport competitive sports] (casual or championship/professional)<br />
** [https://en.wikipedia.org/wiki/Badminton badminton]<br />
** [https://en.wikipedia.org/wiki/Bowling bowling]<br />
** [https://en.wikipedia.org/wiki/Fishing fishing]<br />
** [https://en.wikipedia.org/wiki/Golf golfing]<br />
** [https://en.wikipedia.org/wiki/Table_football table football (foosball)]<br />
** [https://en.wikipedia.org/wiki/Tennis tennis]<br />
*** [https://en.wikipedia.org/wiki/Table_tennis table tennis]<br />
* [https://en.wikipedia.org/wiki/Yoga yoga]<br />
** [https://en.wikipedia.org/wiki/Meditation meditation]<br />
<br />
== Multimedia Franchises ==<br />
* [https://en.wikipedia.org/wiki/Adiboo Adibou (Adiboo)]<br />
* [https://en.wikipedia.org/wiki/Animal_Crossing Animal Crossing]<br />
* [https://en.wikipedia.org/wiki/Benjamin_the_Elephant Benjamin Blümchen (Benjamin the Elephant)] <br />
* [https://en.wikipedia.org/wiki/Bibi_Blocksberg Bibi Blocksberg]<br />
* [https://en.wikipedia.org/wiki/Club_Kidsoft Club Kidsoft]<br />
* [https://en.wikipedia.org/wiki/Crayon_Shin-chan Crayon Shin-chan (Shin-chan)]<br />
* [https://en.wikipedia.org/wiki/Doraemon Doraemon]<br />
* [https://en.wikipedia.org/wiki/Felix_the_Cat Felix the Cat]<br />
* [https://en.wikipedia.org/wiki/Harry_Potter Harry Potter]<br />
* [https://en.wikipedia.org/wiki/Keitai_Denj%C5%AB_Telefang Keitai Denjū Telefang (Telefang)]<br />
* [https://en.wikipedia.org/wiki/Kingdom_Hearts Kingdom Hearts]<br />
* [https://en.wikipedia.org/wiki/Living_Books Living Books]<br />
* [https://en.wikipedia.org/wiki/Medabots Medabots (Medarot)]<br />
* [https://en.wikipedia.org/wiki/Pikmin Pikmin]<br />
* [https://en.wikipedia.org/wiki/Pok%C3%A9mon Pokémon (Pocket Monsters)]<br />
* [https://en.wikipedia.org/wiki/SD_Gundam SD Gundam]<br />
* [https://en.wikipedia.org/wiki/The_Smurfs The Smurfs (Les Schtroumpfs; De Smurfen)]<br />
* [https://en.wikipedia.org/wiki/SpongeBob_SquarePants SpongeBob SquarePants (SpongeBob)]<br />
* [https://en.wikipedia.org/wiki/Yo-kai_Watch Yo-kai Watch]<br />
<br />
== Work/Occupations ==<br />
* [https://en.wikipedia.org/wiki/Animal_training animal training]<br />
* [https://en.wikipedia.org/wiki/Cooking cooking (food preparation)]<br />
** [https://en.wikipedia.org/wiki/Candy_making candy making]<br />
* [https://en.wikipedia.org/wiki/Gardening gardening]<br />
* [https://en.wikipedia.org/wiki/Lifeguard lifeguard]<br />
* [https://en.wikipedia.org/wiki/Marine_biology marine biology]<br />
* [https://en.wikipedia.org/wiki/Soldering soldering]<br />
* [https://en.wikipedia.org/wiki/Woodworking woodwork]<br />
** [https://en.wikipedia.org/wiki/Pyrography wood burning]<br />
<br />
<!-- == Miscellaneous == --><br />
<br />
----<br />
<br />
== Finding Information on These Topics ==<br />
There are numerous places and search queries one can use to find websites, videos, software et al. related to these topics. Below are a few places and search query "templates" you can tweak and try yourself:<br />
<br />
* social media ([[Social_network#List_of_social_networks|see also]])<br />
** [[Facebook]] (usually requires login)<br />
** [[Instagram]] (almost always requires login)<br />
** [[Reddit]]<br />
<!-- *** [https://old.reddit.com/r/all/search?q=searchQuery&nsfw=on basic search] --><br />
** [[TikTok]]<br />
** [[Tumblr]]<br />
** [[Twitter]] (login prompt can be bypassed with a [https://web.archive.org/web/20220713041322/https://old.reddit.com/r/uBlockOrigin/comments/rnay51/twitter_log_in_nag_screen/hpxxcb2/ workaround])<br />
*** [https://twitter.com/explore basic search]<br />
*** [https://twitter.com/search-advanced advanced search]<br />
*** hashtag search: <code>https://twitter.com/search?q=%23<b>hashtagQuery</b></code><br />
** [[VKontakte|VK]]<br />
** [[YouTube]] (often blocks content access behind captchas)<br />
<br />
* search engines commonly utilize special syntax for fine-tuning searches, such as:<br />
** [https://en.wikipedia.org/wiki/DuckDuckGo DuckDuckGo]: [https://help.duckduckgo.com/duckduckgo-help-pages/results/syntax/ search syntax guide]<br />
** [[Google]]: [https://support.google.com/websearch/answer/2466433 search syntax guide]<br />
<br />
* search query "templates," such as:<br />
** <i><b>topic</b> blogs</i><br />
** <i><b>topic</b> blogspot</i><br />
** <i><b>topic</b> wordpress</i><br />
** <i><b>topic</b> for beginners</i><br />
** <i>all about <b>topic</b></i><br />
** <i>intro to <b>topic</b></i><br />
** <i><b>topic</b> [[Reddit|subreddits]]</i><br />
** <i><b>topic</b> videos</i><br />
** <i><b>topic</b> [https://en.wikipedia.org/wiki/Convention_(meeting) conventions]</i><br />
** <i><b>topic</b> [[Tumblr|tumblr]] blogs</i><br />
** <i><b>topic</b> tutorials</i></div>Systwihttps://wiki.archiveteam.org/index.php?title=Topics_of_Archiving_Interest&diff=50028Topics of Archiving Interest2023-06-27T08:16:57Z<p>Systwi: /* Multimedia Franchises */ Added 1 entry.</p>
<hr />
<div>This page lists any general topics that may have little to no archiving coverage. The goal is to look further into these topics/related subtopics and further expand our archival scope in those areas. It is, by all means, <i>not</i> all-encompassing.<br />
<br />
The topics here are listed in no particular order.<br />
<br />
== Art/Entertainment ==<br />
* [https://en.wikipedia.org/wiki/Beadwork beadwork]<br />
** [https://web.archive.org/web/20220630043840id_/https://i.redd.it/iqyw4xictzd21.jpg Perler bead art]<br />
* [https://en.wikipedia.org/wiki/Board_game board games]<br />
* [https://en.wikipedia.org/wiki/Cross-stitch cross-stitching]<br />
* [https://en.wikipedia.org/wiki/Magic_(illusion) magic acts]<br />
<br />
== Businesses ==<br />
* [https://en.wikipedia.org/wiki/Broderbund Broderbund]<br />
* [https://en.wikipedia.org/wiki/Cranium,_Inc. Cranium, Inc.]<br />
* [https://en.wikipedia.org/wiki/Studio_Ghibli Studio Ghibli] ([https://en.wikipedia.org/wiki/List_of_Studio_Ghibli_works related])<br />
<br />
== Celebrities/Famous Individuals ==<br />
* [https://en.wikipedia.org/wiki/Lida_Abdul Lida Abdul]<br />
* [https://en.wikipedia.org/wiki/Gary_Baseman Gary Baseman]<br />
* [https://en.wikipedia.org/wiki/Reggie_Fils-Aim%C3%A9 Reggie Fils-Aimé (<i>colloq.</i> Reggie)]<br />
* [https://en.wikipedia.org/wiki/Rosa_Parks Rosa Parks]<br />
* [https://en.wikipedia.org/wiki/Sylvester_Stallone Sylvester Stallone]<br />
* [https://en.wikipedia.org/wiki/John_Stamos John Stamos]<br />
* [https://en.wikipedia.org/wiki/Barbra_Streisand Barbra Streisand] ([https://en.wikipedia.org/wiki/Streisand_effect related])<br />
* [https://en.wikipedia.org/wiki/Kiefer_Sutherland Kiefer Sutherland]<br />
* [https://en.wikipedia.org/wiki/Risa_Tabata Risa Tabata]<br />
* [https://en.wikipedia.org/wiki/Toro_y_Moi Toro y Moi (Chaz Bear)]<br />
<br />
== Leisure ==<br />
* [https://en.wikipedia.org/wiki/Sport competitive sports] (casual or championship/professional)<br />
** [https://en.wikipedia.org/wiki/Badminton badminton]<br />
** [https://en.wikipedia.org/wiki/Bowling bowling]<br />
** [https://en.wikipedia.org/wiki/Fishing fishing]<br />
** [https://en.wikipedia.org/wiki/Golf golfing]<br />
** [https://en.wikipedia.org/wiki/Table_football table football (foosball)]<br />
** [https://en.wikipedia.org/wiki/Tennis tennis]<br />
*** [https://en.wikipedia.org/wiki/Table_tennis table tennis]<br />
* [https://en.wikipedia.org/wiki/Yoga yoga]<br />
** [https://en.wikipedia.org/wiki/Meditation meditation]<br />
<br />
== Multimedia Franchises ==<br />
* [https://en.wikipedia.org/wiki/Adiboo Adibou (Adiboo)]<br />
* [https://en.wikipedia.org/wiki/Animal_Crossing Animal Crossing]<br />
* [https://en.wikipedia.org/wiki/Benjamin_the_Elephant Benjamin Blümchen (Benjamin the Elephant)] <br />
* [https://en.wikipedia.org/wiki/Bibi_Blocksberg Bibi Blocksberg]<br />
* [https://en.wikipedia.org/wiki/Club_Kidsoft Club Kidsoft]<br />
* [https://en.wikipedia.org/wiki/Crayon_Shin-chan Crayon Shin-chan (Shin-chan)]<br />
* [https://en.wikipedia.org/wiki/Doraemon Doraemon]<br />
* [https://en.wikipedia.org/wiki/Felix_the_Cat Felix the Cat]<br />
* [https://en.wikipedia.org/wiki/Harry_Potter Harry Potter]<br />
* [https://en.wikipedia.org/wiki/Keitai_Denj%C5%AB_Telefang Keitai Denjū Telefang (Telefang)]<br />
* [https://en.wikipedia.org/wiki/Kingdom_Hearts Kingdom Hearts]<br />
* [https://en.wikipedia.org/wiki/Living_Books Living Books]<br />
* [https://en.wikipedia.org/wiki/Medabots Medabots (Medarot)]<br />
* [https://en.wikipedia.org/wiki/Pikmin Pikmin]<br />
* [https://en.wikipedia.org/wiki/Pok%C3%A9mon Pokémon (Pocket Monsters)]<br />
* [https://en.wikipedia.org/wiki/SD_Gundam SD Gundam]<br />
* [https://en.wikipedia.org/wiki/The_Smurfs The Smurfs (Les Schtroumpfs; De Smurfen)]<br />
* [https://en.wikipedia.org/wiki/SpongeBob_SquarePants SpongeBob SquarePants (SpongeBob)]<br />
* [https://en.wikipedia.org/wiki/Yo-kai_Watch Yo-kai Watch]<br />
<br />
== Work/Occupations ==<br />
* [https://en.wikipedia.org/wiki/Animal_training animal training]<br />
* [https://en.wikipedia.org/wiki/Cooking cooking (food preparation)]<br />
** [https://en.wikipedia.org/wiki/Candy_making candy making]<br />
* [https://en.wikipedia.org/wiki/Gardening gardening]<br />
* [https://en.wikipedia.org/wiki/Lifeguard lifeguard]<br />
* [https://en.wikipedia.org/wiki/Marine_biology marine biology]<br />
* [https://en.wikipedia.org/wiki/Woodworking woodwork]<br />
** [https://en.wikipedia.org/wiki/Pyrography wood burning]<br />
<br />
<!-- == Miscellaneous == --><br />
<br />
----<br />
<br />
== Finding Information on These Topics ==<br />
There are numerous places and search queries one can use to find websites, videos, software et al. related to these topics. Below are a few places and search query "templates" you can tweak and try yourself:<br />
<br />
* social media ([[Social_network#List_of_social_networks|see also]])<br />
** [[Facebook]] (usually requires login)<br />
** [[Instagram]] (almost always requires login)<br />
** [[Reddit]]<br />
<!-- *** [https://old.reddit.com/r/all/search?q=searchQuery&nsfw=on basic search] --><br />
** [[TikTok]]<br />
** [[Tumblr]]<br />
** [[Twitter]] (login prompt can be bypassed with a [https://web.archive.org/web/20220713041322/https://old.reddit.com/r/uBlockOrigin/comments/rnay51/twitter_log_in_nag_screen/hpxxcb2/ workaround])<br />
*** [https://twitter.com/explore basic search]<br />
*** [https://twitter.com/search-advanced advanced search]<br />
*** hashtag search: <code>https://twitter.com/search?q=%23<b>hashtagQuery</b></code><br />
** [[VKontakte|VK]]<br />
** [[YouTube]] (often blocks content access behind captchas)<br />
<br />
* search engines commonly utilize special syntax for fine-tuning searches, such as:<br />
** [https://en.wikipedia.org/wiki/DuckDuckGo DuckDuckGo]: [https://help.duckduckgo.com/duckduckgo-help-pages/results/syntax/ search syntax guide]<br />
** [[Google]]: [https://support.google.com/websearch/answer/2466433 search syntax guide]<br />
<br />
* search query "templates," such as:<br />
** <i><b>topic</b> blogs</i><br />
** <i><b>topic</b> blogspot</i><br />
** <i><b>topic</b> wordpress</i><br />
** <i><b>topic</b> for beginners</i><br />
** <i>all about <b>topic</b></i><br />
** <i>intro to <b>topic</b></i><br />
** <i><b>topic</b> [[Reddit|subreddits]]</i><br />
** <i><b>topic</b> videos</i><br />
** <i><b>topic</b> [https://en.wikipedia.org/wiki/Convention_(meeting) conventions]</i><br />
** <i><b>topic</b> [[Tumblr|tumblr]] blogs</i><br />
** <i><b>topic</b> tutorials</i></div>Systwihttps://wiki.archiveteam.org/index.php?title=Topics_of_Archiving_Interest&diff=50027Topics of Archiving Interest2023-06-27T08:16:21Z<p>Systwi: /* Multimedia Franchises */</p>
<hr />
<div>This page lists any general topics that may have little to no archiving coverage. The goal is to look further into these topics/related subtopics and further expand our archival scope in those areas. It is, by all means, <i>not</i> all-encompassing.<br />
<br />
The topics here are listed in no particular order.<br />
<br />
== Art/Entertainment ==<br />
* [https://en.wikipedia.org/wiki/Beadwork beadwork]<br />
** [https://web.archive.org/web/20220630043840id_/https://i.redd.it/iqyw4xictzd21.jpg Perler bead art]<br />
* [https://en.wikipedia.org/wiki/Board_game board games]<br />
* [https://en.wikipedia.org/wiki/Cross-stitch cross-stitching]<br />
* [https://en.wikipedia.org/wiki/Magic_(illusion) magic acts]<br />
<br />
== Businesses ==<br />
* [https://en.wikipedia.org/wiki/Broderbund Broderbund]<br />
* [https://en.wikipedia.org/wiki/Cranium,_Inc. Cranium, Inc.]<br />
* [https://en.wikipedia.org/wiki/Studio_Ghibli Studio Ghibli] ([https://en.wikipedia.org/wiki/List_of_Studio_Ghibli_works related])<br />
<br />
== Celebrities/Famous Individuals ==<br />
* [https://en.wikipedia.org/wiki/Lida_Abdul Lida Abdul]<br />
* [https://en.wikipedia.org/wiki/Gary_Baseman Gary Baseman]<br />
* [https://en.wikipedia.org/wiki/Reggie_Fils-Aim%C3%A9 Reggie Fils-Aimé (<i>colloq.</i> Reggie)]<br />
* [https://en.wikipedia.org/wiki/Rosa_Parks Rosa Parks]<br />
* [https://en.wikipedia.org/wiki/Sylvester_Stallone Sylvester Stallone]<br />
* [https://en.wikipedia.org/wiki/John_Stamos John Stamos]<br />
* [https://en.wikipedia.org/wiki/Barbra_Streisand Barbra Streisand] ([https://en.wikipedia.org/wiki/Streisand_effect related])<br />
* [https://en.wikipedia.org/wiki/Kiefer_Sutherland Kiefer Sutherland]<br />
* [https://en.wikipedia.org/wiki/Risa_Tabata Risa Tabata]<br />
* [https://en.wikipedia.org/wiki/Toro_y_Moi Toro y Moi (Chaz Bear)]<br />
<br />
== Leisure ==<br />
* [https://en.wikipedia.org/wiki/Sport competitive sports] (casual or championship/professional)<br />
** [https://en.wikipedia.org/wiki/Badminton badminton]<br />
** [https://en.wikipedia.org/wiki/Bowling bowling]<br />
** [https://en.wikipedia.org/wiki/Fishing fishing]<br />
** [https://en.wikipedia.org/wiki/Golf golfing]<br />
** [https://en.wikipedia.org/wiki/Table_football table football (foosball)]<br />
** [https://en.wikipedia.org/wiki/Tennis tennis]<br />
*** [https://en.wikipedia.org/wiki/Table_tennis table tennis]<br />
* [https://en.wikipedia.org/wiki/Yoga yoga]<br />
** [https://en.wikipedia.org/wiki/Meditation meditation]<br />
<br />
== Multimedia Franchises ==<br />
* [https://en.wikipedia.org/wiki/Adiboo Adibou (Adiboo)]<br />
* [https://en.wikipedia.org/wiki/Animal_Crossing Animal Crossing]<br />
* [https://en.wikipedia.org/wiki/Benjamin_the_Elephant Benjamin Blümchen (Benjamin the Elephant)] <br />
* [https://en.wikipedia.org/wiki/Bibi_Blocksberg Bibi Blocksberg]<br />
* [https://en.wikipedia.org/wiki/Club_Kidsoft Club Kidsoft]<br />
* [https://en.wikipedia.org/wiki/Crayon_Shin-chan Crayon Shin-chan (Shin-chan)]<br />
* [https://en.wikipedia.org/wiki/Doraemon Doraemon]<br />
* [https://en.wikipedia.org/wiki/Felix_the_Cat Felix the Cat]<br />
* [https://en.wikipedia.org/wiki/Harry_Potter Harry Potter]<br />
* [https://en.wikipedia.org/wiki/Keitai_Denj%C5%AB_Telefang Keitai Denjū Telefang (Telefang)]<br />
* [https://en.wikipedia.org/wiki/Kingdom_Hearts Kingdom Hearts]<br />
* [https://en.wikipedia.org/wiki/Living_Books Living Books]<br />
* [https://en.wikipedia.org/wiki/Medabots Medabots (Medarot)]<br />
* [https://en.wikipedia.org/wiki/Pikmin Pikmin]<br />
* [https://en.wikipedia.org/wiki/Pok%C3%A9mon Pokémon (Pocket Monsters)]<br />
* [https://en.wikipedia.org/wiki/SD_Gundam]<br />
* [https://en.wikipedia.org/wiki/The_Smurfs The Smurfs (Les Schtroumpfs; De Smurfen)]<br />
* [https://en.wikipedia.org/wiki/SpongeBob_SquarePants SpongeBob SquarePants (SpongeBob)]<br />
* [https://en.wikipedia.org/wiki/Yo-kai_Watch Yo-kai Watch]<br />
<br />
== Work/Occupations ==<br />
* [https://en.wikipedia.org/wiki/Animal_training animal training]<br />
* [https://en.wikipedia.org/wiki/Cooking cooking (food preparation)]<br />
** [https://en.wikipedia.org/wiki/Candy_making candy making]<br />
* [https://en.wikipedia.org/wiki/Gardening gardening]<br />
* [https://en.wikipedia.org/wiki/Lifeguard lifeguard]<br />
* [https://en.wikipedia.org/wiki/Marine_biology marine biology]<br />
* [https://en.wikipedia.org/wiki/Woodworking woodwork]<br />
** [https://en.wikipedia.org/wiki/Pyrography wood burning]<br />
<br />
<!-- == Miscellaneous == --><br />
<br />
----<br />
<br />
== Finding Information on These Topics ==<br />
There are numerous places and search queries one can use to find websites, videos, software et al. related to these topics. Below are a few places and search query "templates" you can tweak and try yourself:<br />
<br />
* social media ([[Social_network#List_of_social_networks|see also]])<br />
** [[Facebook]] (usually requires login)<br />
** [[Instagram]] (almost always requires login)<br />
** [[Reddit]]<br />
<!-- *** [https://old.reddit.com/r/all/search?q=searchQuery&nsfw=on basic search] --><br />
** [[TikTok]]<br />
** [[Tumblr]]<br />
** [[Twitter]] (login prompt can be bypassed with a [https://web.archive.org/web/20220713041322/https://old.reddit.com/r/uBlockOrigin/comments/rnay51/twitter_log_in_nag_screen/hpxxcb2/ workaround])<br />
*** [https://twitter.com/explore basic search]<br />
*** [https://twitter.com/search-advanced advanced search]<br />
*** hashtag search: <code>https://twitter.com/search?q=%23<b>hashtagQuery</b></code><br />
** [[VKontakte|VK]]<br />
** [[YouTube]] (often blocks content access behind captchas)<br />
<br />
* search engines commonly utilize special syntax for fine-tuning searches, such as:<br />
** [https://en.wikipedia.org/wiki/DuckDuckGo DuckDuckGo]: [https://help.duckduckgo.com/duckduckgo-help-pages/results/syntax/ search syntax guide]<br />
** [[Google]]: [https://support.google.com/websearch/answer/2466433 search syntax guide]<br />
<br />
* search query "templates," such as:<br />
** <i><b>topic</b> blogs</i><br />
** <i><b>topic</b> blogspot</i><br />
** <i><b>topic</b> wordpress</i><br />
** <i><b>topic</b> for beginners</i><br />
** <i>all about <b>topic</b></i><br />
** <i>intro to <b>topic</b></i><br />
** <i><b>topic</b> [[Reddit|subreddits]]</i><br />
** <i><b>topic</b> videos</i><br />
** <i><b>topic</b> [https://en.wikipedia.org/wiki/Convention_(meeting) conventions]</i><br />
** <i><b>topic</b> [[Tumblr|tumblr]] blogs</i><br />
** <i><b>topic</b> tutorials</i></div>Systwihttps://wiki.archiveteam.org/index.php?title=Tjournal&diff=50026Tjournal2023-06-27T08:06:01Z<p>Systwi: Corrected redirect target from "Tjournal" to "TJ."</p>
<hr />
<div>#REDIRECT [[TJ]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=Tjournal&diff=50025Tjournal2023-06-27T08:04:32Z<p>Systwi: Created an article redirecting "Tjournal" to "TJournal."</p>
<hr />
<div>#REDIRECT [[TJournal]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=V-live&diff=49423V-live2023-01-31T07:38:37Z<p>Systwi: Created an article redirecting "V-live" to "V Live."</p>
<hr />
<div>#REDIRECT [[V_Live]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=Vlive&diff=49227Vlive2022-12-10T17:27:54Z<p>Systwi: Created an article redirecting "Vlive" to "V Live."</p>
<hr />
<div>#REDIRECT [[V_Live]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=Tiktok&diff=49226Tiktok2022-12-10T17:26:13Z<p>Systwi: Created an article redirecting "Tiktok" to "TikTok."</p>
<hr />
<div>#REDIRECT [[TikTok]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=Douyin&diff=49225Douyin2022-12-10T17:25:25Z<p>Systwi: Created an article redirecting "Douyin" to "TikTok."</p>
<hr />
<div>#REDIRECT [[TikTok]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=Topics_of_Archiving_Interest&diff=49224Topics of Archiving Interest2022-12-10T17:22:35Z<p>Systwi: /* Finding Information on These Topics */ Added TikTok.</p>
<hr />
<div>This page lists any general topics that may have little to no archiving coverage. The goal is to look further into these topics/related subtopics and further expand our archival scope in those areas. It is, by all means, <i>not</i> all-encompassing.<br />
<br />
The topics here are listed in no particular order.<br />
<br />
== Art/Entertainment ==<br />
* [https://en.wikipedia.org/wiki/Beadwork beadwork]<br />
** [https://web.archive.org/web/20220630043840id_/https://i.redd.it/iqyw4xictzd21.jpg Perler bead art]<br />
* [https://en.wikipedia.org/wiki/Board_game board games]<br />
* [https://en.wikipedia.org/wiki/Cross-stitch cross-stitching]<br />
* [https://en.wikipedia.org/wiki/Magic_(illusion) magic acts]<br />
<br />
== Businesses ==<br />
* [https://en.wikipedia.org/wiki/Broderbund Broderbund]<br />
* [https://en.wikipedia.org/wiki/Cranium,_Inc. Cranium, Inc.]<br />
* [https://en.wikipedia.org/wiki/Studio_Ghibli Studio Ghibli] ([https://en.wikipedia.org/wiki/List_of_Studio_Ghibli_works related])<br />
<br />
== Celebrities/Famous Individuals ==<br />
* [https://en.wikipedia.org/wiki/Lida_Abdul Lida Abdul]<br />
* [https://en.wikipedia.org/wiki/Gary_Baseman Gary Baseman]<br />
* [https://en.wikipedia.org/wiki/Reggie_Fils-Aim%C3%A9 Reggie Fils-Aimé (<i>colloq.</i> Reggie)]<br />
* [https://en.wikipedia.org/wiki/Rosa_Parks Rosa Parks]<br />
* [https://en.wikipedia.org/wiki/Sylvester_Stallone Sylvester Stallone]<br />
* [https://en.wikipedia.org/wiki/John_Stamos John Stamos]<br />
* [https://en.wikipedia.org/wiki/Barbra_Streisand Barbra Streisand] ([https://en.wikipedia.org/wiki/Streisand_effect related])<br />
* [https://en.wikipedia.org/wiki/Kiefer_Sutherland Kiefer Sutherland]<br />
* [https://en.wikipedia.org/wiki/Risa_Tabata Risa Tabata]<br />
* [https://en.wikipedia.org/wiki/Toro_y_Moi Toro y Moi (Chaz Bear)]<br />
<br />
== Leisure ==<br />
* [https://en.wikipedia.org/wiki/Sport competitive sports] (casual or championship/professional)<br />
** [https://en.wikipedia.org/wiki/Badminton badminton]<br />
** [https://en.wikipedia.org/wiki/Bowling bowling]<br />
** [https://en.wikipedia.org/wiki/Fishing fishing]<br />
** [https://en.wikipedia.org/wiki/Golf golfing]<br />
** [https://en.wikipedia.org/wiki/Table_football table football (foosball)]<br />
** [https://en.wikipedia.org/wiki/Tennis tennis]<br />
*** [https://en.wikipedia.org/wiki/Table_tennis table tennis]<br />
* [https://en.wikipedia.org/wiki/Yoga yoga]<br />
** [https://en.wikipedia.org/wiki/Meditation meditation]<br />
<br />
== Multimedia Franchises ==<br />
* [https://en.wikipedia.org/wiki/Adiboo Adibou (Adiboo)]<br />
* [https://en.wikipedia.org/wiki/Animal_Crossing Animal Crossing]<br />
* [https://en.wikipedia.org/wiki/Benjamin_the_Elephant Benjamin Blümchen (Benjamin the Elephant)] <br />
* [https://en.wikipedia.org/wiki/Bibi_Blocksberg Bibi Blocksberg]<br />
* [https://en.wikipedia.org/wiki/Club_Kidsoft Club Kidsoft]<br />
* [https://en.wikipedia.org/wiki/Crayon_Shin-chan Crayon Shin-chan (Shin-chan)]<br />
* [https://en.wikipedia.org/wiki/Doraemon Doraemon]<br />
* [https://en.wikipedia.org/wiki/Felix_the_Cat Felix the Cat]<br />
* [https://en.wikipedia.org/wiki/Harry_Potter Harry Potter]<br />
* [https://en.wikipedia.org/wiki/Keitai_Denj%C5%AB_Telefang Keitai Denjū Telefang (Telefang)]<br />
* [https://en.wikipedia.org/wiki/Kingdom_Hearts Kingdom Hearts]<br />
* [https://en.wikipedia.org/wiki/Living_Books Living Books]<br />
* [https://en.wikipedia.org/wiki/Medabots Medabots (Medarot)]<br />
* [https://en.wikipedia.org/wiki/Pikmin Pikmin]<br />
* [https://en.wikipedia.org/wiki/Pok%C3%A9mon Pokémon (Pocket Monsters)]<br />
* [https://en.wikipedia.org/wiki/The_Smurfs The Smurfs (Les Schtroumpfs; De Smurfen)]<br />
* [https://en.wikipedia.org/wiki/SpongeBob_SquarePants SpongeBob SquarePants (SpongeBob)]<br />
* [https://en.wikipedia.org/wiki/Yo-kai_Watch Yo-kai Watch]<br />
<br />
== Work/Occupations ==<br />
* [https://en.wikipedia.org/wiki/Animal_training animal training]<br />
* [https://en.wikipedia.org/wiki/Cooking cooking (food preparation)]<br />
** [https://en.wikipedia.org/wiki/Candy_making candy making]<br />
* [https://en.wikipedia.org/wiki/Gardening gardening]<br />
* [https://en.wikipedia.org/wiki/Lifeguard lifeguard]<br />
* [https://en.wikipedia.org/wiki/Marine_biology marine biology]<br />
* [https://en.wikipedia.org/wiki/Woodworking woodwork]<br />
** [https://en.wikipedia.org/wiki/Pyrography wood burning]<br />
<br />
<!-- == Miscellaneous == --><br />
<br />
----<br />
<br />
== Finding Information on These Topics ==<br />
There are numerous places and search queries one can use to find websites, videos, software et al. related to these topics. Below are a few places and search query "templates" you can tweak and try yourself:<br />
<br />
* social media ([[Social_network#List_of_social_networks|see also]])<br />
** [[Facebook]] (usually requires login)<br />
** [[Instagram]] (almost always requires login)<br />
** [[Reddit]]<br />
<!-- *** [https://old.reddit.com/r/all/search?q=searchQuery&nsfw=on basic search] --><br />
** [[TikTok]]<br />
** [[Tumblr]]<br />
** [[Twitter]] (login prompt can be bypassed with a [https://web.archive.org/web/20220713041322/https://old.reddit.com/r/uBlockOrigin/comments/rnay51/twitter_log_in_nag_screen/hpxxcb2/ workaround])<br />
*** [https://twitter.com/explore basic search]<br />
*** [https://twitter.com/search-advanced advanced search]<br />
*** hashtag search: <code>https://twitter.com/search?q=%23<b>hashtagQuery</b></code><br />
** [[VKontakte|VK]]<br />
** [[YouTube]] (often blocks content access behind captchas)<br />
<br />
* search engines commonly utilize special syntax for fine-tuning searches, such as:<br />
** [https://en.wikipedia.org/wiki/DuckDuckGo DuckDuckGo]: [https://help.duckduckgo.com/duckduckgo-help-pages/results/syntax/ search syntax guide]<br />
** [[Google]]: [https://support.google.com/websearch/answer/2466433 search syntax guide]<br />
<br />
* search query "templates," such as:<br />
** <i><b>topic</b> blogs</i><br />
** <i><b>topic</b> blogspot</i><br />
** <i><b>topic</b> wordpress</i><br />
** <i><b>topic</b> for beginners</i><br />
** <i>all about <b>topic</b></i><br />
** <i>intro to <b>topic</b></i><br />
** <i><b>topic</b> [[Reddit|subreddits]]</i><br />
** <i><b>topic</b> videos</i><br />
** <i><b>topic</b> [https://en.wikipedia.org/wiki/Convention_(meeting) conventions]</i><br />
** <i><b>topic</b> [[Tumblr|tumblr]] blogs</i><br />
** <i><b>topic</b> tutorials</i></div>Systwihttps://wiki.archiveteam.org/index.php?title=YouTube/Technical_details&diff=49175YouTube/Technical details2022-11-20T20:29:50Z<p>Systwi: /* Domains */ Adjusted year & added gaming.youtube.com.</p>
<hr />
<div>This page documents some of the publicly known technical details of YouTube.<br />
<br />
== ID formats ==<br />
In most places, IDs are expressed as base64 using the modified character set <code>A-Za-z0-9-_</code>.<br />
<br />
=== Videos ===<br />
Videos have a 64-bit ID. Because an 11-character base64 string is equivalent to 66 bit, the last character of a video ID can only take one of 16 values.<br />
<br />
Video ID regex pattern: <code>[A-Za-z0-9_-]{10}[AEIMQUYcgkosw048]</code><br />
<br />
In the past, YouTube also accepted requests with video IDs that had their last character mangled.<ref>{{URL|https://stackoverflow.com/questions/35731554/every-youtube-video-has-two-links}} (Note that the details in the analysis are incorrect as it uses the ASCII value of <code>k</code>; in the base64 alphabet, it decodes to <code>0b100100</code>.)</ref> For example, an ID ending in <code>A</code> could also be requested with <code>B</code>, <code>C</code>, or <code>D</code> instead, and likewise for the other characters (the final one being <code>9</code>, <code>-</code>, or <code>_</code> instead of <code>8</code>). This stopped working sometime between 2020-01-07 and 2020-08-16, and now YouTube only accepts requests where the last two bits of the base64 ID are zero.<br />
<br />
=== Channels ===<br />
Channels have a 128-bit ID. In base64, this turns into a 22-character string where the last character can take 4 values.<br />
<br />
Channel ID regex pattern: <code>[A-Za-z0-9_-]{21}[AQgw]</code><br />
<br />
(Note that this does not include the <code>UC</code> prefix used e.g. in <code>/channel/</code> URLs. The channel ID appears without that prefix in several places, most notably some playlist IDs.)<br />
<br />
Additionally, all channels created before approximately 2012-03-14<ref>{{URL|https://blog.youtube/news-and-events/seeing-familiar-faces-on-youtube/}}</ref> and some channels created between approximately 2012-03-14 and approximately 2014-11-24<ref>{{URL|https://blog.youtube/news-and-events/want-your-url-to-match-your-channel/}}</ref> have usernames associated with them. These usernames are permanently associated with the channels and cannot be changed or removed from them<ref>{{URL|https://support.google.com/youtube/answer/6180214}}</ref>. They are accessed by using <code>/user/</code> URLs. If a username has not been claimed as a custom URL, the root URL matching that username will redirect to the channel associated with that username. For example: <code>youtube.com/apple</code> redirects to <code>youtube.com/user/Apple</code>.<br />
<br />
The username system has since been replaced with a custom URL system. These custom URLs are case and diacritic insensitive and can always be accessed from the root URL matching that custom URL as well as under <code>/c/</code>. (Note that these URL patterns were different until approximately 2017-05-31 as <code>/c/</code> URLs used to be associated with Google+ while root URLs were only associated with YouTube<ref>{{URL|https://blog.youtube/news-and-events/improving-custom-urls-on-youtube/}}</ref>.) These URLs are not permanently associated with a channel and can be changed<ref>{{URL|https://support.google.com/youtube/answer/2657968}}</ref>.<br />
<br />
Username and custom URLs can be resolved to UCIDs by using a youtubei endpoint instead of downloading the full channel page by using a script like [https://gist.github.com/tech234a/7ec4c215291a4625c3bb83b75f9eeca8 this one].<br />
<br />
=== Playlists ===<br />
Over the years, there has been a large number of playlist types, many of which have since gone the way of the dodo. This section attempts to document them all.<br />
<br />
In the table below, ''pattern'' is a regex with the additional syntax of <code>&lt;videoID&gt;</code> to indicate that the ID contains a video ID (and likewise for channel and playlist IDs).<br />
<br />
{| class="wikitable sortable"<br />
! Pattern !! Purpose !! Status !! Examples !! Notes<br />
|-<br />
| <code>AL[A-Za-z0-9_-]{32}</code> || Artist playlist (Top tracks, albums) || Broken || [https://web.archive.org/web/20121030071819/https://www.youtube.com/playlist?list=ALGLx1orRGw4XggssIZEv5-nCsJlnb8Xc0] [https://web.archive.org/web/20160530212508/https://www.youtube.com/playlist?list=AL94UKMTqg-9BZaAhb-jCKxmoPo01ounO2] || All known IDs for 'Top tracks' start with <code><nowiki>94UKMTqg-9[A-D]|HTd1VmZQRN[o-r]|YL4kY05133[o-r]|zxNYRMVOCR[g-j]</nowiki></code> and for 'Albums' with <code><nowiki>GLx1orRGw4[U-X]|Nb4maWNoT6[Q-T]</nowiki></code><br />
|-<br />
| <code>AV[A-Za-z0-9_-]{32}</code> || Artist on <code>/artist?a=</code> (AV = 'Artist Videos') || Broken || [https://web.archive.org/web/20120309232750/https://www.youtube.com/artist?a=GxdCwVVULXftjkN-p_QIMzXdCKPB-HqD&feature=playlist-comment] || All known IDs start with <code><nowiki>4oVf-d_DwK[A-D]|AYMcY2vx8G[Q-T]|GxdCwVVULX[c-f]|TGnpyrBl25[w-z]|acdYuOj0G3[s-v]|osme4KCrHq[o-r]</nowiki></code>, with the <code>G</code> one being by far the most common.<br />
|-<br />
| <code>BB[0-9]{8}</code> || Billboard top 20 || Unviewable, playnext broken, list param on watch page not functional but not removed by JS || [https://web.archive.org/web/20120504170738/http://www.youtube.com/playlist?list=BB20120505] || The ID is a date in <code>YYYYMMDD</code> format and refers to a Saturday. Known values range from 20110910 to 20121201.<br />
|-<br />
| <code>BLCM&lt;channelID&gt;</code> || Channel Mix || Fully functional || [https://www.youtube.com/playlist?list=BLCMq6VFHwMzcMXbuKyG7SQYIg] || The title is "Latest <Channel name>" while the description is always "Recent videos from this channel, updated weekly". These only work for logged in users.<br />
|-<br />
| <code>BLEM[A-Za-z0-9_-]{22}</code> || Weekly Topic Mix || Fully functional || [https://www.youtube.com/playlist?list=BLEMQ1dJ7wXfLlqCjwV0xfSNbA] || The title is "<Topic> weekly" and the description is "Topical videos, selected for you and updated weekly". These only work for logged in users.<br />
|-<br />
| <code><nowiki>CL([A-Za-z0-9_-]{11}|[A-Za-z0-9_-]{22})</nowiki></code> || Season clips/previews || Fully functional || [https://www.youtube.com/playlist?list=CLnxWDOyerMoJtvI-vnT3QjA] [https://www.youtube.com/playlist?list=CLdJ1NYyglkTE] || In some cases, the playlist exists even though the video doesn't [https://www.youtube.com/playlist?list=CLuSOli2NxJmY]; for these, the list parameter also works on the watch page. In other cases, the playlist doesn't exist but the video does [https://www.youtube.com/playlist?list=CL2yzd6Rg4La0].<br />
|-<br />
| <code><nowiki>EC([0-9A-F]{16}|[A-Za-z0-9_-]{32})</nowiki></code> || Courses || Unviewable, playnext and watch functional || [https://www.youtube.com/playlist?list=EC00615559FE621D9B&playnext=1] || IDs are also fully functional with <code>PL</code>: [https://www.youtube.com/playlist?list=PL00615559FE621D9B]<br />
|-<br />
| <code><nowiki>EL([A-Za-z0-9_-]{11}|[A-Za-z0-9_-]{22})</nowiki></code> || Season episodes || Fully functional || [https://www.youtube.com/playlist?list=ELuSOli2NxJmY] [https://www.youtube.com/playlist?list=EL7B5xbXKtkRaOxLNVdxHYGA] || Some of the short IDs are video IDs, others are not. Even though the contents seem entirely unrelated (e.g. [https://web.archive.org/web/20201024024006/https://www.youtube.com/playlist?list=EL-fNI2W9J9t8&disable_polymer=1] vs [https://www.youtube.com/watch?v=-fNI2W9J9t8]), it's ''highly'' unlikely that these are random collisions since there are quite a lot of them. The longer IDs are unique season playlist IDs (not to be confused with the season IDs seen on the <code>/show/</code> pages). Both short and long IDs can usually be used for <code>CL</code> playlists.<br />
|-<br />
| <code>FL&lt;channelID&gt;</code> || Favourites || Fully functional || [https://www.youtube.com/playlist?list=FLK8sQmJBp8GCxrOtXWBpyEA] ||<br />
|-<br />
| <code>HL[0-9]{10}</code> || || Unviewable, playnext broken || [https://web.archive.org/web/20131106013330/http://www.youtube.com/watch?v=U2-a80nzPAE&list=HL1371590912] [https://web.archive.org/web/20160916111948/http://www.youtube.com/watch?v=Q0gVRten4ss&list=HL1364024750] || Frequently associated with <code>feature=mh_lolz</code>. All WBM snapshots with this are either video pages or <code>/embed/videoseries</code>, and even the older snapshots provide no insight into this playlist type.<br />
|-<br />
| <code>LE&lt;channelID&gt;[APU]</code> || Live events || Broken || [https://web.archive.org/web/20121112140027/http://www.youtube.com/watch?v=rKjbt2wHbhM&list=LED7-53gvqiYTqPmz0EFCTqAP] || Frequently associated with <code>feature=plcp</code><br />
|-<br />
| <code>LL&lt;channelID&gt;</code> || Likes || Private || || Made private in December 2019, also accessible via <code>LL</code> when logged in.<br />
|-<br />
| <code>LM</code> || Liked music || Private || [https://www.youtube.com/playlist?list=LM] || Tied to the account accessing YouTube.<br />
|-<br />
| <code>LM&lt;channelID&gt;</code> || || Broken || ||<br />
|-<br />
| <code>LP&lt;videoID&gt;</code> || Auto-generated channels || Broken || [https://web.archive.org/web/20120831185200/http://www.youtube.com/playlist?list=LP4qRk91tndwg&feature=playlist-comment] [https://web.archive.org/web/20121208025243/https://www.youtube.com/playlist?list=LP-25gYsgSHAs&feature=playlist-comment] [https://web.archive.org/web/20120719164200/https://www.youtube.com/watch?v=DOoXhCyq1-s&list=LP-PgXFUOYEqE&index=1&feature=plcp] || It isn't always clear whether there is any relation between the two video IDs in the watch URL, e.g. [https://web.archive.org/web/20160624202428/http://www.youtube.com/watch?v=2kFbrjfPM_M&list=LP8mCexcOt47s&index=1&feature=plcp] (vs [https://www.youtube.com/watch?v=8mCexcOt47s]) or [https://web.archive.org/web/20121018174159/https://www.youtube.com/watch?v=bxaB85-qAnY&list=LPC_Omnjyzg0A&index=1&feature=plcp] (vs [https://www.youtube.com/watch?v=C_Omnjyzg0A])<br />
|-<br />
| <code>MC([0-9]{8})?</code> || Music Charts || Broken || [https://web.archive.org/web/20140317224526/https://www.youtube.com/playlist?list=MC] [https://web.archive.org/web/20110711043948/http://www.youtube.com:80/watch?v=8gyLR4NfMiI&feature=autoplay&list=MC20110628&index=10&playnext=1] || Daily/weekly global top 100 songs. Dates are Tuesdays, known range is 20110621 to 20110816.<br />
|-<br />
| <code>MC&lt;CC&gt;(\.[0-9]{8}(\.[0-9]+)?)?</code> || Music Charts || Broken || [https://web.archive.org/web/20140317224627/http://www.youtube.com/playlist?list=MCCH] [https://web.archive.org/web/20140622190021/http://www.youtube.com/playlist?list=MCUS.20140617&app=desktop] [https://web.archive.org/web/20140622185202/http://www.youtube.com/playlist?list=MCUS.20140617.453] || <code>&lt;CC&gt;</code> is a country code; known values are AT, CH, DE, RU, and US. Dates are Tuesdays, known range 20110906 to 20140617. The final digits are a genre ID.<br />
|-<br />
| <code>ML</code> || || || ||<br />
|-<br />
| <code>ML[A-Za-z0-9_-]{32}</code> || Artist mix || Broken || [https://web.archive.org/web/20110607085107/https://www.youtube.com/watch?v=9yZ1uI5yPbY&feature=artist&playnext=1&list=ML4oVf-d_DwKBUxbFJjPKyGEuMsx4Z5xum] || Same ID as <code>AV</code><br />
|-<br />
| <code>MLCA&lt;channelID&gt;</code> || Library artist mix || Unviewable, playnext and watch functional || [https://www.youtube.com/playlist?list=MLCAk_XWyHOD2ynJb3ekJBwdew&playnext=1] || Tied to the account accessing YouTube. The channel should be added to the music library via Youtube Music.<br />
|-<br />
| <code>MLCT</code> || Library tracks || Unviewable, playnext and watch functional || [https://www.youtube.com/playlist?list=MLCT&playnext=1] || Tied to the account accessing YouTube. The playlist includes all the tracks added to your Youtube Music library.<br />
|-<br />
| <code>OLAK5uy_[klmn][A-Za-z0-9_-]{32}</code> || Albums || Fully functional || [https://www.youtube.com/playlist?list=OLAK5uy_k--30R4f6IL_A9Nu1SWw4N2Rmlfybp83Y] ||<br />
|-<br />
| <code>PL[0-9A-F]{16}</code> || Normal playlist (old) || Fully functional || [https://www.youtube.com/playlist?list=PLC9B6711584A06935] ||<br />
|-<br />
| <code>PL[A-Za-z0-9_-]{32}</code> || Normal playlist || Fully functional || [https://www.youtube.com/playlist?list=PL7nj3G6Jpv2G6Gp6NvN1kUtQuW8QshBWE] || The 32-character ID consists of a 10-character channel-specific unique ID which is common for all the playlists on a channel and a 22-character playlist-specific ID. Thus, the full playlist ID can be represented as PL[A-Za-z0-9_-]{10}[A-Za-z0-9_-]{22}. It is possible that all the other playlist types IDs which use 32 characters have a similar structure.<br />
|-<br />
| <code>PPSV</code> || Mobile downloads, offline videos || || || Tied to the account accessing YouTube; 'Private Playlist Saved Videos'?<br />
|-<br />
| <code>PU&lt;channelID&gt;</code> || Popular uploads || Unviewable, playnext and watch functional || [https://www.youtube.com/playlist?list=PUZM0q3tj_RstUAUM4ox3R3g&playnext=1] ||<br />
|-<br />
| <code>RD&lt;videoID&gt;</code> || Mix aka radio || Unviewable, playnext and watch functional || [https://www.youtube.com/playlist?list=RDdQw4w9WgXcQ&playnext=1] ||<br />
|-<br />
| <code>RD[0-4][0-9]&lt;videoID&gt;</code> || Mix || Broken || [https://web.archive.org/web/20121206111014/http://www.youtube.com/watch?v=u6Py7wxBVbs&playnext=1&list=RD01_uElJaFMHzM] [https://web.archive.org/web/20130902043345/http://www.youtube.com:80/watch?v=3O1_3zBUKM8&list=RD023O1_3zBUKM8] [https://web.archive.org/web/20130926201345/http://www.youtube.com/watch?v=HNxVNog2Qdw&feature=youtu.be&list=RD03hhBx8eM-Y7Q] [https://web.archive.org/web/20130902093936/http://www.youtube.com/watch?v=0OlY7MwihXY&feature=youtu.be&list=RD04RJvEhVd-ZWU] [https://web.archive.org/web/20131221192452/http://www.youtube.com:80/watch?v=6DUgnZPag04&list=RD05NG4nR_oVk] [https://web.archive.org/web/20131229075110/http://www.youtube.com:80/watch?v=6oW44aeioio&list=RD06zc5JllZC0] [https://web.archive.org/web/20140221003818/http://www.youtube.com/watch?v=Q3__TEQJMvU&list=RD07Ndh2y0m8E] [https://web.archive.org/web/20130513031455/http://www.youtube.com/watch?v=9HPiBJBCOq8&list=RD089bZkp7q19f0] [https://web.archive.org/web/20130302000502/http://www.youtube.com/watch?v=_orfcGnaseE&list=RD13oGtRiCjWmr8] [https://web.archive.org/web/20130509053006/http://www.youtube.com/watch?v=55HTF0h4QAQ&list=RD15N6pxQ7zuCdg] [https://web.archive.org/web/20140125205713/http://www.youtube.com/watch?v=iWJGzR6EdIg&list=RD16rb7wIep6xLg] [https://web.archive.org/web/20130502172329/http://www.youtube.com/watch?v=sUO6-Byrel0&list=RD17XynxTU8ovZ8] [https://web.archive.org/web/20130626103949/http://www.youtube.com/watch?v=cpJpVl2pVUE&list=RD18lExom22Ps68] [https://web.archive.org/web/20131110235936/http://www.youtube.com/watch?v=UDKADsUiE28&list=RD19JTgzyFXmI3w] [https://web.archive.org/web/20140203060741/http://www.youtube.com/watch?v=OBXN1t7K2pY&list=RD20iLTwNuLWLLo] [https://web.archive.org/web/20130815160803/http://www.youtube.com/watch?v=VRaSwAXhEr4&list=RD21KW57OlqEBuA] [https://web.archive.org/web/20131018145156/http://www.youtube.com/watch?v=CevxZvSJLk8&list=RD22My2FRPA3Gf8] [https://web.archive.org/web/20130515041606/http://www.youtube.com/watch?v=JK_DOJa99oo&list=RD24p04TYk4j0zQ] [https://web.archive.org/web/20130715091251/http://www.youtube.com/watch?v=tfdbWutUDAw&list=RD25NIirqJ1Ow8Q] [https://web.archive.org/web/20130515193647/http://www.youtube.com/watch?v=ErLbge1Pl80&list=RD27C7MDjeNZiAI] [https://web.archive.org/web/20130816042714/http://www.youtube.com/watch?v=w1HDYxvBa60&list=RD28grvpTTu09bQ] [https://web.archive.org/web/20130927235020/http://www.youtube.com/watch?v=NVcSNnqRD0c&list=RD29ph1VTId1YNc] [https://web.archive.org/web/20130620223124/http://www.youtube.com/watch?v=qGpXBE33m0g&list=RD32JheYx263ICY] [https://web.archive.org/web/20130904083749/http://www.youtube.com/watch?v=7fNuTf76FlE&list=RD35a2H89ulLtqk] [https://web.archive.org/web/20130814155841/http://www.youtube.com/watch?v=WE-B3J-n_qg&list=RD39iWFw28zJ4HU] [https://web.archive.org/web/20130808015529/http://www.youtube.com:80/watch?v=1kMF6kR1ojY&list=RD40VRcrZ8CC2Vo] [https://web.archive.org/web/20130914150229/https://www.youtube.com/watch?v=GDSQ2Zg2kKM&list=RD47J_kGJApa4DM] || By far the most common prefix among known URLs is <code>RD02</code>. All other confirmed values are listed in the examples, but given the apparent existence of links around the web, it is likely that more existed. They were discontinued in early July 2014 ([https://web.archive.org/web/20140702055455/http://www.youtube.com/watch?v=KLv7rmb6G9c&list=RD02YXUIAqGFYNw] [https://web.archive.org/web/20140704143037/http://www.youtube.com/watch?v=JmM04_OgajU&list=RD02wcLNteez3c4]).<br />
|-<br />
| <code>RDAMVM&lt;videoID&gt;</code> || Mix || Unviewable, playnext and watch functional || [https://www.youtube.com/watch?v=dQw4w9WgXcQ&list=RDAMVMdQw4w9WgXcQ] [https://www.youtube.com/playlist?list=RDAMVMdQw4w9WgXcQ&playnext=1] ||<br />
|-<br />
| <code>RDAMPL&lt;playlistID&gt;</code> || || || ||<br />
|-<br />
| <code>RDAO[A-Za-z0-9_-]{22}</code> || Artist shuffle mix || Unviewable, playnext and watch functional || [https://www.youtube.com/playlist?list=RDAOGTJYzN4zJlP1hWbgi6vt_g&playnext=1] [[https://www.youtube.com/watch?v=edAxujKev1I&list=RDAOGTJYzN4zJlP1hWbgi6vt_g] || Same ID as <code>RDEM</code><br />
|-<br />
| <code>RDCLAK5uy_[klmn][A-Za-z0-9_-]{32}</code> || Compilation || Fully functional || [https://www.youtube.com/playlist?list=RDCLAK5uy_mfut9V_o1n9nVG_m5yZ3ztCif29AHUffI] || Extensively used on https://www.youtube.com/channel/UC-9-kyTW8ZkZNDHQJ6FgpwQ aka https://www.youtube.com/music<br />
|-<br />
| <code>RDCMUC&lt;channelID&gt;</code> || Channel radio mix || Unviewable, playnext and watch functional || [https://www.youtube.com/playlist?list=RDCMUCRvvekKryKLFOj4l3vQgZnA&playnext=1] || Similar to BLCM<br />
|-<br />
| <code>RDEM[A-Za-z0-9_-]{22}</code> || Artist/genre mix || Unviewable, playnext and watch functional || [https://www.youtube.com/playlist?list=RDEMswfFaRBz22qxwFKV9jqZEg&playnext=1] || Same ID as <code>BLEM</code><br />
|-<br />
| <code>RDGMEM[A-Za-z0-9_-]{22}</code> || Genre/mood mix || Unviewable, playnext broken, watch functional || [https://www.youtube.com/watch?v=IZ52lk9wjZI&list=RDGMEM6CZm14o9sc-Q22TIneLI8g] || Can use the same ID as <code>RDEM</code><br />
|-<br />
| <code>RDGMEM[A-Za-z0-9_-]{22}VM&lt;videoID&gt;</code> || Genre/mood mix || Unviewable, playnext and watch functional || [https://www.youtube.com/playlist?list=RDGMEMXdNDEg4wQ96My0DhjI-cIgVMT_lC2O1oIew&playnext=1] [https://www.youtube.com/watch?v=T_lC2O1oIew&list=RDGMEMXdNDEg4wQ96My0DhjI-cIgVMT_lC2O1oIew] || This type behaves like an RDEM and RD hybrid<br />
|-<br />
| <code>RDHC&lt;videoID&gt;</code> || Artist mix? || Broken || [https://web.archive.org/web/20141113195350/http://www.youtube.com/watch?v=50X8EXq6Tlc&list=RDHC-0trwxfNivI] ||<br />
|-<br />
| <code>RDKM[A-Za-z0-9_-]{22}</code> || Topic Radio Mix || Unviewable, playnext and watch functional || [https://www.youtube.com/playlist?list=RDKMafPiHxTAgNfcfWY88IWjBQ&playnext=1] [https://www.youtube.com/watch?v=EfHxUaJDEn8&list=RDKMafPiHxTAgNfcfWY88IWjBQ] || Same ID as <code>RDEM</code><br />
|-<br />
| <code>RDLV&lt;videoID&gt;</code> || Mix || Unviewable, playnext broken, watch functional || [https://www.youtube.com/watch?v=EfHxUaJDEn8&list=RDLVEfHxUaJDEn8] ||<br />
|-<br />
| <code>RDMM</code> || My Mix || || || Tied to the account accessing YouTube<br />
|-<br />
| <code>RDMM&lt;videoID&gt;</code> || My Mix || Unviewable, playnext and watch functional || [https://www.youtube.com/playlist?list=RDMM--41OGPMurU&playnext=1] [https://www.youtube.com/watch?v=--41OGPMurU&list=RDMM--41OGPMurU] ||<br />
|-<br />
| <code>RDQM&lt;videoID&gt;</code> || Video Mix || Unviewable, playnext and watch functional || [https://www.youtube.com/playlist?list=RDQM-0lLj66y6ts&playnext=1] [https://www.youtube.com/watch?v=-0lLj66y6ts&list=RDQM-0lLj66y6ts] || Works even if the target video is unavailable as in the example.<br />
|-<br />
| <code>RDSH</code> || Shorts Mix? || || ||<br />
|-<br />
| <code>RDTMAK5uy_[klmn][A-Za-z0-9_-]{32}</code> || Tailor-made mix || Fully functional || [https://www.youtube.com/playlist?list=RDTMAK5uy_nyhwsGy3aIN3Okw7vCxz-bt2-_AxQiHTM] || Requires authentication<br />
|-<br />
| <code>RDVM&lt;videoID&gt;</code> || Video Mix || Unviewable, playnext and watch functional || [https://www.youtube.com/playlist?list=RDVM-5ZxuLzHxFY&playnext=1] [https://www.youtube.com/watch?v=2MMSc5qAIEg&list=RDVM2MMSc5qAIEg] ||<br />
|-<br />
| <code>RLTD&lt;channelID&gt;</code> || Stories (Reel?) || Unviewable, playnext broken || || [https://www.youtube.com/playlist?list=RLTDBR8-60-B28hp2BmDPdntcQ]<br />
|-<br />
| <code>SL</code> || Season (TV series?) || Broken || [https://web.archive.org/web/20100725132828/http://www.youtube.com/watch?v=GNdhphtB-8Y&list=SL] ||<br />
|-<br />
| <code>SL&lt;videoID&gt;</code> || Season (TV series?) || Broken || [https://web.archive.org/web/20120413213623/http://www.youtube.com:80/watch?v=6vPDvbD1SKg&feature=autoplay&list=SLo5wUQzGudyI&lf=list_related&playnext=2] ||<br />
|-<br />
| <code><nowiki>SP([0-9A-F]{16}|[A-Za-z0-9_-]{32})</nowiki></code> || Series playlist || Fully functional || [https://web.archive.org/web/20120607210346/http://www.youtube.com/watch?v=Rss9T7lu27k&list=SP3C14444766460071] [https://web.archive.org/web/20140819072011/http://www.youtube.com/playlist?list=SP06CD75321C3C3405] [https://web.archive.org/web/20130311214010id_/http://www.youtube.com/watch?v=3AilblBXlWU&list=UU7Edgk9RxP7Fm7vjQ1d-cDA&index=3] || Behaves exactly as the same ID with <code>PL</code> and already did so in 2012 (see example).<br />
|-<br />
| <code>TL[A-Za-z0-9_-]{11} || || Broken || [https://web.archive.org/web/20111206165653/http://www.youtube.com/watch?v=YGKJC22Roxs&playnext=1&list=TLCGgmn2FjODc] [https://web.archive.org/web/20120622153142/http://www.youtube.com/watch?v=-iBduigB4v4&playnext=1&list=TL8Us4LT2BvMM] [https://web.archive.org/web/20131215042652/http://www.youtube.com/watch?v=--7JzR1Hhx0&list=TLcsjn3rKwAEb_osN_QtHc6g12s8oLWzi1] || Details unclear, possibly also temporary: [https://web.archive.org/web/20121108191805/https://www.youtube.com/watch?v=2xeAyliOF_k&playnext=1&list=TLsmhIqd0d9Ck] [https://web.archive.org/web/20121109234527/https://www.youtube.com/watch?v=2xeAyliOF_k&playnext=1&list=TLsmhIqd0d9Ck], [https://web.archive.org/web/20121108191856/https://www.youtube.com/watch?v=165VjNKRNdw&playnext=1&list=TL0RtrN-ER_Ks] [https://web.archive.org/web/20121110013916/https://www.youtube.com/watch?v=165VjNKRNdw&playnext=1&list=TL0RtrN-ER_Ks]<br />
|-<br />
| <code>TL[A-Za-z0-9_-]{22} || 'Templist' || || [https://web.archive.org/web/20160228191431/https://www.youtube.com/watch?v=-Z0_lfGaICM&list=TL-38rKMsdYVMyODAyMjAxNg] [https://web.archive.org/web/20150830130112/https://www.youtube.com/watch?v=mrgtxlkpfFo&list=TL-4j-2ZVIreMzMDA4MjAxNQ] [https://web.archive.org/web/20151103043325/https://www.youtube.com/watch?v=n_FWStKEos4&list=TL-6xoWohYdCEwMjExMjAxNQ] [https://web.archive.org/web/20151001134406/https://www.youtube.com/watch?v=sF8q9nSkD_Y&list=TLA2i7a2zTLjIwMTEwMjAxNQ] ||<br />
|-<br />
| <code>TL[A-Za-z0-9_-]{32} || || || [https://web.archive.org/web/20140623013019/http://www.youtube.com/watch?v=5vB2vo8xoQg&list=TL--0ArGFCzduBjD5Pbg_9XerhS08bDnPI] [https://web.archive.org/web/20140209145851/http://www.youtube.com/watch?v=dyBzYCEyUlE&list=TL--5Dc7-eIUxKYJVQHsShov1TCBXQGDBe] [https://web.archive.org/web/20131128061128/http://www.youtube.com/watch?v=6JQPoaTKiu4&list=TL--2UKPutDV4Klvl6pBLwJRb7YDgYuZZo] [https://web.archive.org/web/20140419052830/http://www.youtube.com/watch?v=WyT_JavQU5s&list=TLB1GbpS4y7ZACK9hDWlo8QCigiEbINi_O] ||<br />
|-<br />
| <code>TLGC[A-Za-z0-9_-]{22} || || || [https://web.archive.org/web/20161014132705/https://www.youtube.com/watch?v=5I8UOOhxx5A&list=TLGC1QbHtr1DxEMxNDEwMjAxNg] [https://web.archive.org/web/20161016222434/https://www.youtube.com/watch?v=8MYZpElW494&list=TLGCH5t7TluDMgsxNjEwMjAxNg] || When decoded, the last eight characters are a date in <code>DDMMYYYY</code> format. However, the timezone is unclear; <code>TLGC1QbHtr1DxEMxNDEwMjAxNg</code> turns out as <code>15102016</code>, even though the list clearly existed at 2016-10-14 13:27 UTC.<br />
|-<br />
| <code>TLGG[A-Za-z0-9_-]{22} || Temporary list || Fully functional || [https://web.archive.org/web/20211201164001/https://www.youtube.com/playlist?list=TLGGoPMGoLZfxBUwMTEyMjAyMQ] [https://web.archive.org/web/20211201163922/https://www.youtube.com/watch?v=dQw4w9WgXcQ&list=TLGGoPMGoLZfxBUwMTEyMjAyMQ] || Produced by <code><nowiki>https://www.youtube.com/watch_videos?video_ids=&lt;videoID&gt;,&lt;videoID&gt;,...</nowiki></code> and <code><nowiki>https://www.youtube-nocookie.com/embed/&lt;videoID&gt;?playlist=&lt;videoID&gt;,&lt;videoID&gt;,...</nowiki></code>. When the full ID (including the prefix) is decoded, the last eight characters are the date of the list's creation in <code>DDMMYYYY</code> format. Lists appear to expire after two days.<br />
|-<br />
| <code>TLPP[A-Za-z0-9_-]{22} || Temporary list || || || Also produced by <code>/watch_videos</code> (example: [https://web.archive.org/web/20211102160829/https://www.youtube.com/watch_videos?type=0&feature=c4-overview-u&video_ids=tk2MP5Bc9qQ%2CJ-cxJl6pwjU%2C3rZEYX47CnE%2Ch9Ch_610aRE%2CkFe4071CY3c%2CPmnLoPSLVIE%2CrQUgYwo5o3c%2CgQV4hl0I9Bo%2CSFm51wTGZPw%2CgQmMwjTlEQY%2C3CQgmSYPI2w%2COjRDK4kQjfY&index=2&title=Popular+uploads&more_url&spf=navigate] → [https://web.archive.org/web/20211102160830/https://www.youtube.com/watch?v=3rZEYX47CnE&list=TLPPMDIxMTIwMjFppbc4wP6OQQ]), most likely dependent on one of the query parameters.<br />
|-<br />
| <code>TLPP[A-Za-z0-9_-]{30} || || || [https://web.archive.org/web/20140729023845/http://www.youtube.com/watch?v=yukIOuF4OYE&list=TLPPmv2nKQLjlofGsTp2Jqxrp_u3c-rtME] [https://web.archive.org/web/20140912065410/https://www.youtube.com/watch?v=-h4zTEwgCpQ&list=TLppmRwJg8CiOzH1vlD9QLVAcbcXnbKp2g] ||<br />
|-<br />
| <code>TLPP[A-Za-z0-9_-]{54} || || || || UNCONFIRMED. No known working example. The oldest snapshots with such IDs in the Wayback Machine ([https://web.archive.org/web/20161202133253/https://www.youtube.com/watch?v=nedtSOuMyzA&list=TLPPdUw4UVRRYlNEWHpCR3F4ekdDYUFiRDZPcElHcVBmVU8wMjEyMjAxNg] [https://web.archive.org/web/20161206174635/https://www.youtube.com/watch?v=PWFabACQ4-Q&list=TLPPdmF5RXBRSzhCeG1tTnplTTNOTzQtN3dqZEZncno2SDcwNjEyMjAxNg]) do not show a playlist. However, there are 117 unique such URLs in the WBM as of 2021-12-07, so this is unlikely that this is just random noise. When decoded, the last eight characters appear to be a date in <code>DDMMYYYY</code> format.<br />
|-<br />
| <code>TLPQ[A-Za-z0-9_-]{22} || Queue || Private || || Created when a video is added to the queue<br />
|-<br />
| <code>UL&lt;videoID&gt;</code> || User uploads || Watch functional || || Only valid on watch pages. Triggers display of uploads by the same user as the watched video. The ID in the list parameter must be a valid video ID from any channel; in the past, any 11 characters were accepted, and even further in the past, a sole <code>UL</code> worked as well.<br />
|-<br />
| <code>UU&lt;channelID&gt;</code> || User/channel uploads || Fully functional || [https://www.youtube.com/playlist?list=UUZM0q3tj_RstUAUM4ox3R3g] ||<br />
|-<br />
| <code>UUMO&lt;channelID&gt;</code> || User Uploads Members Only || Fully functional || [https://www.youtube.com/playlist?list=UUMOLyKM-d75GEeJGcYI6sr2rQ] ||<br />
|-<br />
| <code>UUSH&lt;channelID&gt;</code> || User shorts || Fully functional || [https://www.youtube.com/playlist?list=UUSHK8sQmJBp8GCxrOtXWBpyEA] ||<br />
|-<br />
| <code>WL</code> || Watch Later || Private || || Tied to the account accessing YouTube<br />
|-<br />
| <code><nowiki>WL([0-9A-F]{16}|[A-Za-z0-9_-]{32})</nowiki></code> || User's Watch Later || Broken || [https://web.archive.org/web/20111223165251/https://www.youtube.com/playlist?list=WL224B512296894D40&feature=playlist-comment] || WL list of a channel, but actually internally a normal <code>PL</code> playlist as evidenced by the profile page corresponding to the example: [https://web.archive.org/web/20111223165246/https://www.youtube.com/user/nikkie755467]. Later, these started to redirect to <code>/my_watch_later_videos</code> instead: [https://web.archive.org/web/20130815105059/http://www.youtube.com/playlist?list=WL9BA14E4873F54DBF]. No working example is known for the 32-char version; the oldest known snapshots in the WBM are from late 2012 and already don't display a playlist: [https://web.archive.org/web/20121121175844/http://www.youtube.com/watch?v=GCrViXx4lus&feature=player_detailpage&list=WLb41BoOUqQxK1mA4i1BrjuqbHerdFwsMY] [https://web.archive.org/web/20121217184226/http://www.youtube.com/watch?v=qtJPAv1UiAE&feature=BFp&list=WLwpw4Htj7V2SFZE7D05LAH0Y8_klT4F1X].<br />
|-<br />
| <code>WL&lt;channelID&gt;</code> || User's Watch Later || Broken || || No known direct example, but there are plenty of URLs in the WBM that were captured after this was discontinued that contain valid channel IDs, e.g. [https://web.archive.org/web/*/https://www.youtube.com/playlist?list=WLDwVhYk6Fqx2l3MzcZQWaJg].<br />
|}<br />
<br />
Further prefixes that have been claimed elsewhere to have existed but whose details aren't known yet: <code>BP</code> ('Branded Playlist'), <code>MQ</code>, <code>QL</code> ('Quicklist'), <code>SV</code> ('Station'), <code>TT</code>, <code>MLPT</code><br />
<br />
== URL formats ==<br />
A large number of URL formats have been in use over the years, too many to be listed here. [[User:JustAnotherArchivist]]'s [https://gitea.arpa.li/JustAnotherArchivist/little-things/src/branch/master/youtube-extract youtube-extract script] contain regex patterns for most of the ones that were at least somewhat common.<br />
<br />
== Domains ==<br />
Domains actively serving content as of 2022:<br />
<br />
* www.youtube.com (main site)<br />
* m.youtube.com (mobile site)<br />
* youtu.be (short URLs)<br />
* www.youtube-nocookie.com (embeds only)<br />
* gaming.youtube.com (video game-centric videos)<br />
* music.youtube.com (YouTube Music)<br />
* www.youtubekids.com (YouTube Kids)<br />
* tv.youtube.com (YouTube TV)<br />
* i.ytimg.com (static images like thumbnails; avatars are on subdomains of ggpht.com as they are shared between Google services)<br />
* subdomains of googlevideo.com (video content)<br />
* studio.youtube.com (the place where creators upload their videos and manage their channels, also has an instance of youtubei)<br />
* youtube.googleapis.com (can load embeds with titles and thumbnails but doesn't actually support playback anymore)<br />
* youtubei.googleapis.com (primarily hosts youtubei for mobile apps)<br />
* www.googleapis.com/youtube/v3/* (official public, rate-limited data API)<br />
<br />
A long time ago, there were YouTube domains under a number of ccTLDs. Nowadays, these all redirect to the main site: youtube.at, by, ca, co.uk, cz, de, dk, ee, es, fi, fr, gr, hr, hu, it, lt, lv, no, pl, pt, ro, rs, ru, se, si, sk.<br />
<br />
In addition, there are some subdomains under youtube.com which redirect to the main site but add the <code>gl</code> parameter to the query string, which changes the interface language: br.youtube.com and likewise for es, it, jp, pl, ru, uk.</div>Systwihttps://wiki.archiveteam.org/index.php?title=File:Tj-anim.gif&diff=48957File:Tj-anim.gif2022-09-10T04:43:43Z<p>Systwi: Systwi uploaded a new version of File:Tj-anim.gif</p>
<hr />
<div>== Summary ==<br />
Lossy animated version of https://tjournal.ru/ closure message.</div>Systwihttps://wiki.archiveteam.org/index.php?title=File:Tj-anim.png&diff=48955File:Tj-anim.png2022-09-09T23:29:12Z<p>Systwi: Lossless animated version of https://tjournal.ru/ closure message.</p>
<hr />
<div>== Summary ==<br />
Lossless animated version of https://tjournal.ru/ closure message.</div>Systwihttps://wiki.archiveteam.org/index.php?title=File:Tj-anim.gif&diff=48954File:Tj-anim.gif2022-09-09T23:28:25Z<p>Systwi: Lossy animated version of https://tjournal.ru/ closure message.</p>
<hr />
<div>== Summary ==<br />
Lossy animated version of https://tjournal.ru/ closure message.</div>Systwihttps://wiki.archiveteam.org/index.php?title=File:Tj-static.png&diff=48953File:Tj-static.png2022-09-09T23:23:48Z<p>Systwi: Static version of https://tjournal.ru/ closure message.</p>
<hr />
<div>== Summary ==<br />
Static version of https://tjournal.ru/ closure message.</div>Systwihttps://wiki.archiveteam.org/index.php?title=TJ&diff=48952TJ2022-09-09T23:27:04Z<p>Systwi: Marked as offline, added webpage screenshot (closure message).</p>
<hr />
<div>{{Infobox project<br />
| URL = https://tjournal.ru/<br />
| logo = Tjournal-logo.png<br />
| image = Tj-static.png<br />
| project_status = {{closed}} on September 10, 2022 (00:00 UTC+03:00)<br />
| archiving_status = {{inprogress}}<br />
| archiving_type = DPoS<br />
| source = [https://github.com/ArchiveTeam/tjournal-grab tjournal-grab]<br />
| tracker = [http://tracker.archiveteam.org/tjournal/ tjournal]<br />
| irc = journalthis<br />
| lead = [[User:Arkiver]]<br />
}}<br />
<br />
'''TJ''' (formerly '''TJournal''') is a Russian news platform. It shut down on 2022-09-10 (00:00 Moscow Time)<ref>{{URL|https://tjournal.ru/team/714914-istoriya-tj-zavershaetsya}}</ref> and, as of 2022-09-09 23:25:41 UTC, its homepage has been replaced with a closure message. All subpages appear to return 500s when retrieved.<br />
<br />
== References ==<br />
<references /></div>Systwihttps://wiki.archiveteam.org/index.php?title=Mobileme&diff=48944Mobileme2022-09-08T22:21:00Z<p>Systwi: Created an article redirecting "Mobileme" to "MobileMe."</p>
<hr />
<div>#REDIRECT [[MobileMe]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=File:Screenshot-MobileMe_Sign_In_-_Google_Chrome.png&diff=48934File:Screenshot-MobileMe Sign In - Google Chrome.png2022-09-08T20:12:09Z<p>Systwi: Systwi uploaded a new version of File:Screenshot-MobileMe Sign In - Google Chrome.png</p>
<hr />
<div>Screenshot of Me.com</div>Systwihttps://wiki.archiveteam.org/index.php?title=TED_Talks&diff=48933TED Talks2022-09-08T14:13:46Z<p>Systwi: /* Backups */ Updated dead torrent date.</p>
<hr />
<div>{{Infobox project<br />
| title = TED Talks<br />
| image = Ted.png<br />
| description = TED mainpage in 2010-12-22<br />
| URL = http://www.ted.com<br />
| project_status = {{online}}<br />
| archiving_status = {{notsaved}} (TPB torrent is gone)<br />
}}<br />
<br />
'''TED Talks''' is a website that provides free talks from famous figures.<br />
<br />
== Backups ==<br />
<br />
* [http://thepiratebay.org/torrent/5420151 All TED talks in mp4, well organized] 35 GB - {{offline}} as of 2022-09-08 14:13:48 UTC<br />
<br />
== Vital signs ==<br />
<br />
Alive and well.<br />
==Downloading TED Talks==<br />
Anyone may download a TED Talks video from an individual video in mp4 in medium resolution under Share > Tools > Download > Download Video. The podcast version from TED Talks Daily may also be downloaded under the Tools section. <br />
== External links ==<br />
<br />
* http://www.ted.com<br />
<br />
{{Navigation box}}<br />
<br />
[[Category:Video hosting]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=File:Screenshot-MobileMe_Sign_In_-_Google_Chrome.png&diff=48931File:Screenshot-MobileMe Sign In - Google Chrome.png2022-09-08T06:32:27Z<p>Systwi: Systwi uploaded a new version of File:Screenshot-MobileMe Sign In - Google Chrome.png</p>
<hr />
<div>Screenshot of Me.com</div>Systwihttps://wiki.archiveteam.org/index.php?title=TikTok&diff=48905TikTok2022-09-01T18:06:09Z<p>Systwi: /* Notes */ Added caution regarding the use of vm.tiktok.com URLs.</p>
<hr />
<div>{{Infobox project<br />
| title = TikTok<br />
| logo = TikTok logo.png<br />
| image = TikTok home page 2019-11-09.png<br />
| description = tiktok home page as seen on November 9, 2019<br />
| URL = https://www.tiktok.com/<br />
| project_status = {{online}} but possibly {{endangered}}<br />
| archiving_status = {{nosavedyet}}<br />
| irc = TikOff<br />
}}<br />
<br />
'''TikTok''' is a social media app for short, user generated videos developed by the Chinese company ByteDance. The app was launched internationally in September 2017 and merged with musical.ly on 9 November 2017. It is currently under intense scrutiny for its collection and use of user data, primarily in its relationship with China.<ref>https://en.wikipedia.org/wiki/TikTok#User_privacy_concerns</ref><br />
<br />
== Douyin ==<br />
<br />
Douyin is the China-only version of TikTok.<br />
<br />
Sites:<br />
* https://www.douyin.com | Main page<br />
* https://sso.douyin.com | Some sort of login thing? Unsure<br />
* https://renzheng.douyin.com | Unsure<br />
* https://v.douyin.com | Redirects to iesdouyin, where videos are<br />
* https://open.douyin.com | Unsure<br />
* https://musician.douyin.com | Unsure<br />
* https://creator.douyin.com | Unsure<br />
* https://www.iesdouyin.com | Videos are seen on this page<br />
<br />
If anyone who can read Chinese would like to help out, that'd be great.<br />
<br />
== Archiving tools ==<br />
<br />
[https://github.com/ytdl-org/youtube-dl youtube-dl] is able to grab videos straight from the site<br />
<br />
A tool to [https://github.com/carcabot/tiktok-signature generate the signature] TikTok uses to sign all their requests exists and is actively updated<br />
<br />
=== TikTokTake ===<br />
<br />
Work is being done to create a tool that can get all of a users videos and WARC them up, that's being implemented in a cli tool and irc bot [https://github.com/JackDallas/TikTokTake TikTokTake]<br />
<br />
=== TikUp ===<br />
<br />
TikUp is a Python package that gets all videos from a user or a hashtag, and uploads them to the Internet Archive under `tiktok-<video id>`. [https://github.com/Coloradohusky/TikUp Github link]<br />
<br />
== Vital Signs ==<br />
<br />
* The first request for archival was in [https://old.reddit.com/r/Archiveteam/comments/dazw5l/tiktok/ October 2019] due to removal of pro-LGBTQ content, even in countries where homosexuality has never been illegal. <ref>[https://www.theguardian.com/technology/2019/sep/26/tiktoks-local-moderation-guidelines-ban-pro-lgbt-content "TikTok's local moderation guidelines ban pro-LGBT content"] - Article on ''The Guardian''</ref><br />
* Videos from Douyin (Chinese TikTok) can not be seen outside of China. <ref>[https://www.wired.com/story/tiktok-is-the-latest-window-into-chinas-police-state/ "TikTok - Yes, TikTok - Is the Latest Window Into China’s Police State"] - Article on ''Wired''</ref><br />
* In August 2020 Trump said "As far as TikTok is concerned, we're banning them from the United States," <ref>[https://web.archive.org/web/20200801075923/https://edition.cnn.com/2020/07/31/tech/tiktok-trump-bytedance-sale/index.html "Trump says he will ban TikTok"] - Article on CNN</ref> <br />
* In 2020-09-18, the Trump administrator will ban downloads of the TikTok app starting 2020-09-20, and access will be banned on November 12.<ref>[https://www.abc.net.au/news/2020-09-19/us-bans-wechat-and-tik-tok-over-national-security-concerns-china/12680516]</ref><br />
<br />
== Archival Locations ==<br />
<br />
[https://archive.org/details/tiktoks Tiktok Video Bin] on the [[Internet Archive]] is where the TikToks uploaded with TikUp and all videos uploaded with the identifier 'tiktok-<video id>' go. The Internet Archive deindexed this collection some time in 2020.<ref>https://archive.fart.website/bin/irclogger_log/archiveteam-ot?date=2021-04-13,Tue</ref><br />
<br />
== Notes ==<br />
<br />
* https://m.tiktok.com/v/<id> is a better way to archive videos, versus https://www.tiktok.com/@<username>/video/<id>. (because it requires the ID only, not the username and the ID)<br />
* Videos without sound can't be viewed or downloaded(it just says 'Video currently unavailable') on desktop, and their links don't work, but they can be viewed and downloaded on mobile.<br />
* If we wanted to make a Warrior, we could use https://vm.tiktok.com/*******/, I think it's alphanumeric, but I'm not sure if these links work for every TikTok.<br />
** A minor caution regarding using https://vm.tiktok.com/*******/ URLs; they redirect to the full video URLs mentioned earlier, but also include telemetry parameters (<code>_t=...</code> and <code>_r=...</code>). It is believed that the former parameter equates to a unique identifier TikTok has assigned to both authenticated and non-authenticated users.<br />
<br />
{{Navigation box}}</div>Systwihttps://wiki.archiveteam.org/index.php?title=TED_Talks&diff=48757TED Talks2022-07-14T22:23:13Z<p>Systwi: /* Backups */ Updated dead torrent date.</p>
<hr />
<div>{{Infobox project<br />
| title = TED Talks<br />
| image = Ted.png<br />
| description = TED mainpage in 2010-12-22<br />
| URL = http://www.ted.com<br />
| project_status = {{online}}<br />
| archiving_status = {{notsaved}} (TPB torrent is gone)<br />
}}<br />
<br />
'''TED Talks''' is a website that provides free talks from famous figures.<br />
<br />
== Backups ==<br />
<br />
* [http://thepiratebay.org/torrent/5420151 All TED talks in mp4, well organized] 35 GB - {{offline}} as of 2022-07-14 22:22:47 UTC<br />
<br />
== Vital signs ==<br />
<br />
Alive and well.<br />
==Downloading TED Talks==<br />
Anyone may download a TED Talks video from an individual video in mp4 in medium resolution under Share > Tools > Download > Download Video. The podcast version from TED Talks Daily may also be downloaded under the Tools section. <br />
== External links ==<br />
<br />
* http://www.ted.com<br />
<br />
{{Navigation box}}<br />
<br />
[[Category:Video hosting]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=MediaFire&diff=48756MediaFire2022-07-14T22:14:09Z<p>Systwi: /* Projects */ Marked https://urls.ajay.app/ as offline.</p>
<hr />
<div>{{Infobox project<br />
| title = Mediafire<br />
| url = https://www.mediafire.com/<br />
| logo = Mediafire-logo-blue.png<br />
| image = Mediafire.com-frontpage.png<br />
| project_status = {{endangered}}<br />
| archiving_status = {{inprogress}}<br />
| tracker = [https://tracker.archiveteam.org/mediafire/ mediafire]<br />
| source = [https://github.com/ArchiveTeam/mediafire-items mediafire-items], [https://github.com/ArchiveTeam/mediafire-grab mediafire-grab]<br />
| irc = mediaonfire<br />
| data = {{IA id|archiveteam_mediafire}}<br />
}}<br />
<br />
'''Mediafire''' is a file hosting service with file synchronisation features. It launched around June 2006.<ref>{{URL|https://en.wikipedia.org/wiki/MediaFire}}</ref><br />
<br />
On December 18 2020, users reported that they began receiving emails from MediaFire how they plan to classify accounts as abandoned if they fail to meet certain criteria, starting in January.<ref>{{URL|https://twitter.com/textfiles/status/1339625133363912706}}</ref> This policy, however, had been in effect for many years prior, as later confirmed by MediaFire.<ref>{{URL|https://twitter.com/tlangridge/status/1349515837351231488}}</ref><br />
<br />
== Projects ==<br />
* '''[https://urls.ajay.app/ URL Submission Site]''': URLs submitted will be periodically added to the archiving queue. - {{offline}} as of 2022-07-14 22:13:37 UTC<br />
<br />
{{Navigation box}}</div>Systwihttps://wiki.archiveteam.org/index.php?title=Topics_of_Archiving_Interest&diff=48755Topics of Archiving Interest2022-07-17T07:31:25Z<p>Systwi: Added more topic categories; added more topics; added information on finding topic content.</p>
<hr />
<div>This page lists any general topics that may have little to no archiving coverage. The goal is to look further into these topics/related subtopics and further expand our archival scope in those areas. It is, by all means, <i>not</i> all-encompassing.<br />
<br />
The topics here are listed in no particular order.<br />
<br />
== Art/Entertainment ==<br />
* [https://en.wikipedia.org/wiki/Beadwork beadwork]<br />
** [https://web.archive.org/web/20220630043840id_/https://i.redd.it/iqyw4xictzd21.jpg Perler bead art]<br />
* [https://en.wikipedia.org/wiki/Board_game board games]<br />
* [https://en.wikipedia.org/wiki/Cross-stitch cross-stitching]<br />
* [https://en.wikipedia.org/wiki/Magic_(illusion) magic acts]<br />
<br />
== Businesses ==<br />
* [https://en.wikipedia.org/wiki/Broderbund Broderbund]<br />
* [https://en.wikipedia.org/wiki/Cranium,_Inc. Cranium, Inc.]<br />
* [https://en.wikipedia.org/wiki/Studio_Ghibli Studio Ghibli] ([https://en.wikipedia.org/wiki/List_of_Studio_Ghibli_works related])<br />
<br />
== Celebrities/Famous Individuals ==<br />
* [https://en.wikipedia.org/wiki/Lida_Abdul Lida Abdul]<br />
* [https://en.wikipedia.org/wiki/Gary_Baseman Gary Baseman]<br />
* [https://en.wikipedia.org/wiki/Reggie_Fils-Aim%C3%A9 Reggie Fils-Aimé (<i>colloq.</i> Reggie)]<br />
* [https://en.wikipedia.org/wiki/Rosa_Parks Rosa Parks]<br />
* [https://en.wikipedia.org/wiki/Sylvester_Stallone Sylvester Stallone]<br />
* [https://en.wikipedia.org/wiki/John_Stamos John Stamos]<br />
* [https://en.wikipedia.org/wiki/Barbra_Streisand Barbra Streisand] ([https://en.wikipedia.org/wiki/Streisand_effect related])<br />
* [https://en.wikipedia.org/wiki/Kiefer_Sutherland Kiefer Sutherland]<br />
* [https://en.wikipedia.org/wiki/Risa_Tabata Risa Tabata]<br />
* [https://en.wikipedia.org/wiki/Toro_y_Moi Toro y Moi (Chaz Bear)]<br />
<br />
== Leisure ==<br />
* [https://en.wikipedia.org/wiki/Sport competitive sports] (casual or championship/professional)<br />
** [https://en.wikipedia.org/wiki/Badminton badminton]<br />
** [https://en.wikipedia.org/wiki/Bowling bowling]<br />
** [https://en.wikipedia.org/wiki/Fishing fishing]<br />
** [https://en.wikipedia.org/wiki/Golf golfing]<br />
** [https://en.wikipedia.org/wiki/Table_football table football (foosball)]<br />
** [https://en.wikipedia.org/wiki/Tennis tennis]<br />
*** [https://en.wikipedia.org/wiki/Table_tennis table tennis]<br />
* [https://en.wikipedia.org/wiki/Yoga yoga]<br />
** [https://en.wikipedia.org/wiki/Meditation meditation]<br />
<br />
== Multimedia Franchises ==<br />
* [https://en.wikipedia.org/wiki/Adiboo Adibou (Adiboo)]<br />
* [https://en.wikipedia.org/wiki/Animal_Crossing Animal Crossing]<br />
* [https://en.wikipedia.org/wiki/Benjamin_the_Elephant Benjamin Blümchen (Benjamin the Elephant)] <br />
* [https://en.wikipedia.org/wiki/Bibi_Blocksberg Bibi Blocksberg]<br />
* [https://en.wikipedia.org/wiki/Club_Kidsoft Club Kidsoft]<br />
* [https://en.wikipedia.org/wiki/Crayon_Shin-chan Crayon Shin-chan (Shin-chan)]<br />
* [https://en.wikipedia.org/wiki/Doraemon Doraemon]<br />
* [https://en.wikipedia.org/wiki/Felix_the_Cat Felix the Cat]<br />
* [https://en.wikipedia.org/wiki/Harry_Potter Harry Potter]<br />
* [https://en.wikipedia.org/wiki/Keitai_Denj%C5%AB_Telefang Keitai Denjū Telefang (Telefang)]<br />
* [https://en.wikipedia.org/wiki/Kingdom_Hearts Kingdom Hearts]<br />
* [https://en.wikipedia.org/wiki/Living_Books Living Books]<br />
* [https://en.wikipedia.org/wiki/Medabots Medabots (Medarot)]<br />
* [https://en.wikipedia.org/wiki/Pikmin Pikmin]<br />
* [https://en.wikipedia.org/wiki/Pok%C3%A9mon Pokémon (Pocket Monsters)]<br />
* [https://en.wikipedia.org/wiki/The_Smurfs The Smurfs (Les Schtroumpfs; De Smurfen)]<br />
* [https://en.wikipedia.org/wiki/SpongeBob_SquarePants SpongeBob SquarePants (SpongeBob)]<br />
* [https://en.wikipedia.org/wiki/Yo-kai_Watch Yo-kai Watch]<br />
<br />
== Work/Occupations ==<br />
* [https://en.wikipedia.org/wiki/Animal_training animal training]<br />
* [https://en.wikipedia.org/wiki/Cooking cooking (food preparation)]<br />
** [https://en.wikipedia.org/wiki/Candy_making candy making]<br />
* [https://en.wikipedia.org/wiki/Gardening gardening]<br />
* [https://en.wikipedia.org/wiki/Lifeguard lifeguard]<br />
* [https://en.wikipedia.org/wiki/Marine_biology marine biology]<br />
* [https://en.wikipedia.org/wiki/Woodworking woodwork]<br />
** [https://en.wikipedia.org/wiki/Pyrography wood burning]<br />
<br />
<!-- == Miscellaneous == --><br />
<br />
----<br />
<br />
== Finding Information on These Topics ==<br />
There are numerous places and search queries one can use to find websites, videos, software et al. related to these topics. Below are a few places and search query "templates" you can tweak and try yourself:<br />
<br />
* social media ([[Social_network#List_of_social_networks|see also]])<br />
** [[Facebook]] (usually requires login)<br />
** [[Instagram]] (almost always requires login)<br />
** [[Reddit]]<br />
<!-- *** [https://old.reddit.com/r/all/search?q=searchQuery&nsfw=on basic search] --><br />
** [[Tumblr]]<br />
** [[Twitter]] (login prompt can be bypassed with a [https://web.archive.org/web/20220713041322/https://old.reddit.com/r/uBlockOrigin/comments/rnay51/twitter_log_in_nag_screen/hpxxcb2/ workaround])<br />
*** [https://twitter.com/explore basic search]<br />
*** [https://twitter.com/search-advanced advanced search]<br />
*** hashtag search: <code>https://twitter.com/search?q=%23<b>hashtagQuery</b></code><br />
** [[VKontakte|VK]]<br />
** [[YouTube]] (often blocks content access behind captchas)<br />
<br />
* search engines commonly utilize special syntax for fine-tuning searches, such as:<br />
** [https://en.wikipedia.org/wiki/DuckDuckGo DuckDuckGo]: [https://help.duckduckgo.com/duckduckgo-help-pages/results/syntax/ search syntax guide]<br />
** [[Google]]: [https://support.google.com/websearch/answer/2466433 search syntax guide]<br />
<br />
* search query "templates," such as:<br />
** <i><b>topic</b> blogs</i><br />
** <i><b>topic</b> blogspot</i><br />
** <i><b>topic</b> wordpress</i><br />
** <i><b>topic</b> for beginners</i><br />
** <i>all about <b>topic</b></i><br />
** <i>intro to <b>topic</b></i><br />
** <i><b>topic</b> [[Reddit|subreddits]]</i><br />
** <i><b>topic</b> videos</i><br />
** <i><b>topic</b> [https://en.wikipedia.org/wiki/Convention_(meeting) conventions]</i><br />
** <i><b>topic</b> [[Tumblr|tumblr]] blogs</i><br />
** <i><b>topic</b> tutorials</i></div>Systwihttps://wiki.archiveteam.org/index.php?title=Topics_of_Archiving_Interest&diff=48733Topics of Archiving Interest2022-06-30T04:43:28Z<p>Systwi: Creating page with working title, content and structure.</p>
<hr />
<div>This page lists any general topics that may have little to no archiving coverage. The goal is to look further into these topics/related subtopics and further expand our archival scope in those areas.<br />
<br />
The topics here are listed in no particular order.<br />
<br />
== Art/Entertainment ==<br />
* [https://en.wikipedia.org/wiki/Beadwork beadwork]<br />
** [https://web.archive.org/web/20220630043840id_/https://i.redd.it/iqyw4xictzd21.jpg Perler bead art]<br />
* [https://en.wikipedia.org/wiki/Cross-stitch cross-stitching]<br />
* [https://en.wikipedia.org/wiki/Magic_(illusion) magic acts]<br />
<br />
== Leisure ==<br />
* [https://en.wikipedia.org/wiki/Fishing fishing]<br />
* [https://en.wikipedia.org/wiki/Yoga yoga]<br />
** [https://en.wikipedia.org/wiki/Meditation meditation]<br />
* [https://en.wikipedia.org/wiki/Bowling bowling/championship bowling]<br />
<br />
== Work ==<br />
* [https://en.wikipedia.org/wiki/Animal_training animal training]<br />
* [https://en.wikipedia.org/wiki/Woodworking woodwork]<br />
** [https://en.wikipedia.org/wiki/Pyrography wood burning]<br />
* [https://en.wikipedia.org/wiki/Cooking cooking (food preparation)]<br />
* [https://en.wikipedia.org/wiki/Marine_biology marine biology]<br />
* [https://en.wikipedia.org/wiki/Gardening gardening]<br />
<br />
== Miscellaneous ==</div>Systwihttps://wiki.archiveteam.org/index.php?title=Template:Wikis&diff=48653Template:Wikis2022-05-28T15:58:56Z<p>Systwi: Formatting correction.</p>
<hr />
<div>{{navbox<br />
|name=Template:Wikis<br />
|status={{{status|autocollapse}}}<br />
|title=Knowledge and Wikis<br />
<br />
|group1=Software<br />
|list1=<br />
[[DokuWiki]]{{·}}<br />
[[MediaWiki]]{{·}}<br />
[[MoinMoin]]{{·}}<br />
[[Oddmuse]]{{·}}<br />
[[PukiWiki]]{{·}}<br />
[[UseModWiki]]{{·}}<br />
[[YukiWiki]]<br />
<br />
|group2=[[WikiTeam#Wikifarms|Wikifarms]]<br />
|list2=<br />
[[atwiki]]{{·}} <br />
[[Battlestar Wiki]]{{·}} <br />
[[BluWiki]]{{·}} <br />
[[Communpedia]]{{·}} <br />
[[EditThis]]{{·}} <br />
[[elwiki.com]]{{·}} <br />
[[Miraheze]]{{·}} <br />
[[Neoseeker.com]]{{·}} <br />
[[Orain]]{{·}} <br />
[[Referata]]{{·}} <br />
[[ScribbleWiki]]{{·}} <br />
[[ShoutWiki]]{{·}} <br />
[[Sourceforge]]{{·}} <br />
[[TropicalWikis]]{{·}} <br />
[[Wik.is]]{{·}} <br />
[[Wiki.Wiki]]{{·}} <br />
[[Wiki-Site]]{{·}} <br />
[[Wikia]]{{·}} <br />
[[Wikidot]]{{·}} <br />
[[WikiHub]]{{·}} <br />
[[Wikispaces]]{{·}} <br />
[[Wikkii]]{{·}} <br />
[[YourWiki.net]]<br />
<br />
|group3=[[Wikimedia]]<br />
|list3=<br />
[[Wikipedia]]{{·}} <br />
[[Wikimedia Commons]]{{·}} <br />
[[Wikibooks]]{{·}} <br />
[[Wikidata]]{{·}} <br />
[[Wikinews]]{{·}} <br />
[[Wikiquote]]{{·}} <br />
[[Wikisource]]{{·}} <br />
[[Wikispecies]]{{·}} <br />
[[Wiktionary]]{{·}} <br />
[[Wikiversity]]{{·}} <br />
[[Wikivoyage]]{{·}}<br />
[[Wikimedia Incubator]]{{·}}<br />
[[Meta-Wiki]]<br />
<br />
|group4=Other<br />
|list4=<br />
[[Anarchopedia]]{{·}} <br />
[[Citizendium]]{{·}} <br />
[[Conservapedia]]{{·}} <br />
[[Creation Wiki]]{{·}} <br />
[[EcuRed]]{{·}} <br />
[[Enciclopedia Libre Universal en Español]]{{·}} <br />
[[GNUPedia]]{{·}} <br />
[[Moegirlpedia]]{{·}} <br />
[[Niconico#Nico Nico Pedia|Nico Nico Pedia]]{{·}} <br />
[[Nupedia]]{{·}} <br />
[[OmegaWiki]]{{·}} <br />
[[OpenStreetMap]]{{·}} <br />
[[Pixiv#Pixiv Encyclopedia|Pixiv Encyclopedia]]<br />
<br />
|group5=Indexes<br />
|list5=<br />
[[WikiApiary]]{{·}} <br />
[[WikiIndex]]<br />
<br />
|end=<br />
}}[[Category:Wikis]]<noinclude>[[Category:Templates]]</noinclude></div>Systwihttps://wiki.archiveteam.org/index.php?title=Template:Wikis&diff=48639Template:Wikis2022-05-24T19:18:05Z<p>Systwi: Added MoinMoin wiki software (https://moinmo.in/).</p>
<hr />
<div>{{navbox<br />
|name=Template:Wikis<br />
|status={{{status|autocollapse}}}<br />
|title=Knowledge and Wikis<br />
<br />
|group1=Software<br />
|list1=<br />
[[DokuWiki]]{{·}}<br />
[[MediaWiki]]{{·}}<br />
[[MoinMoin]]{{·}}<br />
[[PukiWiki]]{{·}}<br />
[[YukiWiki]]<br />
<br />
|group2=[[WikiTeam#Wikifarms|Wikifarms]]<br />
|list2=<br />
[[atwiki]]{{·}} <br />
[[Battlestar Wiki]]{{·}} <br />
[[BluWiki]]{{·}} <br />
[[Communpedia]]{{·}} <br />
[[EditThis]]{{·}} <br />
[[elwiki.com]]{{·}} <br />
[[Miraheze]]{{·}} <br />
[[Neoseeker.com]]{{·}} <br />
[[Orain]]{{·}} <br />
[[Referata]]{{·}} <br />
[[ScribbleWiki]]{{·}} <br />
[[ShoutWiki]]{{·}} <br />
[[Sourceforge]]{{·}} <br />
[[TropicalWikis]]{{·}} <br />
[[Wik.is]]{{·}} <br />
[[Wiki.Wiki]]{{·}} <br />
[[Wiki-Site]]{{·}} <br />
[[Wikia]]{{·}} <br />
[[Wikidot]]{{·}} <br />
[[WikiHub]]{{·}} <br />
[[Wikispaces]]{{·}} <br />
[[Wikkii]]{{·}} <br />
[[YourWiki.net]]<br />
<br />
|group3=[[Wikimedia]]<br />
|list3=<br />
[[Wikipedia]]{{·}} <br />
[[Wikimedia Commons]]{{·}} <br />
[[Wikibooks]]{{·}} <br />
[[Wikidata]]{{·}} <br />
[[Wikinews]]{{·}} <br />
[[Wikiquote]]{{·}} <br />
[[Wikisource]]{{·}} <br />
[[Wikispecies]]{{·}} <br />
[[Wiktionary]]{{·}} <br />
[[Wikiversity]]{{·}} <br />
[[Wikivoyage]]{{·}}<br />
[[Wikimedia Incubator]]{{·}}<br />
[[Meta-Wiki]]<br />
<br />
|group4=Other<br />
|list4=<br />
[[Anarchopedia]]{{·}} <br />
[[Citizendium]]{{·}} <br />
[[Conservapedia]]{{·}} <br />
[[Creation Wiki]]{{·}} <br />
[[EcuRed]]{{·}} <br />
[[Enciclopedia Libre Universal en Español]]{{·}} <br />
[[GNUPedia]]{{·}} <br />
[[Moegirlpedia]]{{·}} <br />
[[Niconico#Nico Nico Pedia|Nico Nico Pedia]]{{·}} <br />
[[Nupedia]]{{·}} <br />
[[OmegaWiki]]{{·}} <br />
[[OpenStreetMap]]{{·}} <br />
[[Pixiv#Pixiv Encyclopedia|Pixiv Encyclopedia]]<br />
<br />
|group5=Indexes<br />
|list5=<br />
[[WikiApiary]]{{·}} <br />
[[WikiIndex]]<br />
<br />
|end=<br />
}}[[Category:Wikis]]<noinclude>[[Category:Templates]]</noinclude></div>Systwihttps://wiki.archiveteam.org/index.php?title=Why_Back_Up%3F&diff=48541Why Back Up?2022-05-04T06:54:33Z<p>Systwi: Added slideshow GIF that briefly summarizes the topic.</p>
<hr />
<div>[[Image:Whybackup.gif|right]]<br />
<br />
== Because They Don't Care About You ==<br />
<br />
Technology is business, and in business, money often takes priority over ethics. With the law perpetually lagging behind technology and business, there is often nothing between you and getting screwed over besides your own volition.<br />
<br />
Corporations do not contemplate their own inevitable end. At least, they don't do it in public, unless they are in really bad shape. When times are good, those thoughts are pushed away, and end users are encouraged to do the same. When times are bad, they tend to go '''very''' bad, '''very''' quickly - if you're lucky, you'll have an announcement. '''Your data is never totally safe. Backing up your data is always necessary, even if it's stored elsewhere.'''<br />
<br />
Some operators of user-generated content sites may have the dignity to inform contributors and visitors through a site notice well in advance of a possible shutdown, ideally at least three months before a planned shutdown, so users have appropriate time to save what they wish. Site operators may release users' text contributions as a compressed archive, since text compresses efficiently, and may send users who have provided an email address an export of their account data. However, as seen through numerous examples in Internet history, sites can shut down spontaneously with little indication.<br />
<br />
Therefore, consider any content on sites without a proper backup plan or shutdown plan to be on borrowed time, even sites which appear to have operated well for long, as [[Yahoo Answers]] (†2021) and [[LiveLeak]] (†2021) have shown. Larger sites, while less likely to shut down as a whole, may happen to spontaneously purge sections with little to no announcement, as can be seen with [[Reddit|subreddit]] bans, [[YouTube]] annotations (†2019), and their purge of unlisted videos from before 2017 (†2021).<br />
<br />
== Disaster Will Strike ==<br />
<br />
Entropy will rear its head, if you leave things up to chance. You will lose 4 years worth of email, including communications from the early days of your marriage, and the receipt to that flat panel monitor you bought online that now has tons of dead pixels. It doesn't take much for catastrophic data loss; one absent-minded mistake, or the ravages of time, can wipe out years of data.<br />
<br />
There is real convenience to web services like [[Google]] Apps. It's tempting to get wrapped up in that convenience and never take a step outside of it. With a little work, convenience can be evenly matched with user control and agency. <br />
<br />
Businesses can be extremely helpful, but they are also self-interested. As benevolent as Web services present themselves to be, your data is valuable to them - they aren't running this for your benefit. And it should be valuable to ''you'', too.<br />
<br />
== Get rid of your worries. ==<br />
Just like you wear a helmet while riding a bicycle or motorcycle, or wear seat belts while driving, even if data loss is not likely, having a backup solely for the reason of getting rid of data loss worries that bother you, is worthed it.<br />
<br />
[[File:Backupyourdata.gif|center]]<br />
<br />
== But There Is Still Hope ==<br />
<br />
Luckily, '''[[Introduction|a few basic (and cheap) precautions]]''' can bring the long-term care of your data into your own hands, away from the short-term world of the Internet.<br />
<br />
In information theory, entropy is often defined as "the loss of useful information." [http://books.google.com/books?id=oUnzaEDO1yYC&pg=PA185&lpg=PA185&dq=entropy&source=web&ots=abZfwAl4lN&sig=whhqna8q-c6amYq_H_DBCogr2qs&hl=en&sa=X&oi=book_result&resnum=8&ct=result] In historical research, all information is useful. For the historical benefit of future generations it is essential to organize a concentrated effort against information entropy. <br />
<br />
Despite our best efforts, some information will always be lost, but what is saved may help form the foundations of future cultures.<br />
<br />
== See also ==<br />
* [[YouTube#Why_download?]]<br />
<br />
{{Navigation pager<br />
| previous = Introduction<br />
| next = Software<br />
}}<br />
{{Navigation box}}<br />
<br />
[[Category:Archive Team]] [[Category:Backup]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=File_talk:Whybackup.gif&diff=48540File talk:Whybackup.gif2022-05-04T06:51:11Z<p>Systwi: Created discussion page with sizing notes.</p>
<hr />
<div>== Image Size Concerns ==<br />
Sorry if this image is a bit big. Resizing it any smaller than source makes the text difficult to read. --[[User:Systwi|Systwi]] ([[User talk:Systwi|talk]]) 06:50, 4 May 2022 (UTC)</div>Systwihttps://wiki.archiveteam.org/index.php?title=Clown_hosting&diff=48539Clown hosting2022-05-04T06:43:57Z<p>Systwi: Added comical photo (and caption) relevant to the topic.</p>
<hr />
<div>[[Image:Rmscloud.jpg|right|frame|Stallman doesn't like cloud computing, AND NEITHER SHOULD YOU!]]<br />
<br />
== Essential Information ==<br />
When adding an entry below please specify:<br />
* Plan name<br />
** the cost per hour or per month<br />
** RAM<br />
** Hard drive space, partition type (ext3, ext4, jfs, etc...)<br />
** CPUs/cores<br />
** bandwidth per month<br />
** operating system options<br />
** Free trial options<br />
<br />
== Hosting Providers ==<br />
<br />
* [https://www.gandi.net/hosting/simple gandi.net] - $5 a month for a small instance, 30gb traffic<br />
** GANDI also provides VPS with unmetered 10Mbps link and I can vouch that using at least 2/3 of this 24/7 gets no complaints. Billing is a big pain in the ass though, there's no way to auto-bill so you have to remember to fill your prepaid account. [[User:Troofmaster|Troofmaster]] 14:57, 24 March 2013 (EDT)<br />
* Amazon EC2 - [https://aws.amazon.com/ec2/pricing/#spot pricing]<br />
* BandwagonHost - '''Avoid like the plague.''' Host seems to consider sales more important than user safety; response to inquiry about lack of SSL in the customer panel was "''Very few people actually care about it. We did all kinds of experiments on our websites throughout the years, and SSL is one of those things that does not actually impact the number of new orders received.''" ([http://www.lowendbox.com/blog/bandwagonhost-1-99month-512mb-openvz-vps-in-phoenix/#comment-111493 source]). <br />
<br />
== [https://www.digitalocean.com/ DigitalOcean] ==<br />
<br />
* Smallest plan<br />
** $0.007 per hour<br />
** 512 MB RAM<br />
** 20 GB storage<br />
** 1 cpu core<br />
** 1 TB transfer<br />
<br />
* Most Popular Plan<br />
** $0.015 per hour<br />
** 1 GB RAM<br />
** 30 GB storage<br />
** 1 cpu core<br />
** 2 TB transfer<br />
<br />
* These plans have no names<br />
** $0.030 per hour<br />
** 2 GB RAM<br />
** 40 GB storage<br />
** 2 cpu cores<br />
** 3 TB transfer<br />
<br />
* These plans have no names<br />
** $0.119 per hour<br />
** 8 GB RAM<br />
** 80 GB storage<br />
** 4 cpu cores<br />
** transfer<br />
<br />
* Unmetered BW<br />
* Fedora 16-17, CentOS 5.8-6.3, Ubuntu 10.04 - 12.10, Debian 6.0, Arch Linux 2013.03<br />
* Coupon code "SSDTWEET" gives $11.15 account credit<br />
* Supports launching new instances from an image<br />
* Submit a support ticket to lift the standard 5 instance limit<br />
* API can launch instances<br />
<br />
=== Problems ===<br />
* Cannot resize a droplet unless it is powered down.<br />
* Resize options only include CPU and RAM. No way to increase base storage. Additional Disks can be attached.<br />
<br />
== [http://joyent.com/products/joyent-cloud/pricing Joyent] ==<br />
I've had good experiences running high-traffic, low-CPU jobs at Joyent due to the incredible 20TB/mo included (with a single account, last I checked, not per-instance). [[User:Troofmaster|Troofmaster]] 14:57, 24 March 2013 (EDT)<br />
''(Note Joyent has changed the free bandwidth quote.)''<br />
<br />
Been running an extra small instance for a few weeks. No uptime or connectivity issues. A decent service but the problems listed below make this a second not a first choice. [[User:DukeNukem|omf_]] 0:44, 27 March 2013 (EDT)<br />
<br />
* Extra small Linux instance<br />
** $0.03 per hour<br />
** 512 MB RAM<br />
** /dev/vda1 9 GB ext4 partition, minus 750mb for the operating system<br />
** /dev/vdb1 15 GB ext4 partition. mounted as /data/<br />
** 1 cpu<br />
* Small Linux instance<br />
** $0.085 per hour<br />
** 1 GB RAM<br />
** 30 GB storage<br />
** 1 cpu<br />
* '''free bandwidth per month: All in, 1 TB out'''<br />
* Debian 6.x and Centos<br />
<br />
=== Problems ===<br />
<br />
* No support for installing custom Linux distributions. Meaning you have to install everything you need every single time since you always start off with one of their fresh installs.<br />
* Cannot find a way to add more hard disk and RAM to a running instance.<br />
* The wiki is currently down :(<br />
* Some scripts on there site are just broken links :(<br />
* No way to view your current account balance or hours usage.<br />
<br />
== [https://www.linode.com/ linode] ==<br />
<br />
* Linode 1G<br />
** $19.95 per month<br />
** 1 GB RAM<br />
** 24 GB storage<br />
** 2 TB per month<br />
* Linode 2G<br />
* Linux<br />
** $39.95 per month<br />
** 2 GB RAM<br />
** 48 GB storage<br />
** 4 TB per month<br />
<br />
Linode upgraded all their plans on April 9, 2013. http://blog.linode.com/2013/04/09/linode-nextgen-ram-upgrade/<br />
<br />
== [http://buyvm.net/ BuyVM] (VPS) ==<br />
<br />
* OpenVZ plan<br />
** $15/year<br />
** 128 MB RAM (burst to 256 MB)<br />
** 15 GB storage<br />
** 1 CPU<br />
** 500 GB per month bandwidth<br />
<br />
* CentOS/Debian/Fedora/Ubuntu<br />
* OpenVZ means you can't change the kernel<br />
* Burstable RAM is available as long as there is capacity on the host node, which seems to be almost all the time<br />
* Additional IP v4 addresses $1/month, but they probably won't let you cycle them to avoid bans<br />
* Additional bandwidth $2.50/month for 1TB extra with no maximum<br />
* Larger plans also available with similar pricing<br />
* Plans often go out of stock, especially the smaller ones<br />
* IRC support and information about stock in [irc://irc.dairc.net/frantech #Frantech on irc.dairc.net]<br />
<br />
== [http://www.edis.at/en/home/ EDIS.at] ==<br />
I've so far had good experiences running Tor relays and other small jobs on the Micro VPSes with EDIS. [[User:Troofmaster|Troofmaster]] 14:57, 24 March 2013 (EDT)<br />
<br />
* KVM VPS in many countries, including Iceland, Isle of Man ('offshore'), Sweden, Switzerland, Spain for those concerned about local privacy laws<br />
* Micro<br />
** €1.99 per month, yearly<br />
** 128 MB RAM<br />
** 1 GB storage (was plenty to run [[Yahoo Messages]] grab scripts)<br />
** 500 GB per month traffic<br />
<br />
* Smart<br />
** €3.99 per month, yearly<br />
** 256 MB RAM<br />
** 5 GB storage<br />
** 1 TB per month traffic<br />
<br />
* It also has OpenVZ plans:<br />
* Micro<br />
** €1.45 per month, yearly<br />
** 128 MB RAM<br />
** 2 GB storage<br />
** 500 GB per month traffic<br />
* Archlinux, CentOS 5/6, Debian 5/6/7, Elastix, Fedora 17/18, FreeBSD, Gentoo, GRML, MikroTik, NetBSD, OpenBSD, openSUSE, Ubuntu, Windows Server 2008/2012, Zeroshell<br />
<br />
<br />
{{Navigation box}}</div>Systwihttps://wiki.archiveteam.org/index.php?title=File:Rmscloud.jpg&diff=48525File:Rmscloud.jpg2022-04-30T04:49:12Z<p>Systwi: Stallman doesn't like clown computing, AND NEITHER SHOULD YOU!</p>
<hr />
<div>== Summary ==<br />
Stallman doesn't like clown computing, AND NEITHER SHOULD YOU!</div>Systwihttps://wiki.archiveteam.org/index.php?title=File:Whybackup-icon.png&diff=48524File:Whybackup-icon.png2022-04-30T03:57:33Z<p>Systwi: Finder icon for "Why Back Up."</p>
<hr />
<div>== Summary ==<br />
Finder icon for "Why Back Up."</div>Systwihttps://wiki.archiveteam.org/index.php?title=File:Whybackup.gif&diff=48523File:Whybackup.gif2022-04-30T03:33:42Z<p>Systwi: Lossless screenshots from "Why Back Up," a mini slideshow from 1994 included as an extra on the restore CD for the Apple Macintosh Performa 630CD.</p>
<hr />
<div>== Summary ==<br />
Lossless screenshots from "Why Back Up," a mini slideshow from 1994 included as an extra on the restore CD for the Apple Macintosh Performa 630CD.</div>Systwihttps://wiki.archiveteam.org/index.php?title=Template:Wikis&diff=48522Template:Wikis2022-04-30T17:01:30Z<p>Systwi: Added two more wiki software.</p>
<hr />
<div>{{navbox<br />
|name=Template:Wikis<br />
|status={{{status|autocollapse}}}<br />
|title=Knowledge and Wikis<br />
<br />
|group1=Software<br />
|list1=<br />
[[DokuWiki]]{{·}}<br />
[[MediaWiki]]{{·}}<br />
[[PukiWiki]]{{·}}<br />
[[YukiWiki]]<br />
<br />
|group2=[[WikiTeam#Wikifarms|Wikifarms]]<br />
|list2=<br />
[[atwiki]]{{·}} <br />
[[Battlestar Wiki]]{{·}} <br />
[[BluWiki]]{{·}} <br />
[[Communpedia]]{{·}} <br />
[[EditThis]]{{·}} <br />
[[elwiki.com]]{{·}} <br />
[[Miraheze]]{{·}} <br />
[[Neoseeker.com]]{{·}} <br />
[[Orain]]{{·}} <br />
[[Referata]]{{·}} <br />
[[ScribbleWiki]]{{·}} <br />
[[ShoutWiki]]{{·}} <br />
[[Sourceforge]]{{·}} <br />
[[TropicalWikis]]{{·}} <br />
[[Wik.is]]{{·}} <br />
[[Wiki.Wiki]]{{·}} <br />
[[Wiki-Site]]{{·}} <br />
[[Wikia]]{{·}} <br />
[[Wikidot]]{{·}} <br />
[[WikiHub]]{{·}} <br />
[[Wikispaces]]{{·}} <br />
[[Wikkii]]{{·}} <br />
[[YourWiki.net]]<br />
<br />
|group3=[[Wikimedia]]<br />
|list3=<br />
[[Wikipedia]]{{·}} <br />
[[Wikimedia Commons]]{{·}} <br />
[[Wikibooks]]{{·}} <br />
[[Wikidata]]{{·}} <br />
[[Wikinews]]{{·}} <br />
[[Wikiquote]]{{·}} <br />
[[Wikisource]]{{·}} <br />
[[Wikispecies]]{{·}} <br />
[[Wiktionary]]{{·}} <br />
[[Wikiversity]]{{·}} <br />
[[Wikivoyage]]{{·}}<br />
[[Wikimedia Incubator]]{{·}}<br />
[[Meta-Wiki]]<br />
<br />
|group4=Other<br />
|list4=<br />
[[Anarchopedia]]{{·}} <br />
[[Citizendium]]{{·}} <br />
[[Conservapedia]]{{·}} <br />
[[Creation Wiki]]{{·}} <br />
[[EcuRed]]{{·}} <br />
[[Enciclopedia Libre Universal en Español]]{{·}} <br />
[[GNUPedia]]{{·}} <br />
[[Moegirlpedia]]{{·}} <br />
[[Niconico#Nico Nico Pedia|Nico Nico Pedia]]{{·}} <br />
[[Nupedia]]{{·}} <br />
[[OmegaWiki]]{{·}} <br />
[[OpenStreetMap]]{{·}} <br />
[[Pixiv#Pixiv Encyclopedia|Pixiv Encyclopedia]]<br />
<br />
|group5=Indexes<br />
|list5=<br />
[[WikiApiary]]{{·}} <br />
[[WikiIndex]]<br />
<br />
|end=<br />
}}[[Category:Wikis]]<noinclude>[[Category:Templates]]</noinclude></div>Systwihttps://wiki.archiveteam.org/index.php?title=Talk:YouTube&diff=48521Talk:YouTube2022-04-30T04:10:54Z<p>Systwi: /* Saving MORE stuff from YouTube? */ Further discoveries on getting original image sizes.</p>
<hr />
<div>==youtube2internearchive==<br />
https://github.com/emijrp/youtube2internetarchive contains a script which also handled upload to Internet Archive, but I can't find it any longer. --[[User:Nemo_bis|Nemo]] 06:28, 26 January 2015 (EST)<br />
:I've found [https://code.google.com/p/emijrp/source/browse/trunk/scrapers/youtube2internetarchive.py something] with Google. [[User:bzc6p|bzc6p]] ([[User_talk:bzc6p|talk]]) 12:25, 26 January 2015 (EST)<br />
<br />
If YouTube needed to be quickly captured for some unforeseen reason, it might make sense to download only the XML and SRT files, so then at least some record would be saved. Google's subtitle recognition is currently far from accurate, but it's certainly improving. [[User:wtron|wtron]] 06:48, 12 June 2015 (EST)<br />
<br />
== Options ==<br />
<br />
Is it really necessary to explicitly call best A/V when youtube-dl it by default? <br />
<br />
Also, why not embed the subs and thumbnail instead of make a separate file? Also why not xattrs for those of us with unix filesystems? Xattrs is only one extra flag. <br />
<br />
My command is currently <br />
<br />
youtube-dl -t --embed-subs --add-metadata --xattrs --console-title --embed-thumbnails<br />
<br />
although I'm going to be incorporating elements from the suggested one into mine. The reasoning behind this is it's one file to send. That command is how I archive currently, it's changing though.<br />
<br />
I'd appreciate hearing your input about why I may be wrong though. Thanks in advance,<br />
<br />
--[[User:Vxbinaca|Vxbinaca]] 21:24, 29 May 2015 (EDT)<br />
<br />
On your second note, I strongly believe it's better to have different things (video, thumbnail, subtitle) in separate files. Easier to access, process, categorize, recognize. I think it's worth the "trouble" of having three files (with the same name) instead of one.<br />
<br />
[[User:bzc6p|bzc6p]] ([[User_talk:bzc6p|talk]]) 07:08, 31 May 2015 (EDT)<br />
<br />
xattrs are not portable and will get lost when copying to a file system that doesn't have them (or when uploading it somewhere, like to IA) --[[User:Darkstar|Darkstar]] 08:53, 31 May 2015 (EDT)<br />
<br />
Solid reasoning. I've now switched to your way of doing things. --[[User:Vxbinaca|Vxbinaca]] 19:32, 2 August 2015 (EDT)<br />
<br />
== --ignore-errors shouldn't be youtube-dl archiving default best practices ==<br />
<br />
Theres a myriad of reasons this isn't a good idea to have by default. Downloads getting snapped off on channel rips could go unnoticed (I search for these with ls *.part). Problems with various versions of youtube-dl could lead to a channel rip with half-processed videos, see [https://github.com/rg3/youtube-dl/issues/6413 this issue on github].<br />
<br />
Perhaps for a well-tested version that works on YouTube running in a warrior, --ignore-errors is appropriate, but for an attended rip we should by default suggest people not use it and instead just make sure all of it got ripped and if theres an error try to work resolve that particular video, and if it's a problem they can't get around then just go --ignore-errors.<br />
<br />
I'm open to being told why I may be wrong though. --[[User:Vxbinaca|Vxbinaca]] 19:32, 2 August 2015 (EDT)<br />
<br />
== Paid subscription ==<br />
<br />
On September 1, 2015 [https://www.theverge.com/2015/4/8/8371131/youtube-paid-subscription-offline-video The Verge reported] about oncoming paid subscription option (ad-free+premium videos), although paid ad-free videos may not be a nail in the coffin, previously free content could become a premium. According to the article, transition could happen in a few months -- [[User:Vitzli|Vitzli]] 06:46, 2 September 2015 (EDT)<br />
<br />
YouTube has never been profitable. [[User:bzc6p|bzc6p]] ([[User_talk:bzc6p|talk]]) 08:46, 2 September 2015 (EDT)<br />
<br />
== As we don't have 500 petabytes of free storage... ==<br />
<br />
... a solution may be to discard "low-value" videos. I mean, if we discard duplicates (films, music, etc) and we set a limit of 1 PB (1000 TB) for quality content, what could be the lucky videos that would be downloaded and preserved? We can work in a approach like this. Just because we don't have space for all, doesn't mean that we don't download anything. [[User:Emijrp|Emijrp]] 09:33, 20 October 2015 (EDT)<br />
<br />
Google will probably give a notice long before that date. So we'll have the time to find stuff worth saving. We could create a "warroom" or such, and users (ArchiveTeam members and other people) could suggest channels worth saving, with description, average views and size estimate, say, in a table, clear-cut. If someone suggests videos that are otherwise available or not too popular, they can be striked out (with proper reasoning). (The reviewing of suggestions can be done by everyone continously.) A deadline of suggestions could be given (say, 2 months before the end), and after that a Committee could select the "lucky" 1000 TB that could end up in the Archive.<br />
* In the meanwhile, the Archive itself would be queried for already saved videos, and those wouldn't be saved again.<br />
* The Archive and the Team should expect other preserving actions, and ours should be in accordance with those (no duplicates).<br />
* There could be national limits, e.g. not only a global limit, but also language or country-specific, say, 500 TB English videos, and 10–50 terabytes per other countries (just ad-hoc numbers, see the concept).<br />
<br />
The importance of the last point and the importance of saving some of YouTube at all is, I think, in the fact that, without too much exaggaration, a substantial part of today's culture is stored and represented there, on the most popular video sharing site on Earth. [[User:bzc6p|bzc6p]] ([[User_talk:bzc6p|talk]]) 15:41, 20 October 2015 (EDT)<br />
<br />
: We could also save all videos starting at a specific view count such as 500 or 1000 and/or from channels above a subscriber threshold such as 200. But maybe better first start from e.g. 1000000 views and 10000 subscribers. --[[User:ATrescue|ATrescue]] ([[User talk:ATrescue|talk]]) 22:29, 13 May 2019 (UTC)<br />
<br />
== Saving YouTube Comments ==<br />
<br />
YouTube comments are a surefire sign of just how awful the internet can be at times. Shouldn't they be archived as well? There's already a script for it, [https://github.com/egbertbouman/youtube-comment-downloader youtube-comment-downloader]. --[[User:Powerkitten|Powerkitten]] ([[User talk:Powerkitten|talk]]) 16:49, 26 October 2016 (EDT)<br />
<br />
: Hello, [[User:Powerkitten|Powerkitten]]. Thank you for mentioning this.<br />
<br />
There is also [https://github.com/philbot9/youtube-comment-scraper-cli youtube-comment-scraper-cli] which can do this as well and supports .csv and .json outputs. This tool logs comment ID, channel name, like count, post date (in both Unix timestamp and UTC date format), user profile picture and number of replies. An online version is also available [http://ytcomments.klostermann.ca/ here]. [[User:Systwi|Systwi]] ([[User talk:Systwi|talk]]) 06:53, 21 July 2019 (UTC)<br />
<br />
<br />
Now, there is some information about comment archival here:<br />
<br />
* [[YouTube#Comment_loading]]<br />
* [[Archive.is#YouTube_comment_archival]]<br />
=== Celebrations ===<br />
When the great wall of YTimg's [[robots.txt]] fell on 20120229 (leap year's day of 2012), it was a prideful feeling. Until then, http://ythistory.weebly.com/ and http://ythistory.co.cc/ were the only places where one could access the YouTube swf (flash-based) ActionScript 2 and ActionScript-3-players between circa 2008 (when YouTube's YTimg domain started, likely due to Google's acquisition) and 2012. Additionally, when browsing YouTube via the Wayback Machine, stylesheet information and all images were also blocked by [[robots.txt]], which was suddenly gone.<br />
<br />
For a few hours after it's removal, there was an error stating “Couldn't load Robots.txt for this page”, “Unable to read Robots.txt” or “Couldn't find robots.txt for this site” or something similar (I don't remember the exact words), because it was assuming that the robots.txt was just temporarily inaccessible. After these few hours past, suddently, all the YouTube pages blocked by robots.txt were suddently visible in their full beauty.<br />
<br />
Now, because one of the biggest Wayback dreams I and {{URL|2=other people|1=https://archive.org/post/406632/why-does-the-wayback-machine-pay-attention-to-robotstxt}} had apparently has come true (read more at [[Internet_Archive#robots.txt_and_the_Wayback_Machine]] too), I can now load YouTube comments through the Wayback Machine! ([http://web.archive.org/web/20190512223447/https://www.youtube.com/watch?v=W3GrSMYbkBE&lc=UgxC238Gea0KGOditl54AaABAg Try here]) --[[User:ATrescue|ATrescue]] ([[User talk:ATrescue|talk]]) 22:52, 12 May 2019 (UTC)<br />
<br />
: @Powerkitten: http://ytcomments.klostermann.ca/ (title element={YouTube Comment Scraper}; title on the webpage={Download comments from YouTube};) can be used to get all the comments on a video, which can be downloaded as JSON and CSV at http://ytcomments.klostermann.ca/scrape. You can then upload those files to archive.org. --[[User:Usernam|Usernam]] ([[User talk:Usernam|talk]]) 10:57, 20 May 2019 (UTC)<br />
<br />
: archive.is can also archive replies to comments, given you have the link to a reply to a comment. --[[User:Usernam|Usernam]] ([[User talk:Usernam|talk]]) 00:17, 15 July 2019 (UTC)<br />
<br />
== Many YouTube video items are archived in low quality because they were archived using an old version of tubeup. Is there a way I can replace the files of old youtube items (using the "youtube-[id]" naming format the tubeup uses) with updated, high quality files? --[[User:Hiccup|Hiccup]] ([[User talk:Hiccup|talk]]) 17:01, 6 February 2018 (UTC) ==<br />
<br />
Many YouTube video items are archived in low quality because they were archived using an old version of tubeup. Is there a way I can replace the files of old youtube items (using the "youtube-[id]" naming format the tubeup uses) with updated, high quality files? --[[User:Hiccup|Hiccup]] ([[User talk:Hiccup|talk]]) 17:01, 6 February 2018 (UTC)<br />
: Let's say that the youtube-TFkwwyRqXRA identifier has already been taken. Unless you uploaded youtube-TFkwwyRqXRA or are an admin at archive.org, youtube-TFkwwyRqXRA is read only for you. If it is a small channel, or if you feel up to downloading all the videos of a medium or large-amount-of-videos channel, then just download all the videos and upload them at youtube_[channel_ID]. Examples: youtube_UCPycIJJaP8rv0XTfioIpcOA, youtube_UCfLuMSIDmeWRYpuCQL0OJ6A --[[User:Usernam|Usernam]] ([[User talk:Usernam|talk]]) 12:59, 27 July 2020 (UTC)<br />
<br />
== Thumbnail rescue ==<br />
<br />
Some YouTube uploaders change thumbnails of their videos, which would lead to the loss of the existing thumbnail if archived nowhere else.<br />
<br />
Some YouTube thumbnails can be retrieved by searching the video ID using web search (try different search engine).<br />
<br />
To save as many thumbnails as possible from a specific channel, please archive the video page of their channel using the help of [[chromebot]]'s page scrolling.<br />
<br />
One can also manually scroll down a page and save it using the web browser, then use the “<code>sed</code>” command to extract thumbnail URL's from the HTML page source code.<br />
<br />
: (command will be added here.)<br />
<br />
With <s>the help of {{W2|regular expressions}}</s> simple text replacement, you can put all “mqdefault” and “hqdefault” URL's into “maxresdefault”, although “maxresdefault” is somehow not available for all videos, thus better feed both hqdefaults and maxresdefaults into ArchiveBot.<br />
<br />
Then, you can upload them to https://transfer.sh/ or https://transfer.notkiska.pw/ and feed them into [[ArchiveBot]] using the [https://archivebot.readthedocs.io/en/latest/commands.html#archiveonly-file !ao <file] command. --[[User:ATrescue|ATrescue]] ([[User talk:ATrescue|talk]]) 13:27, 13 May 2019 (UTC)<br />
<br />
== Mass Deletion in Internet Archive??==<br />
There was more than 300,000 videos in https:// archive.org/details/archiveteam_youtube . Now just 12,000 videos, are there a mass deletion?? --[[User:Gridkr|Gridkr]] ([[User talk:Gridkr|talk]]) 14:05, 18 July 2019 (UTC)<br />
<br />
: We are not the Internet Archive. But no, they haven't been deleted, only delisted. [[User:JustAnotherArchivist|JustAnotherArchivist]] ([[User talk:JustAnotherArchivist|talk]]) 16:39, 18 July 2019 (UTC)<br />
<br />
: They are still accessible via the URLs just not through search any more. Flashfire42<br />
<br />
==Saving MORE stuff from YouTube?==<br />
<br />
On the topic of saving stuff from YouTube, are there any methods available for scraping more content such as a user's profile picture, channel banners, channel description, channel discussion page and video watermarks? I was thinking for things such as their profile picture it would be as simple as using wget, however I don't know if the url I am grabbing (such as [https://yt3.ggpht.com/a/AGF-l7_GPnSqhtxkf5pgDj4jdL3EfgJkG09iAXg3Og this one]) is the largest resolution available, especially since the URLs don't seem to follow a specific format. --[[User:Systwi|Systwi]] ([[User talk:Systwi|talk]]) 18:38, 21 July 2019 (UTC)<br />
<br />
: I've discovered that appending <code>=w0</code> to a channel banner's/profile picture's URL, such as the one above ([https://yt3.ggpht.com/a/AGF-l7_GPnSqhtxkf5pgDj4jdL3EfgJkG09iAXg3Og=w0 like this]), appears to return the originally-uploaded image. This is unconfirmed by anyone else, but try it out! It seems to me like this is the case. Remove or change any other URL parameters before appending (primarily ones with <code>=w<digit(s)></code>). --[[User:Systwi|Systwi]] ([[User talk:Systwi|talk]]) 04:10, 30 April 2022 (UTC)</div>Systwihttps://wiki.archiveteam.org/index.php?title=Google_Drive&diff=47093Google Drive2021-09-01T02:02:54Z<p>Systwi: Changed "...an URL..." to "...a URL..."; changed IRC channel and server info.</p>
<hr />
<div>{{Infobox project<br />
| title = Google Drive<br />
| image = Google drive logo.png<br />
| URL = {{url|http://drive.google.com/}}<br />
| project_status = {{specialcase}} (Technically files are only becoming inaccessible)<br />
| archiving_status = {{upcoming}}<br />
| irc_network = hackint<br />
| irc = googlecrash<br />
}}<br />
<br />
'''Google Drive''' is a filehosting service, a la Dropbox, run by Google (not to be confused with [https://cloud.google.com/ Google Cloud Storage] and similar more technical storage solutions). It is popular both for personal storage and for sharing of files.<br />
<br />
Google Drive IDs are not random (anecdotally, IDs of folders in the same tree often share long parts), which makes them predictable, a problem which Google has been trying to rectify across its products (others of which have similar issues) throughout 2021<ref>https://www.theverge.com/google/2021/6/23/22548007/youtube-google-drive-security-link-sharing</ref>. As such, on September 13, 2021, Google will be requiring that, in order to access files and folders, users either have permissions tied to their signed-in Google Accounts, or access the item through a URL with a random per-item parameter called resourceKey, apparently introduced in 2021.<ref>https://support.google.com/a/answer/10685032</ref> The result of this will be that at least millions of links across the Web will effectively break.<br />
<br />
== Grab ==<br />
The upcoming grab script has 3 item types, folder:, file:, and user:. It is intended that all folder: items be run first, to get a pool (through backfeed) of file: items, that can be randomly sampled to determine a size threshold that the Internet Archive will accept; then files will be run. Users contain some user metadata but not links to files or folders.<br />
<br />
Playback is theoretically possible with a flexible, POST-capable Wayback Machine, but this does not yet exist. In the meantime, it may be possible to get files with vanilla wget or similar from the WBM, and hopefully a tool to list folders can be created as well.</div>Systwihttps://wiki.archiveteam.org/index.php?title=TED_Talks&diff=45501TED Talks2020-09-18T12:56:27Z<p>Systwi: Marked TPB torrent link rot.</p>
<hr />
<div>{{Infobox project<br />
| title = TED Talks<br />
| image = Ted.png<br />
| description = TED mainpage in 2010-12-22<br />
| URL = http://www.ted.com<br />
| project_status = {{online}}<br />
| archiving_status = {{notsaved}} (TPB torrent is gone)<br />
}}<br />
<br />
'''TED Talks''' is a website that provides free talks from famous figures.<br />
<br />
== Backups ==<br />
<br />
* [http://thepiratebay.org/torrent/5420151 All TED talks in mp4, well organized] 35 GB - {{offline}} as of 2020-09-18 12:50.26 UTC<br />
<br />
== Vital signs ==<br />
<br />
Alive and well.<br />
<br />
== External links ==<br />
<br />
* http://www.ted.com<br />
<br />
{{Navigation box}}<br />
<br />
[[Category:Video hosting]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=Valhalla&diff=45479Valhalla2020-09-12T07:12:58Z<p>Systwi: /* Physical Options */ Added ODA cartridges to the table.</p>
<hr />
<div>[[Image:Ms internet on a disc.jpg|300px|right]]<br />
This wiki page is a collection of ideas for Project '''Valhalla'''.<br />
<br />
This project/discussion has come around because there is a class of data currently existing, several times a year, as a massive amount of data with "large, but nominal" status within the Internet Archive. The largest example is currently MobileMe, which is hundreds of terabytes in the Internet Archive system (and in need of WARC conversion), which represents a cost amount far outstripping its use. Another is TwitPic, which is currently available (and might continue to be available) but which has shown itself to be a bad actor with regards to longevity and predictability for its sunset. <br />
<br />
Therefore, there is an argument that there could be a "third place" that data collected by Archive Team could sit, until the Internet Archive (or another entity) grows its coffers/storage enough that 80-100tb is "no big deal", just like 1tb of data was annoying in 2009 and now is totally understandable for the value, i.e. Geocities. <br />
<br />
This is for short-term (or potentially also long-term) storage options, say five years or less, of data generated by Archive Team.<br />
<br />
* What options are out there, generally?<br />
* What are the costs, roughly?<br />
* What are the positives and negatives?<br />
<br />
There has been a lot of study in this area over the years, of course, so links to known authorities and debates will be welcome as well.<br />
<br />
Join the discussion in [irc://irc.efnet.org/huntinggrounds #huntinggrounds].<br />
<br />
== Goals ==<br />
<br />
We want to:<br />
<br />
* Dump an unlimited<ref>Unlimited doesn't mean infinite, but it does mean that we shouldn't worry about running out of space. We won't be the only expanding data store.</ref> amount of data into something.<br />
* Recover that data at any point.<br />
<br />
We do not care about:<br />
<br />
* Immediate or continuous availability.<br />
<br />
We absolutely require:<br />
<br />
* Low (ideally, zero) human time for maintenance. If we have substantial human maintenance needs, we're probably going to need a Committee of Elders or something.<br />
* Data integrity. The storage medium must be impossibly durable or make it inexpensive/easy to copy and verify the data onto a fresh medium.<br />
<br />
It would be nice to have:<br />
<br />
* No special environmental requirements that could not be handled by a third party. (So nobody in Archive Team would have to set up some sort of climate-controlled data-cave; however, if this is already something that e.g. IA does and they are willing to lease space, that's cool.)<br />
<br />
== What does the Internet Archive do for this Situation, Anyway? ==<br />
<br />
''This section has not been cleared by the Internet Archive, and so should be considered a rough sketch.''<br />
<br />
The Internet Archive primarily wants "access" to the data it stores, so the primary storage methodology is spinning hard drives connected to a high-speed connection from multiple locations. These hard drives are between 4-6tb (as of 2014) and are of general grade, as is most of the hardware - the theory is that replacing cheap hardware is better than spending a lot of money on super-grade hardware (whatever that may be) and not being able to make the dollars stretch. Hundreds of drives die in a month and the resiliency of the system allows them all to hot-swap in replacements. <br />
<br />
There are multiple warehouses for storing the original books that are scanned, as well as materials like CD-ROMs and even hard drives. There are collections of tapes and CD-ROMs from previous iterations of storage, although they are thought of as drop-dead options instead of long-term archival storage - the preference is, first and foremost, the spinning hard drives.<br />
<br />
The Archive does not generally use tape technology, having run into the classic "whoops, no tape drive on earth reads these any more" and "whoops, this tape no longer works properly".<br />
<br />
The Archive has indicated that if Archive Team uses a physical storage method, such as tapes, paper, hard drives or anything else, that they are willing to store these materials "as long as they are exceedingly labelled".<br />
<br />
== Physical Options ==<br />
{| class="wikitable sortable"<br />
! Storage type<br />
! Cost ($/TB/year)<br />
! Storage density (m³/TB)<br />
! Theoretical lifespan<br />
! Practical, tested lifespan<br />
! Notes<br />
|-<br />
| Hard drives (simple distributed pool)<br />
| $15 (external USB HD's on sale at 8TB to 10TB)<br />
| <br />
| <br />
| <br />
| December 2018, best reasonable 1TB+ external HD is [http://thewirecutter.com/reviews/the-best-external-desktop-hard-drive/ a 4TB WD]. 25+ pool members would need one HD each plus a computer plus software to distribute data across the entire pool.<br />
|-<br />
| Hard drives (dedicated distributed pool)<br />
| <br />
| <br />
| <br />
| <br />
| An off-the-shelf or otherwise specified, dedicated, network storage device used exclusively as part of a distributed pool.<br />
|-<br />
| Hard drives (SPOF) <ref>The [[Internet Archive]]'s cost per TB, with 24/7 online hard drives, is approximately $2000 for forever.</ref><br />
| $37 - $44 (but you have to buy 240TB - 480TB)<br />
| <br />
| <br />
| <br />
| For a single location to provide all storage needs, building a [https://www.backblaze.com/blog/open-source-data-storage-server/ Backblaze Storage Pod 6.0] runs an average of $10,364, providing 240TB of [http://bioteam.net/2011/08/why-you-should-never-build-a-backblaze-pod/ non-redundant, not-highly-available] storage. (You really want more than one pod mirroring your data, but this is the most effective way to get that much storage in one place.)<br />
|-<br />
| Commercial / archival-grade tapes<br />
| <br />
| <br />
| <br />
| <br />
| <br />
|-<br />
| Consumer tape systems (VHS, Betamax, cassette tapes, ...)<br />
| <br />
| <br />
| <br />
| <br />
| <br />
|-<br />
| Vinyl<br />
| <br />
| <br />
| <br />
| <br />
| <br />
|-<br />
| [http://www.ollydbg.de/Paperbak/index.html PaperBack]<br />
| <br />
| <br />
| <br />
| <br />
| 500KB per letter sheet means 1TB is 2,199,024 sheets, or ~4400 reams (500 sheets each), or an 8'x16' room filled with 6' tall stacks. It would take 63.6 days of continuous printing to do this.<ref>A HP LaserJet 5Si printing 24 pages per minute which generates the 500K bytes per page, yielding approximately 200,000 bytes per second.</ref><br />
|-<br />
| [http://ronja.twibright.com/optar/ Optar]<br />
| <br />
| <br />
| <br />
| <br />
| At 200KB per page, this has less than half the storage density of Paperback.<br />
|-<br />
|Archival DVDR<br />
|<br />
|<br />
|<br />
|<br />
|Reference Book<ref>[https://clir.wordpress.clir.org/wp-content/uploads/sites/6/pub121.pdf]</ref><br />
|-<br />
| Blu-Ray R (HTL only)<br />
| $16-$24 (using 50 pack spindle of 25GB BD-Rs)<br />
| <br />
| 30 years<ref>On the basis of the described studies and assuming adequate consideration of the specified conditions for storage and handling, as well as verification of data after writing, we estimate the Imation CD, DVD or Blu-ray media to have a theoretical readability of up to 30 years. The primary caveat is how you handle and store the media. http://support.tdkperformance.com/app/answers/detail/a_id/1685/~/life-expectancy-of-optical-media </ref><br />
| <br />
| Lasts a LOT longer than CD/DVD, but should not be assumed to last more than a decade. [http://arstechnica.com/information-technology/2014/01/why-facebook-thinks-blu-ray-discs-are-perfect-for-the-data-center/ Raidz3 with Blu-rays Doing a backup in groups of 15 disks]. Comes to under $.04/GB which is cheap, and low initial investment (drives) too!<br><br />
<br>Specifically, a 50pack spindle of 25GB BD-Rs could readily hold 1TB of data for $20-$33 per spindle. 50GB and 100GB discs are DL so carry different risk profile.<br />
|-<br />
| [http://en.wikipedia.org/wiki/M-DISC M-DISC]<br />
| <br />
| <br />
| <br />
| <br />
| Unproven technology, but potentially interesting.<br />Note that the main selling point of M-DISC DVDs – not using fast-detoriating organic dyes – is moot for BDs because those (with the exception of BD-R LTH) do not use such dyes. There are some indications that M-DISC BDs handle very extreme conditions slightly better than standard BDs, but the difference is much smaller than for DVDs.<br />
|-<br />
| [https://en.wikipedia.org/wiki/Optical_Disc_Archive Optical Disc Archive (ODA) cartridges]<br />
| <br />
| <br />
| <br />
| <br />
| <br />
|-<br />
| Flash media<br />
| <br />
| <br />
| <br />
| <br />
| Very durable for online use, and usually fails from lots of writes. A drive might never wear out from cold-storage usage. Newer drives can have 10-year warranties. But capacitors may leak charge over time. JEDEC JESD218A only specifies 101 weeks (almost two years) retention without power, so we'd have to check the spec of the specific drives, or power them up and re-write the data to refresh it about once a year. Soliciting donations for old flash media from people, or sponsorship from flash companies?<br />
|-<br />
| Glass/metal etching<br />
| <br />
| <br />
| <br />
| <br />
| <br />
|-<br />
| Amazon Glacier<br />
| $49.15 (storage only, retrieval billed separately)<br />
| <br />
| average annual durability of 99.999999999% <ref>"Amazon Glacier is designed to provide average annual durability of 99.999999999% for an archive. The service redundantly stores data in multiple facilities and on multiple devices within each facility. To increase durability, Amazon Glacier synchronously stores your data across multiple facilities before returning SUCCESS on uploading archives. Glacier performs regular, systematic data integrity checks and is built to be automatically self-healing." Maciej Ceglowski thinks that's [https://blog.pinboard.in/2014/04/cloudy_snake_oil/ kinda bullshit compared to the failure events you don't plan for], of course.</ref><br />
| <br />
| Retrieval is billed separately. Free to S3 or CloudFront. $50-$100/TB otherwise. Outdated?-> Data can be copied out from S3 to a SATA HD for $2.50/hr. plus media handling and shipping fees. Downloading 5TB from S3 would cost $614.40 (~$122.88/TB), but only $44.82 to transfer to HD via USB 3 or SATA (USB 2 is slower).<br />
|-<br />
|Amazon Glacier Deep Archive<br />
|$12.16 (storage only, retrieval billed separately)<br />
|<br />
|99.99999999999% <br />
|<br />
|<br />
|-<br />
|BackBlaze<br />
|$61.44<br />
|<br />
|<br />
|<br />
|$10 retrieval<br />
|-<br />
| Dropbox for Business<br />
| ~ $15 ("as much as you need" for $720/year, with soft caps)<br />
| <br />
| <br />
| <br />
| Dropbox for Business Advanced is currently billed at $20 per user per month, with a minimum of 3 users, for "as much as you need" storage. They'll probably get cranky as we near 50/100TB though.<br />
|-<br />
| Box.com for Business<br />
| $180* ("unlimited" storage for $900/year)<br />
| <br />
| <br />
| <br />
| Box.com for Business provides "unlimited" storage at $15/user/month, five user minimum, or $900/year.<br />
|-<br />
| Google Apps Unlimited<br />
| $120* ("unlimited" storage for $600/year, free for Google Apps for Education/Nonprofits)<br />
|<br />
|<br />
|<br />
| Google Apps Unlimited, Google Apps for Education and Google Apps for Nonprofits provide 1TB of storage per user for domains with less than 5 users, and "unlimited" storage for domains with five or more users. Google Apps Unlimited starts at $10/user/month, and at least $600/year for unlimited storage. Google Apps for Education and Nonprofits is free but requires fulfillment of certain criteria, such as being a 501(c)(3) nonprofit.<br />
|-<br />
| Dedicated colocated storage servers<br />
| $100* (e.g. $1300 for one year of 12TB rackmount server rental)<br />
|<br />
|<br />
|<br />
| Rent [http://www.ovh.com/us/dedicated-servers/storage/ storage servers from managed hosting colocation providers], and pool data across them. Benefits include bandwidth and electricity being included in the cost, and files could be made available online immediately. Negatives include needing to administer tens of servers.<br />
|}<br />
<br />
== Software Options ==<br />
<br />
Some of the physical options require supporting software.<br />
<br />
Removable media requires a centralized index of who has what discs, where they are, how they are labeled, and what the process for retrieval/distribution is. It could just be a wiki page, but it does require something.<br />
<br />
A simple pool of HDs ("simple pool"), one without a shared filesystem, just people offering up HDs, requires software running on Windows, Linux and/or Mac hardware to allow Archive Team workers to learn who has free disk space, and to save content to those disks. This could be just an IRC conversation and SFTP, but the more centralized and automated, the more likely available disk space will be able to be utilized. Software that is not cross-platform cannot be used here.<br />
<br />
A simple distributed and redundant pool of HDs ("distributed pool") requires software running on Windows, Linux and Mac hardware to manage a global filesystem or object store, and distribute uploads across the entire pool of available space, and make multiple copies on an ongoing basis to ensure preservation of data if a pool member goes offline. This has to be automated and relatively maintenance-free, and ideally low-impact on CPU and memory if it will be running on personal machines with multi-TB USB drives hanging off them. Software that is not cross-platform cannot be used here.<br />
<br />
A dedicated distributed and redundant pool of HDs ("dedicated pool") requires a selection of dedicated hardware and disks for maximum availability, and software to run on that hardware to manage a global filesystem or object store. It has to be automated and relatively maintenance-free, but would be the only thing running on its dedicated hardware, and as such does not have to be cross-platform.<br />
<br />
{| class="wikitable sortable"<br />
! Software name<br />
! Filesystem or Object Store?<br />
! Platform(s)<br />
! License<br />
! Good for which pool?<br />
! Pros<br />
! Cons<br />
! Notes<br />
|-<br />
| Tahoe-LAFS<br />
| Filesystem<br />
| Windows, Mac, Linux<br />
| GPL 2+<br />
| Distributed, dedicated<br />
| Uses what people already have, can spread expenses out, could be a solution done with only software<br />
| Barrier to leaving is non-existent, might cause data-loss even with auto-fixing infrastructure. Too slow to be a primary offloading site. <ref>"Practically the following results have been reported: 16Mbps in throughput for writing and about 8.8Mbps in reading" -- from https://tahoe-lafs.org/trac/tahoe-lafs/wiki/FAQ, making it non-competitive with the 1-2 gigabit speeds needed when archiving twitch.tv.</ref><br />
| Accounting is experimental, meaning "in practice is that anybody running a storage node can also automatically shove shit onto it, with no way to track down who uploaded how much or where or what it is" -joepie91 on IRC<br />
|-<br />
| Ceph<br />
| Object store, Filesystem<br />
| Linux<br />
| LGPL<br />
| Dedicated<br />
|<br />
|<br />
|<br />
|-<br />
| GlusterFS<br />
| Filesystem<br />
| Linux, BSD, OpenSolaris<br />
| GPL 3<br />
| Dedicated<br />
|<br />
|<br />
|<br />
|-<br />
| Gfarm<br />
| Filesystem<br />
| Mac, Linux, BSD, Solaris<br />
| X11<br />
| Dedicated<br />
|<br />
|<br />
|<br />
|-<br />
| Quantcast<br />
| Filesystem<br />
| Linux<br />
| Apache<br />
| Dedicated<br />
|<br />
| Like HDFS, intended for MapReduce processing, which writes large files, and doesn't delete them. Random access and erasing or moving data around may not be performant.<br />
| <br />
|-<br />
| GlusterFS<br />
| Filesystem<br />
| Mac, Linux, BSD, Solaris<br />
| GPL 3<br />
| Dedicated<br />
|<br />
|<br />
|<br />
|-<br />
| HDFS<br />
| Filesystem<br />
| Java<br />
| Apache<br />
| Distributed, dedicated<br />
|<br />
| Like Quantcast, intended for MapReduce processing, which writes large files, and doesn't delete them. Random access and erasing or moving data around may not be performant.<br />
| <br />
|-<br />
| XtreemFS<br />
| Filesystem<br />
| Linux, Solaris<br />
| BSD<br />
| Dedicated<br />
|<br />
|<br />
|<br />
|-<br />
| MogileFS<br />
| Object store<br />
| Linux<br />
| GPL<br />
| Dedicated<br />
| Understands distributing files across multiple networks, not just multiple disks<br />
|<br />
| As an object store, you can't just mount it as a disk and dump files onto it, you have to push them into it through its API, and retrieve them the same way.<br />
|-<br />
| Riak CS<br />
| Object store<br />
| Mac, Linux, BSD<br />
| Apache<br />
| Dedicated<br />
| S3 API compatible<br />
| Multi-datacenter replication (which might be what you consider having multiple disparate users on different networks) is only available in the commercial offering.<br />
| A former Basho employee suggests this might not be a good fit due to the high latency and unstable connections we'd be dealing with. Datacenter-to-datacenter sync is an "entirely different implementation" than local replication, and would require the enterprise offering.<br />
|-<br />
| MongoDB GridFS<br />
| Object store<br />
| Windows, Mac, Linux<br />
| AGPL<br />
| Distributed, dedicated<br />
|<br />
| MongoDB fares poorly at data integrity, performance is dubious, and scaling is a nightmare. Not recommended.<br />
|<br />
|-<br />
| LeoFS<br />
| Object store<br />
| Mac, Linux<br />
| Apache<br />
| Dedicated<br />
| S3-compatible interface, beta NFS interface, supports multi-datacenter replication, designed with GUI administration in mind<br />
|<br />
|<br />
|-<br />
| Resilio Sync (BitTorrent Sync)<br />
| Synchronization<br />
| Windows, Mac, Linux, BSD, NAS<br />
| Proprietary<br />
| Simple<br />
| Commercially supported software<br />
| As straight synchronization software, it mirrors folders across devices. Individual users would have to make synched folders available to get copies of archives, and then they would be mirrored, and that's it.<br />
| Poor performance with large files (10 GB+) in large folders (200 GB+).<br />
Synchronization software in general is not the right solution for this problem.<br />
|-<br />
| [http://syncthing.net/ Syncthing]<br />
| Synchronization<br />
| Windows, Mac, Linux, BSD, NAS<br />
| GPL<br />
| Simple<br />
| Open Source Software, active Developement, individual rights<br />
| As straight synchronization software, it mirrors folders across devices. Individual users would have to make synched folders available to get copies of archives, and then they would be mirrored, and that's it. Rightsmanagment allows to only download but not change the files in the cloud.<br />
| Synchronization software in general is not the right solution for this problem.<br />
|-<br />
| BitTorrent<br />
| Filesystem<br />
| All<br />
| various<br />
| Distributed<br />
| Readily available technology, easily understood distribution model, contributors can join or leave at any time<br />
| Harder to get people interested in contribute if they have to join bittorrent swarms<br />
| Breaking a large item up into smaller torrents makes contributing smaller chunks of space possible, and a custom client could be created which would let the user dedicate some space and automatically join the swarms which have the fewest peers. Getting the initial seeds requires coordination to distribute the data across available seeds by other means, creating the sub-torrents, etc.<br />
|-<br />
| Git Annex<br />
| Filesystem<br />
| linux, mac, windows, android (in that order)<br />
| GPL (mostly)<br />
| Distributed<br />
| Once you join a repository (by using git clone), you can choose which of the files within that repository you will actually store locally.<br />
| Pretty complicated to use on the command-line; git annex assistant isn't quite geared for this use-case but could handle it anyway.<br />
| [[INTERNETARCHIVE.BAK/git-annex_implementation]]<br />
|}<br />
<br />
== Non-options ==<br />
* Ink-based Consumer Optical Media (CDs, DVD, etc.) <br />
** Differences between Blu-Ray and DVD? DVDs do not last very long. The fact is, the history of optical writable media has been on of chicanery, failure, and overpromising while under-delivering. Some DVDs failed within a year. There are claims Blu-Ray is different, but fool me 3,504 times, shame on me.<br />
* Resilio Sync (BitTorrent Sync)<br />
** Proprietary (currently), so not a good idea to use as an archival format/platform<br />
* Amazon S3 / Google Cloud Storage / Microsoft Azure Storage<br />
** Amazon S3 might be a viable waypoint for intra-month storage ($30.68/TB), but retrieval over the internet, as with Glacier, is expensive, $8499.08 for 100TB. Google's and Microsoft's offerings are all in the same price range.<br />
* Floppies<br />
** ''"Because 1.4 trillion floppies exists less than 700 billion floppies. HYPOTHETICALLY, if you set twenty stacks side by side, figure a quarter centimeter per floppy thickness, excluded the size of the drive needed to read the floppies you would still need a structure 175,000 ft. high to house them. Let's also assume that the failure rate for floppies is about 5% (everyone knows that varies by brand, usage, time of manufacture, materials used, etc, but lets say 5% per year). 70 million of those 1.4 trillion floppies are unusuable. Figuring 1.4 MB per floppy disk, you are losing approximately 100MB of porn each year. Assuming it takes 5 seconds to replace a bad floppy, you would have to spend 97,222 hrs/yr to replace them. Considering there are only 8,760 hrs per year, you would require a staff of 12 people replacing floppies around the clock or 24 people on 12 hr shifts. Figuring $7/hr you would spend $367,920 on labor alone. Figuring a nickel per bad floppy, you would need $3,500,000 annually in floppy disks, bringing your 1TB floppy raid operating costs (excluding electricity, etc) to $3,867, 920 and a whole landfill of corrupted porn. Thank you for destroying the planet and bankrupting a small country with your floppy based porn RAID."'' ([http://gizmodo.com/5431497/why-its-better-to-pretend-you-dont-know-anything-about-computers?comment=17793028#comments source])<br />
<br />
== Alternatives ==<br />
<br />
For completeness sake:<br />
<br />
* Fund raising. If IA had more donations coming in then the problem would be less important.<br />
* Grant writing. A more formal form of fund raising, but most grant writers would expect the grantee to be doing the work. Works well in combination with the other methods mentioned above.<br />
* Better accessibility. If the problem is that these archives are simultaneously large and infrequently accessed, then making them more accessible would make the size easier to swallow.<br />
<br />
== From IRC ==<br />
<br />
<Drevkevac> we are looking to store 100TB+ of media offline for 25+ years<br />
<Drevkevac> if anyone wants to drop in, I will pastebin the chat log<br />
<rat> DVDR and BR-R are not high volume. When you have massive amounts of data, raid arrays have too many points of failure.<br />
<rat> Drevkevac: I work in a tv studio. We have 30+ years worth of tapes. And all of them are still good.<br />
<rat> find a hard drive from 30 years ago and see how well it hooks up ;)<br />
<brousch_> 1500 Taiyo Yuden Gold CD-Rs http://www.mediasupply.com/taiyo-yuden-gold-cd-rs.html<br />
<br />
<Drevkevac> still, if its true, you could do, perhaps, raidz3s in groups of 15 disks or so?<br />
<SketchCow> Please add paperbak to the wiki page.<br />
<SketchCow> Fuck Optical Media. not an option;.<br />
<Drevkevac> that would give you ~300GB per disk group, with 3 disks<br />
<br />
== Where are you going to put it? ==<br />
<br />
Okay, so you have the tech. Now you need a place for it to live.<br />
<br />
Possibilities:<br />
<br />
* The Internet Archive Physical Warehouse, Richmond, CA<br />
** The Internet Archive has several physical storage facilities, including warehouses in Richmond, CA (home of the Physical Archive) and the main location in San Francisco, CA. They have indicated they are willing to take copies of Archive Team-sponsored physical materials with the intent of them being ingested into the Archive at large over time, as costs lower and 100tb collections are not as big a drain (or a rash of funding arrives elsewhere).<br />
<br />
* Living Computer Museum, Seattle, WA<br />
** In discussions with Jason Scott, the Living Computer Museum has indicated they will have physical storage available for computer historical materials. Depending on the items being saved by Archive Team, they may be willing to host/hold copies for the forseable future.<br />
<br />
* Library of Congress, Washington, DC<br />
** The Library of Congress may be willing to take a donation of physical storage, although it is not indicated what they may do long-term with it.<br />
<br />
Multiple copies would of course be great.<br />
<br />
== No, seriously, how are you going to actually DO it ==<br />
<br />
There are only a few practical hardware+software+process combinations. In order of cost to each volunteer:<br />
<br />
* A pool of volunteers with Blu-ray burners commit to ("the Blu-ray option"): <br />
** buying a 50-disc spindle of 25GB discs per TB per project,<br />
** burning them,<br />
** verifying them,<br />
** storing them somewhere climate-controlled (a shelf in a house with AC and heat is fine, an attic/garage/flooded basement is not),<br />
** verifying them regularly (monthly? quarterly?) and replacing discs if necessary, and<br />
** shipping them somewhere else upon request, with no expectation of return (permanent storage, consolidation, etc.).<br />
<br />
This probably requires a minimum of three volunteers per TB per project. Probably best to pre-split the data into < 25GB chunks so each disc can be labeled the same and expected to have the same data on it. Fifty 25GB discs is a little more than a TB, and it's expected you'll lose a few to bad burns each time, but it might be worth buying more than a spindle and generating parity files onto additional discs.<br />
<br />
* A pool of volunteers commit to ("the simple pool"):<br />
** buying a best reasonable external HD,<br />
** downloading archives to it,<br />
** keeping it spun up, or spinning it up regularly (monthly? quarterly?) and running filesystem and content checks on it,<br />
** storing it somewhere climate-controlled (a shelf in a house with AC and heat is fine, an attic/garage/flooded basement is not),<br />
** buying additional HDs once it's full or if there are drive errors, and<br />
** shipping it somewhere else upon request, with no expectation of return (permanent storage, consolidation, etc.).<br />
<br />
Same as with Blu-rays, and not really any more expensive ($150 == $37.50 for one 1TB of Blu-rays * 4, or one 4TB HD), except look at all that disc-swapping time and effort you don't have to do. You don't have to split data into chunks, but you do want to download it in a resumable fashion and verify it afterwards, so, checksums, parity files, something. You also risk losing a lot more if a drive fails, and the cost per-volunteer is higher (replacing a whole drive versus replacing individual discs or spindles). As such, you still probably want a minimum of three volunteers per TB per project (so a 2TB project needs six volunteers with 1TB each, not three volunteers holding all 2TB each).<br />
<br />
* A pool of volunteers commit to ("the distributed pool"):<br />
** all buying the same, standard, inexpensive, hackable, RAID 1, NAS,<br />
*** WD My Cloud Mirror (starts at $300 for 2TB [called "4TB," only 2TB with mirroring])<br />
*** QNAP (2-bay starts at $140 without HDs)<br />
*** Synology (2-bay starts at $200 without HDs)<br />
*** Pogoplug Series 4 + two best reasonable external HD + software RAID 1, or a download script that manually mirrors files ($20 without HDs)<br />
** keeping it spun up, online, and possibly accessible by external AT admins,<br />
** storing it somewhere climate-controlled (a shelf in a house with AC and heat is fine, an attic/garage/flooded basement is not),<br />
** buying entire additional units once they are full or if there are drive errors, and<br />
** shipping the drives (or the entire My Cloud Mirror unit, if that's the one selected) somewhere else upon request, with no expectation of return (permanent storage, consolidation, etc.).<br />
<br />
These units provide dramatically improved reliability for content, enough that perhaps you only need two volunteers per project, and no need to split by TB, since each volunteer would have two copies. Having everyone buy the same hardware means reduced administration time overall, especially if custom scripts are involved. QNAP and Synology both have official SDKs, and all of them run some flavor of Linux, with Synology supporting SSH logins out of the box. The Pogoplug is the most underpowered of the options, but even it should be powerful enough to run a MogileFS storage node, or a script that downloads to one HD and copies to the other. (Checksums would be really slow, though.) This is moderately expensive per-volunteer, with an upfront cost of $320-$500.<br />
<br />
Consumer NAS devices have severe firmware issues, potentially causing full data loss on a trivial operation. Such a case was previously observed after flashing a new official firmware image onto a QNAP Pro series 4 bay NAS (700€ empty) while the RAID was presumably resyncing. It has to be expected that the device prefers reinitialization over being stuck with an error.<br />
<br />
HDD compatibility is limited and has needs close investigation, WD Green 2TB for example tend to frequently degrade the RAID array and accumulate load cycles from frequent head parking.<br />
<br />
* A pool of volunteers commit to ("the dedicated pool"):<br />
** all buying the same, standard, expensive NAS,<br />
*** iXsystems FreeNAS Mini (starts at $1000 without HDs),<br />
*** A DIY FreeNAS box ($300+ without HDs),<br />
*** A DIY NexentaStor box (probably the same as the DIY FreeNAS box)<br />
** keeping it spun up, online, and possibly accessible by external AT admins,<br />
** storing it somewhere climate-controlled and well-ventilated (a shelf with no airflow is not fine),<br />
** replacing drives if there are drive errors,<br />
** migrating the pool to larger disks once it starts getting full, and<br />
** shipping the drives somewhere else upon request, with no expectation of return (permanent storage, consolidation, etc.).<br />
<br />
A set of volunteers with (comparatively) expensive network-attached storage gives you a lot of storage in a lot of locations, potentially tens of redundant TB in each one, depending on the size of the chassis. You want everyone running the same NAS software, but the hardware can vary somewhat; however, the hardware should all have ECC RAM, and the more the better. MogileFS storage nodes are known to run on NexentaStor, and FreeNAS supports plugins, so it could be adapted to run there, or you could figure out e.g. LeoFS (which also expects ZFS). This is the most expensive option per-volunteer, upfront costs starting at around $1300 for a DIY box with four 4TB WD Red drives.<br />
<br />
* A pool of volunteers set up a recurring payment to fund ("the server option"):<br />
** one or more rented, managed, storage servers; or<br />
** saving up to buy one or more storage servers, and then hosting it somewhere.<br />
<br />
A rented server has no hardware maintenance costs; replacing a failed HD is the responsibility of the hosting provider, both in terms of materials cost and in labor cost. This is not the case with a purchased server, where someone would have to buy a replacement hard drive, bring it to the colocation center, and replace the drive; or someone would have to buy a replacement disk, ship it to the colocation center, and then they would bill someone for the labor involved in replacing it.<br />
<br />
== What Can You Contribute? == <br />
<br />
{| class="wikitable"<br />
! Name<br />
! What You Can Contribute<br />
! For How Long?<br />
! Exit Strategy<br />
|-<br />
| ExampleArchiver<br />
| Describe what you are willing to buy/build/write/do. Talk about the connection you would use, the storage conditions, etc. How much money can you put into it? <br />
| For how long can you truly commit to this?<br />
| If you need to quit or wind down your contribution, what are you willing to do? Can you guarantee a period of notice? Are you willing to ship your hardware or media to another volunteer anywhere in the world, or will you want to keep it? <br />
|-<br />
| vitorio<br />
|<br />
* Participating in the simple pool (I only have a laptop, so I'd store the HDs offline at home and check them monthly/quarterly)<br />
* Participating in the distributed pool (residential 30/10 connection)<br />
* Contributing $100/mo. for the server option<br />
| Indefinitely<br />
| Can give ample notice for either full upload and/or shipping of all hardware anywhere in the world.<br />
|-<br />
| pluesch<br />
|<br />
* Willing to provide 12 TB storage space (3x 4TB drives) on a rented ovh.com server; Access via SSH; Uploaded stuff can be made available via Rsync and HTTP<br />
| As long as I have a job. It's a very stable position atm.<br />
| If I can't provide the storage anymore I'll inform archiveteam at least 3 months before.<br />
|-<br />
|}<br />
<br />
== Project-specific suggestions ==<br />
<br />
=== Twitch.tv (and other video services) ===<br />
<br />
* Keep the original video files in (semi-)offline storage, and store transcoded (compressed) versions on the Internet Archive.<br />
<br />
== See Also ==<br />
* [[Storage Media]]<br />
* [[INTERNETARCHIVE.BAK]]<br />
* [[Places to store data]]<br />
<br />
== References ==<br />
<references/><br />
<br />
{{Navigation box}}</div>Systwihttps://wiki.archiveteam.org/index.php?title=Backup_Tips&diff=45424Backup Tips2020-08-25T20:27:32Z<p>Systwi: Changed wording.</p>
<hr />
<div>Personal digital archiving is all the rage nowadays. This article will give you a basic overview of why you should do it and how.<br />
<br />
==How Is Data Lost?==<br />
<br />
Here's a short list of the ways to lose data:<br />
<br />
* Disk failure<br />
* Software failure<br />
* Malicious software<br />
* Natural disaster<br />
* Clumsy user<br />
* Accidental deletion<br />
* Accidental overwriting<br />
* Cat hair<br />
* Refrigerator magnets<br />
* Solar radiation<br />
* You forgot where you put it<br />
* Your parents/roommate/spouse moved it and didn't tell you<br />
* The feds paid a surprise visit to you/your storage provider<br />
* Your storage provider went under/got bought/got bored<br />
<br />
Any one of these can erase decades of data in a second. The goal of good backups is to contain the damage any one of these can cause, ideally to nearly zero. The ways to lose data can be summarized into these categories:<br />
<br />
# Operational - drives wear down, software writes garbage, user error<br />
# Environmental - building catches fire, hurricane knocks your house over<br />
# Access - you lose track of data, or lose ability to get to it<br />
<br />
Thus, a good backup plan is resilient against each of these types of failure. We'll use [http://thewakilibrarian.wordpress.com/2012/02/23/personal-digital-archiving-2012-keynote-by-mike-ashenfelder/ Michael Ashenfelder's four-step process] as a model.<br />
<br />
==Identify/Decide==<br />
<br />
Before starting any sort of backup plan, it helps to identify what you're saving. Terabytes save differently than megabytes; knowing which you plan on saving can save money and grief.<br />
<br />
One way to get started is to envision the following scenarios (which also serve as excellent fire drills):<br />
<br />
* You get a call from your lawyer, telling you that someone opened a suit against you. Your lawyer says that it's easy to take care of, but it requires as much documentation as possible to build your case. ''This encompasses things like financial data (Quicken, Excel, etc.), legal documentation, and the like.''<br />
* You get a call from a client/boss, saying that the Big Project needs some crucial information from some old work of yours to save it. They don't have a specific date, but they'll sift through everything more than one year old to find it. ''This encompasses anything you do for work, be it media projects, code, reports, etc. Essentially, stuff your livelihood is based on.''<br />
* You learn the hard way that nobody keeps backups of people. Your next of kin go through your effects, and come across your digital data. ''Here, they'll find personal things - photographs/videos, special emails, etc. - essentially, stuff that's meaningful to you.''<br />
* You have to move to a developing country for a short time - not long enough to think long-term, but long enough that you'll want some amenities. Due to arcane customs laws, you can only bring one small hard drive into the country. ''These are things that you really like and would rather not lose, but don't fall into the above. Things like contact information, game saves, hard-to-replace data, favorite porn, etc. Think of this as a catch-all.''<br />
<br />
Here are things you probably shouldn't save:<br />
<br />
* Program and system files. Unless you run a high-reliability business server, there's little need to have a ready copy of ''explorer.exe''. If you have the install discs handy, then there's no real reason to back these up. Note that software published by very small houses (music software comes to mind) can be hard to track down later - it may be prudent to archive these applications.<br />
<br />
That being said, remember that '''storage is cheap, but your data is priceless.''' When in doubt, save it - the cost of doing so is nearly zero, and the cost of losing it is not.<br />
<br />
==Organize==<br />
<br />
Make sure you assess '''all possible data sources''' when deciding what to back up. There's nothing more embarrassing than losing your vacation photos because you didn't copy them off your phone before pitching it. If you have anything stored remotely - on a webhost, in an email server, in the cloud - copy it locally! Odds are good that the service won't be there when you need it most.<br />
<br />
When backing up, it helps to keep everything together in one large archive. This solves a number of problems:<br />
<br />
* You won't forget where you put that Really Important Data - don't be like [http://jordanmechner.com/blog/2012/03/prince-of-persia-source-code-found/ Jordan Mechner]! - because it's all in one place.<br />
* It's easier to keep one big archive reliable than many small ones (think economy of scale)<br />
* Buying a few big hard drives is cheaper than buying many small ones (and they tend to be more reliable)<br />
<br />
==Save Copies==<br />
<br />
The goal here is to mitigate the damage caused by sudden catastrophic data loss, so that your valuable data (from above) is kept safe.<br />
<br />
===Scheme===<br />
<br />
First, buy some hard drives. Mechanical (traditional) drives are the cheapest for the storage and their longevity/flaws are well-documented. For purposes of personal archiving, consumer drives are sufficient - so long as it's not a disastrously bad line, any drive is sufficient. 1-1.5TB drives are roomy and cheap, and are recommended.<br />
<br />
A basic backup scheme may look like this:<br />
<br />
# Primary storage (your PC/phone/tablet/etc) - changes constantly as you use it<br />
# Secondary local storage (a hard drive in a closet) - changes once every 2-3 weeks<br />
# Secondary offsite storage (a hard drive in a safety deposit box) - changes once or twice a year [optional but highly recommended]<br />
<br />
This scheme provides resilence against most common failures: if one drive dies, there are two backups; if you delete something, you have two; if your house floods, you have one. So long as you are vigilant, the chance of total data loss is negligible, even in case of total disaster. You may wish to add more drives to each area in accordance with your paranoia.<br />
<br />
Keeping backup cycles up is important for both the longevity of the hardware and security of your data. Not only does it allow you to keep your data current, but it can show early signs of hardware failure as you read/write to the disks.<br />
<br />
====Why Not the Cloud?====<br />
<br />
You may be thinking "why not use cloud backups as offsite storage?" The answer is: you can, but it's risky, and you should only use it to supplement an already solid scheme.<br />
<br />
The cloud offers many seductive features, such as high disk reliability, easy access, and cheap storage. However, there's a hidden cost: by using cloud storage, you '''lose control of your data'''. By trusting a storage provider with your data, you trust them to be there tomorrow. This has shown to be a very risky agreement, as cloud storage providers tend not to be long-lived, and those that fall don't give a damn about your data loss. The provider may lose interest ([[MobileMe]]), close shop ([[Deathwatch#Dead as a doornail]]), or have a surprise party thrown by the Department of Justice ([[MegaUpload]]).<br />
<br />
In short, cloud storage is a lot like real clouds - insubstantial, fleeting, and really bad to build on. This isn't to say it's useless - cloud storage can be a useful function in a storage scheme due to easy access - but it's not something you should trust your backups to. Put another way, AT wouldn't be working overtime to save dying clouds if they were reliable long-term.<br />
<br />
===Doing It===<br />
<br />
Here's a quick, dirty, and platform-agnostic backup method:<br />
<br />
# Connect the external drive to your computer. <br />
# In the root of the external hard drive, create a folder called 'backups'. <br />
# In the 'backups' folder, create another folder named after your computer's name (e.g. "POSEIDON"). <br />
# In this new folder, create another folder of today's date in the format of YYYY-MM-DD (i.e. {{CURRENTYEAR}}-{{CURRENTMONTH}}-{{CURRENTDAY}}). This format ensures that Windows will properly sort the folders in date order. <br />
# Copy everything you selected (if you're extra-thorough, everything from your C: drive) into this date folder. A program like [http://www.codesector.com/teracopy.php TeraCopy] is exceedingly useful, as it supports copy verification, pause/resume and most importantly, won't randomly die if it runs into any problems.<br />
<br />
You'll probably have space left over (unless you do a lot of media editing), so you can repeat this once per backup cycle on the same drive. This gives you some extra peace of mind in case one cycle's backup is corrupted, as you can use the next most recent one.<br />
<br />
==The 3-2-1 Backup Strategy==<br />
<br />
This rule of thumb can greatly reduce the likelihood of data loss:<br />
<br />
* Keep a total of at least '''3''' copies of your data,<br />
* '''2''' of which are on different media formats and at least<br />
* '''1''' off-site copy.<br />
<br />
Keyword is "at least." The more copies of your data you have the lower the chance of data loss.<br />
<br />
==Conclusion==<br />
<br />
Congrats! You're now resistant against catastrophic data loss. This is only the beginning of good archiving - '''vigilance''' is the watchword of digital archiving. Keep an eye on disk health, run through fire drills (either literally or figuratively), and stay consistent with backups.<br />
<br />
As the old joke goes, there are two kinds of people in the world: those that keep backups, and those that haven't lost data yet. Don't let it happen to you.<br />
<br />
==See also==<br />
* [[Places to store data]]<br />
<br />
{{Navigation box}}<br />
<br />
[[Category:Backup]][[Category:Data Loss Prevention]]</div>Systwihttps://wiki.archiveteam.org/index.php?title=Move_Archiveteam_to_Hackint&diff=45414Move Archiveteam to Hackint2020-08-23T01:53:12Z<p>Systwi: Added myself (systwi) to the list of those in favor of moving to Hackint.</p>
<hr />
<div>__NOTOC__<br />
<big>PROPOSED: MOVE ARCHIVETEAM IRC COMMUNICATION PRIMARILY TO HACKINT FROM EFNET</big><br />
<br />
Proposed by JAA a good long while ago, the proposal is to move Archive Team's IRC channels (and many project sub-channels) from EFNet to HackINT.<br />
<br />
As is typical, we're currently split between the two networks, with many channels in HackInt and many others in EFNet, depending on the preferences and inclinations of various members. Honestly, this can't continue. As most activity is happening in Hackint anyway, and because we might as well use this for a quorum discussion, this page exists for discussion (along with the talk page) *NOW* (Mid-August) to September 30th, at which point it will (hopefully) be very clear which direction we should go. This page is likely to get increased changes and traffic as time goes, so check back often.<br />
<br />
* Information on hackint is here: https://www.hackint.org/<br />
* Information on EFNet is here: http://www.efnet.org/<br />
<br />
= Arguments for moving Archive Team to Hackint =<br />
* IRC Services: No need to micromanage ops in each project channel, ease of administration<br />
* Stability: Handful of netsplits that lasted just a few minutes on Hackint compared to the ones on EFnet<br />
* Support: IRC staff has proven to be very helpful and useful and helped us get running with channel and user administration<br />
* Limits: Per-connection channel limits are much higher on Hackint and do not differ across servers (123 is the current joined-channels per connection limit and opers are willing to increase this network-wide limit in case we hit it)<br />
<br />
= Arguments for keeping Archive Team on EFnet =<br />
* IRC Services: Inherently leads to a certralisation of permissions vs current system.<br />
* It has Always Been EFNet, we shouldn't uproot our long-standing relationship and work with that network.<br />
* EFNet is the longest-lived Network, showing it's here to stay.<br />
* We should just engage with EFNet to make them more hospitable for Archive Team needs.<br />
<br />
== How Moving to HackInt would Work ==<br />
<br />
Jason says "There would almost certainly be a #archiveteam channel on EFNet forever, with some people sitting in related long-time channels like #archiveteam-bs and #archiveteam-ot - but project channels would shut down and move to HackInt. So we'd still have a split, but the channels on EFNet would be more like either social hangouts or represent outreach to guide people to the other location."<br />
<br />
== Signatories ==<br />
<br />
* This is not a vote; this is a show of support in one direction.<br />
<br />
Edit and add your name to one of the lists below if you have a strong opinion one way or another. Describe your thinking, if you'd like. Do not add others if they are not on the Wiki.<br />
<br />
=== In Favor of Move to Hackint ===<br />
* [[User:Kiska]]<br />
* [[User:wessel1512]]<br />
* [[User:Aoede]]<br />
* [[User:Fusl]]<br />
* [[User:Katocala]]<br />
* [[User:ivan]]<br />
* [[User:JAA]]<br />
* [[User:Flashfire42]]<br />
* [[User:Jake]]<br />
* [[User:Kaz]]<br />
* [[User:Maxfan8]]<br />
* [[User:Jrwr]]<br />
* [[User:Craigle]]<br />
* [[User:systwi]]<br />
<br />
=== In Favor of Staying at EFNet ===</div>Systwihttps://wiki.archiveteam.org/index.php?title=Talk:YouTube&diff=41003Talk:YouTube2019-07-21T18:41:41Z<p>Systwi: Updated user profile URL</p>
<hr />
<div>==youtube2internearchive==<br />
https://github.com/emijrp/youtube2internetarchive contains a script which also handled upload to Internet Archive, but I can't find it any longer. --[[User:Nemo_bis|Nemo]] 06:28, 26 January 2015 (EST)<br />
:I've found [https://code.google.com/p/emijrp/source/browse/trunk/scrapers/youtube2internetarchive.py something] with Google. [[User:bzc6p|bzc6p]] ([[User_talk:bzc6p|talk]]) 12:25, 26 January 2015 (EST)<br />
<br />
If YouTube needed to be quickly captured for some unforeseen reason, it might make sense to download only the XML and SRT files, so then at least some record would be saved. Google's subtitle recognition is currently far from accurate, but it's certainly improving. [[User:wtron|wtron]] 06:48, 12 June 2015 (EST)<br />
<br />
== Options ==<br />
<br />
Is it really necessary to explicitly call best A/V when youtube-dl it by default? <br />
<br />
Also, why not embed the subs and thumbnail instead of make a separate file? Also why not xattrs for those of us with unix filesystems? Xattrs is only one extra flag. <br />
<br />
My command is currently <br />
<br />
youtube-dl -t --embed-subs --add-metadata --xattrs --console-title --embed-thumbnails<br />
<br />
although I'm going to be incorporating elements from the suggested one into mine. The reasoning behind this is it's one file to send. That command is how I archive currently, it's changing though.<br />
<br />
I'd appreciate hearing your input about why I may be wrong though. Thanks in advance,<br />
<br />
--[[User:Vxbinaca|Vxbinaca]] 21:24, 29 May 2015 (EDT)<br />
<br />
On your second note, I strongly believe it's better to have different things (video, thumbnail, subtitle) in separate files. Easier to access, process, categorize, recognize. I think it's worth the "trouble" of having three files (with the same name) instead of one.<br />
<br />
[[User:bzc6p|bzc6p]] ([[User_talk:bzc6p|talk]]) 07:08, 31 May 2015 (EDT)<br />
<br />
xattrs are not portable and will get lost when copying to a file system that doesn't have them (or when uploading it somewhere, like to IA) --[[User:Darkstar|Darkstar]] 08:53, 31 May 2015 (EDT)<br />
<br />
Solid reasoning. I've now switched to your way of doing things. --[[User:Vxbinaca|Vxbinaca]] 19:32, 2 August 2015 (EDT)<br />
<br />
== --ignore-errors shouldn't be youtube-dl archiving default best practices ==<br />
<br />
Theres a myriad of reasons this isn't a good idea to have by default. Downloads getting snapped off on channel rips could go unnoticed (I search for these with ls *.part). Problems with various versions of youtube-dl could lead to a channel rip with half-processed videos, see [https://github.com/rg3/youtube-dl/issues/6413 this issue on github].<br />
<br />
Perhaps for a well-tested version that works on YouTube running in a warrior, --ignore-errors is appropriate, but for an attended rip we should by default suggest people not use it and instead just make sure all of it got ripped and if theres an error try to work resolve that particular video, and if it's a problem they can't get around then just go --ignore-errors.<br />
<br />
I'm open to being told why I may be wrong though. --[[User:Vxbinaca|Vxbinaca]] 19:32, 2 August 2015 (EDT)<br />
<br />
== Paid subscription ==<br />
<br />
On September 1, 2015 [https://www.theverge.com/2015/4/8/8371131/youtube-paid-subscription-offline-video The Verge reported] about oncoming paid subscription option (ad-free+premium videos), although paid ad-free videos may not be a nail in the coffin, previously free content could become a premium. According to the article, transition could happen in a few months -- [[User:Vitzli|Vitzli]] 06:46, 2 September 2015 (EDT)<br />
<br />
YouTube has never been profitable. [[User:bzc6p|bzc6p]] ([[User_talk:bzc6p|talk]]) 08:46, 2 September 2015 (EDT)<br />
<br />
== As we don't have 500 petabytes of free storage... ==<br />
<br />
... a solution may be to discard "low-value" videos. I mean, if we discard duplicates (films, music, etc) and we set a limit of 1 PB (1000 TB) for quality content, what could be the lucky videos that would be downloaded and preserved? We can work in a approach like this. Just because we don't have space for all, doesn't mean that we don't download anything. [[User:Emijrp|Emijrp]] 09:33, 20 October 2015 (EDT)<br />
<br />
Google will probably give a notice long before that date. So we'll have the time to find stuff worth saving. We could create a "warroom" or such, and users (ArchiveTeam members and other people) could suggest channels worth saving, with description, average views and size estimate, say, in a table, clear-cut. If someone suggests videos that are otherwise available or not too popular, they can be striked out (with proper reasoning). (The reviewing of suggestions can be done by everyone continously.) A deadline of suggestions could be given (say, 2 months before the end), and after that a Committee could select the "lucky" 1000 TB that could end up in the Archive.<br />
* In the meanwhile, the Archive itself would be queried for already saved videos, and those wouldn't be saved again.<br />
* The Archive and the Team should expect other preserving actions, and ours should be in accordance with those (no duplicates).<br />
* There could be national limits, e.g. not only a global limit, but also language or country-specific, say, 500 TB English videos, and 10–50 terabytes per other countries (just ad-hoc numbers, see the concept).<br />
<br />
The importance of the last point and the importance of saving some of YouTube at all is, I think, in the fact that, without too much exaggaration, a substantial part of today's culture is stored and represented there, on the most popular video sharing site on Earth. [[User:bzc6p|bzc6p]] ([[User_talk:bzc6p|talk]]) 15:41, 20 October 2015 (EDT)<br />
<br />
: We could also save all videos starting at a specific view count such as 500 or 1000 and/or from channels above a subscriber threshold such as 200. But maybe better first start from e.g. 1000000 views and 10000 subscribers. --[[User:ATrescue|ATrescue]] ([[User talk:ATrescue|talk]]) 22:29, 13 May 2019 (UTC)<br />
<br />
== Saving YouTube Comments ==<br />
<br />
YouTube comments are a surefire sign of just how awful the internet can be at times. Shouldn't they be archived as well? There's already a script for it, [https://github.com/egbertbouman/youtube-comment-downloader youtube-comment-downloader]. --[[User:Powerkitten|Powerkitten]] ([[User talk:Powerkitten|talk]]) 16:49, 26 October 2016 (EDT)<br />
<br />
: Hello, [[User:Powerkitten|Powerkitten]]. Thank you for mentioning this.<br />
<br />
There is also [https://github.com/philbot9/youtube-comment-scraper-cli youtube-comment-scraper-cli] which can do this as well and supports .csv and .json outputs. This tool logs comment ID, channel name, like count, post date (in both Unix timestamp and UTC date format), user profile picture and number of replies. An online version is also available [http://ytcomments.klostermann.ca/ here]. [[User:Systwi|Systwi]] ([[User talk:Systwi|talk]]) 06:53, 21 July 2019 (UTC)<br />
<br />
<br />
Now, there is some information about comment archival here:<br />
<br />
* [[YouTube#Comment_loading]]<br />
* [[Archive.is#YouTube_comment_archival]]<br />
=== Celebrations ===<br />
When the great wall of YTimg's [[robots.txt]] fell on 20120229 (leap year's day of 2012), it was a prideful feeling. Until then, http://ythistory.weebly.com/ and http://ythistory.co.cc/ were the only places where one could access the YouTube swf (flash-based) ActionScript 2 and ActionScript-3-players between circa 2008 (when YouTube's YTimg domain started, likely due to Google's acquisition) and 2012. Additionally, when browsing YouTube via the Wayback Machine, stylesheet information and all images were also blocked by [[robots.txt]], which was suddenly gone.<br />
<br />
For a few hours after it's removal, there was an error stating “Couldn't load Robots.txt for this page”, “Unable to read Robots.txt” or “Couldn't find robots.txt for this site” or something similar (I don't remember the exact words), because it was assuming that the robots.txt was just temporarily inaccessible. After these few hours past, suddently, all the YouTube pages blocked by robots.txt were suddently visible in their full beauty.<br />
<br />
Now, because one of the biggest Wayback dreams I and {{URL|2=other people|1=https://archive.org/post/406632/why-does-the-wayback-machine-pay-attention-to-robotstxt}} had apparently has come true (read more at [[Internet_Archive#robots.txt_and_the_Wayback_Machine]] too), I can now load YouTube comments through the Wayback Machine! ([http://web.archive.org/web/20190512223447/https://www.youtube.com/watch?v=W3GrSMYbkBE&lc=UgxC238Gea0KGOditl54AaABAg Try here]) --[[User:ATrescue|ATrescue]] ([[User talk:ATrescue|talk]]) 22:52, 12 May 2019 (UTC)<br />
<br />
: @Powerkitten: http://ytcomments.klostermann.ca/ (title element={YouTube Comment Scraper}; title on the webpage={Download comments from YouTube};) can be used to get all the comments on a video, which can be downloaded as JSON and CSV at http://ytcomments.klostermann.ca/scrape. You can then upload those files to archive.org. --[[User:Usernam|Usernam]] ([[User talk:Usernam|talk]]) 10:57, 20 May 2019 (UTC)<br />
<br />
: archive.is can also archive replies to comments, given you have the link to a reply to a comment. --[[User:Usernam|Usernam]] ([[User talk:Usernam|talk]]) 00:17, 15 July 2019 (UTC)<br />
<br />
== Many YouTube video items are archived in low quality because they were archived using an old version of tubeup. Is there a way I can replace the files of old youtube items (using the "youtube-[id]" naming format the tubeup uses) with updated, high quality files? --[[User:Hiccup|Hiccup]] ([[User talk:Hiccup|talk]]) 17:01, 6 February 2018 (UTC) ==<br />
<br />
Many YouTube video items are archived in low quality because they were archived using an old version of tubeup. Is there a way I can replace the files of old youtube items (using the "youtube-[id]" naming format the tubeup uses) with updated, high quality files? --[[User:Hiccup|Hiccup]] ([[User talk:Hiccup|talk]]) 17:01, 6 February 2018 (UTC)<br />
<br />
== Thumbnail rescue ==<br />
<br />
Some YouTube uploaders change thumbnails of their videos, which would lead to the loss of the existing thumbnail if archived nowhere else.<br />
<br />
Some YouTube thumbnails can be retrieved by searching the video ID using web search (try different search engine).<br />
<br />
To save as many thumbnails as possible from a specific channel, please archive the video page of their channel using the help of [[chromebot]]'s page scrolling.<br />
<br />
One can also manually scroll down a page and save it using the web browser, then use the “<code>sed</code>” command to extract thumbnail URL's from the HTML page source code.<br />
<br />
: (command will be added here.)<br />
<br />
With <s>the help of {{W2|regular expressions}}</s> simple text replacement, you can put all “mqdefault” and “hqdefault” URL's into “maxresdefault”, although “maxresdefault” is somehow not available for all videos, thus better feed both hqdefaults and maxresdefaults into ArchiveBot.<br />
<br />
Then, you can upload them to https://transfer.sh/ or https://transfer.notkiska.pw/ and feed them into [[ArchiveBot]] using the [https://archivebot.readthedocs.io/en/latest/commands.html#archiveonly-file !ao <file] command. --[[User:ATrescue|ATrescue]] ([[User talk:ATrescue|talk]]) 13:27, 13 May 2019 (UTC)<br />
<br />
== Mass Deletion in Internet Archive??==<br />
There was more than 300,000 videos in https:// archive.org/details/archiveteam_youtube . Now just 12,000 videos, are there a mass deletion?? --[[User:Gridkr|Gridkr]] ([[User talk:Gridkr|talk]]) 14:05, 18 July 2019 (UTC)<br />
<br />
: We are not the Internet Archive. But no, they haven't been deleted, only delisted. [[User:JustAnotherArchivist|JustAnotherArchivist]] ([[User talk:JustAnotherArchivist|talk]]) 16:39, 18 July 2019 (UTC)<br />
<br />
: They are still accessible via the URLs just not through search any more. Flashfire42<br />
<br />
==Saving MORE stuff from YouTube?==<br />
<br />
On the topic of saving stuff from YouTube, are there any methods available for scraping more content such as a user's profile picture, channel banners, channel description, channel discussion page and video watermarks? I was thinking for things such as their profile picture it would be as simple as using wget, however I don't know if the url I am grabbing (such as [https://yt3.ggpht.com/a/AGF-l7_GPnSqhtxkf5pgDj4jdL3EfgJkG09iAXg3Og this one]) is the largest resolution available, especially since the URLs don't seem to follow a specific format. [[User:Systwi|Systwi]] ([[User talk:Systwi|talk]]) 18:38, 21 July 2019 (UTC)</div>Systwihttps://wiki.archiveteam.org/index.php?title=Talk:YouTube&diff=41002Talk:YouTube2019-07-21T18:38:00Z<p>Systwi: Added signature</p>
<hr />
<div>==youtube2internearchive==<br />
https://github.com/emijrp/youtube2internetarchive contains a script which also handled upload to Internet Archive, but I can't find it any longer. --[[User:Nemo_bis|Nemo]] 06:28, 26 January 2015 (EST)<br />
:I've found [https://code.google.com/p/emijrp/source/browse/trunk/scrapers/youtube2internetarchive.py something] with Google. [[User:bzc6p|bzc6p]] ([[User_talk:bzc6p|talk]]) 12:25, 26 January 2015 (EST)<br />
<br />
If YouTube needed to be quickly captured for some unforeseen reason, it might make sense to download only the XML and SRT files, so then at least some record would be saved. Google's subtitle recognition is currently far from accurate, but it's certainly improving. [[User:wtron|wtron]] 06:48, 12 June 2015 (EST)<br />
<br />
== Options ==<br />
<br />
Is it really necessary to explicitly call best A/V when youtube-dl it by default? <br />
<br />
Also, why not embed the subs and thumbnail instead of make a separate file? Also why not xattrs for those of us with unix filesystems? Xattrs is only one extra flag. <br />
<br />
My command is currently <br />
<br />
youtube-dl -t --embed-subs --add-metadata --xattrs --console-title --embed-thumbnails<br />
<br />
although I'm going to be incorporating elements from the suggested one into mine. The reasoning behind this is it's one file to send. That command is how I archive currently, it's changing though.<br />
<br />
I'd appreciate hearing your input about why I may be wrong though. Thanks in advance,<br />
<br />
--[[User:Vxbinaca|Vxbinaca]] 21:24, 29 May 2015 (EDT)<br />
<br />
On your second note, I strongly believe it's better to have different things (video, thumbnail, subtitle) in separate files. Easier to access, process, categorize, recognize. I think it's worth the "trouble" of having three files (with the same name) instead of one.<br />
<br />
[[User:bzc6p|bzc6p]] ([[User_talk:bzc6p|talk]]) 07:08, 31 May 2015 (EDT)<br />
<br />
xattrs are not portable and will get lost when copying to a file system that doesn't have them (or when uploading it somewhere, like to IA) --[[User:Darkstar|Darkstar]] 08:53, 31 May 2015 (EDT)<br />
<br />
Solid reasoning. I've now switched to your way of doing things. --[[User:Vxbinaca|Vxbinaca]] 19:32, 2 August 2015 (EDT)<br />
<br />
== --ignore-errors shouldn't be youtube-dl archiving default best practices ==<br />
<br />
Theres a myriad of reasons this isn't a good idea to have by default. Downloads getting snapped off on channel rips could go unnoticed (I search for these with ls *.part). Problems with various versions of youtube-dl could lead to a channel rip with half-processed videos, see [https://github.com/rg3/youtube-dl/issues/6413 this issue on github].<br />
<br />
Perhaps for a well-tested version that works on YouTube running in a warrior, --ignore-errors is appropriate, but for an attended rip we should by default suggest people not use it and instead just make sure all of it got ripped and if theres an error try to work resolve that particular video, and if it's a problem they can't get around then just go --ignore-errors.<br />
<br />
I'm open to being told why I may be wrong though. --[[User:Vxbinaca|Vxbinaca]] 19:32, 2 August 2015 (EDT)<br />
<br />
== Paid subscription ==<br />
<br />
On September 1, 2015 [https://www.theverge.com/2015/4/8/8371131/youtube-paid-subscription-offline-video The Verge reported] about oncoming paid subscription option (ad-free+premium videos), although paid ad-free videos may not be a nail in the coffin, previously free content could become a premium. According to the article, transition could happen in a few months -- [[User:Vitzli|Vitzli]] 06:46, 2 September 2015 (EDT)<br />
<br />
YouTube has never been profitable. [[User:bzc6p|bzc6p]] ([[User_talk:bzc6p|talk]]) 08:46, 2 September 2015 (EDT)<br />
<br />
== As we don't have 500 petabytes of free storage... ==<br />
<br />
... a solution may be to discard "low-value" videos. I mean, if we discard duplicates (films, music, etc) and we set a limit of 1 PB (1000 TB) for quality content, what could be the lucky videos that would be downloaded and preserved? We can work in a approach like this. Just because we don't have space for all, doesn't mean that we don't download anything. [[User:Emijrp|Emijrp]] 09:33, 20 October 2015 (EDT)<br />
<br />
Google will probably give a notice long before that date. So we'll have the time to find stuff worth saving. We could create a "warroom" or such, and users (ArchiveTeam members and other people) could suggest channels worth saving, with description, average views and size estimate, say, in a table, clear-cut. If someone suggests videos that are otherwise available or not too popular, they can be striked out (with proper reasoning). (The reviewing of suggestions can be done by everyone continously.) A deadline of suggestions could be given (say, 2 months before the end), and after that a Committee could select the "lucky" 1000 TB that could end up in the Archive.<br />
* In the meanwhile, the Archive itself would be queried for already saved videos, and those wouldn't be saved again.<br />
* The Archive and the Team should expect other preserving actions, and ours should be in accordance with those (no duplicates).<br />
* There could be national limits, e.g. not only a global limit, but also language or country-specific, say, 500 TB English videos, and 10–50 terabytes per other countries (just ad-hoc numbers, see the concept).<br />
<br />
The importance of the last point and the importance of saving some of YouTube at all is, I think, in the fact that, without too much exaggaration, a substantial part of today's culture is stored and represented there, on the most popular video sharing site on Earth. [[User:bzc6p|bzc6p]] ([[User_talk:bzc6p|talk]]) 15:41, 20 October 2015 (EDT)<br />
<br />
: We could also save all videos starting at a specific view count such as 500 or 1000 and/or from channels above a subscriber threshold such as 200. But maybe better first start from e.g. 1000000 views and 10000 subscribers. --[[User:ATrescue|ATrescue]] ([[User talk:ATrescue|talk]]) 22:29, 13 May 2019 (UTC)<br />
<br />
== Saving YouTube Comments ==<br />
<br />
YouTube comments are a surefire sign of just how awful the internet can be at times. Shouldn't they be archived as well? There's already a script for it, [https://github.com/egbertbouman/youtube-comment-downloader youtube-comment-downloader]. --[[User:Powerkitten|Powerkitten]] ([[User talk:Powerkitten|talk]]) 16:49, 26 October 2016 (EDT)<br />
<br />
: Hello, [[User:Powerkitten|Powerkitten]]. Thank you for mentioning this.<br />
<br />
There is also [https://github.com/philbot9/youtube-comment-scraper-cli youtube-comment-scraper-cli] which can do this as well and supports .csv and .json outputs. This tool logs comment ID, channel name, like count, post date (in both Unix timestamp and UTC date format), user profile picture and number of replies. An online version is also available [http://ytcomments.klostermann.ca/ here]. [[User:Systwi|Systwi]] ([[User talk:Systwi|talk]]) 06:53, 21 July 2019 (UTC)<br />
<br />
<br />
Now, there is some information about comment archival here:<br />
<br />
* [[YouTube#Comment_loading]]<br />
* [[Archive.is#YouTube_comment_archival]]<br />
=== Celebrations ===<br />
When the great wall of YTimg's [[robots.txt]] fell on 20120229 (leap year's day of 2012), it was a prideful feeling. Until then, http://ythistory.weebly.com/ and http://ythistory.co.cc/ were the only places where one could access the YouTube swf (flash-based) ActionScript 2 and ActionScript-3-players between circa 2008 (when YouTube's YTimg domain started, likely due to Google's acquisition) and 2012. Additionally, when browsing YouTube via the Wayback Machine, stylesheet information and all images were also blocked by [[robots.txt]], which was suddenly gone.<br />
<br />
For a few hours after it's removal, there was an error stating “Couldn't load Robots.txt for this page”, “Unable to read Robots.txt” or “Couldn't find robots.txt for this site” or something similar (I don't remember the exact words), because it was assuming that the robots.txt was just temporarily inaccessible. After these few hours past, suddently, all the YouTube pages blocked by robots.txt were suddently visible in their full beauty.<br />
<br />
Now, because one of the biggest Wayback dreams I and {{URL|2=other people|1=https://archive.org/post/406632/why-does-the-wayback-machine-pay-attention-to-robotstxt}} had apparently has come true (read more at [[Internet_Archive#robots.txt_and_the_Wayback_Machine]] too), I can now load YouTube comments through the Wayback Machine! ([http://web.archive.org/web/20190512223447/https://www.youtube.com/watch?v=W3GrSMYbkBE&lc=UgxC238Gea0KGOditl54AaABAg Try here]) --[[User:ATrescue|ATrescue]] ([[User talk:ATrescue|talk]]) 22:52, 12 May 2019 (UTC)<br />
<br />
: @Powerkitten: http://ytcomments.klostermann.ca/ (title element={YouTube Comment Scraper}; title on the webpage={Download comments from YouTube};) can be used to get all the comments on a video, which can be downloaded as JSON and CSV at http://ytcomments.klostermann.ca/scrape. You can then upload those files to archive.org. --[[User:Usernam|Usernam]] ([[User talk:Usernam|talk]]) 10:57, 20 May 2019 (UTC)<br />
<br />
: archive.is can also archive replies to comments, given you have the link to a reply to a comment. --[[User:Usernam|Usernam]] ([[User talk:Usernam|talk]]) 00:17, 15 July 2019 (UTC)<br />
<br />
== Many YouTube video items are archived in low quality because they were archived using an old version of tubeup. Is there a way I can replace the files of old youtube items (using the "youtube-[id]" naming format the tubeup uses) with updated, high quality files? --[[User:Hiccup|Hiccup]] ([[User talk:Hiccup|talk]]) 17:01, 6 February 2018 (UTC) ==<br />
<br />
Many YouTube video items are archived in low quality because they were archived using an old version of tubeup. Is there a way I can replace the files of old youtube items (using the "youtube-[id]" naming format the tubeup uses) with updated, high quality files? --[[User:Hiccup|Hiccup]] ([[User talk:Hiccup|talk]]) 17:01, 6 February 2018 (UTC)<br />
<br />
== Thumbnail rescue ==<br />
<br />
Some YouTube uploaders change thumbnails of their videos, which would lead to the loss of the existing thumbnail if archived nowhere else.<br />
<br />
Some YouTube thumbnails can be retrieved by searching the video ID using web search (try different search engine).<br />
<br />
To save as many thumbnails as possible from a specific channel, please archive the video page of their channel using the help of [[chromebot]]'s page scrolling.<br />
<br />
One can also manually scroll down a page and save it using the web browser, then use the “<code>sed</code>” command to extract thumbnail URL's from the HTML page source code.<br />
<br />
: (command will be added here.)<br />
<br />
With <s>the help of {{W2|regular expressions}}</s> simple text replacement, you can put all “mqdefault” and “hqdefault” URL's into “maxresdefault”, although “maxresdefault” is somehow not available for all videos, thus better feed both hqdefaults and maxresdefaults into ArchiveBot.<br />
<br />
Then, you can upload them to https://transfer.sh/ or https://transfer.notkiska.pw/ and feed them into [[ArchiveBot]] using the [https://archivebot.readthedocs.io/en/latest/commands.html#archiveonly-file !ao <file] command. --[[User:ATrescue|ATrescue]] ([[User talk:ATrescue|talk]]) 13:27, 13 May 2019 (UTC)<br />
<br />
== Mass Deletion in Internet Archive??==<br />
There was more than 300,000 videos in https:// archive.org/details/archiveteam_youtube . Now just 12,000 videos, are there a mass deletion?? --[[User:Gridkr|Gridkr]] ([[User talk:Gridkr|talk]]) 14:05, 18 July 2019 (UTC)<br />
<br />
: We are not the Internet Archive. But no, they haven't been deleted, only delisted. [[User:JustAnotherArchivist|JustAnotherArchivist]] ([[User talk:JustAnotherArchivist|talk]]) 16:39, 18 July 2019 (UTC)<br />
<br />
: They are still accessible via the URLs just not through search any more. Flashfire42<br />
<br />
==Saving MORE stuff from YouTube?==<br />
<br />
On the topic of saving stuff from YouTube, are there any methods available for scraping more content such as a user's profile picture, channel banners, channel description, channel discussion page and video watermarks? I was thinking for things such as their profile picture it would be as simple as using wget, however I don't know if the url I am grabbing (such as [https://yt3.ggpht.com/a/AGF-l7_GPnSqhtxkf5pgDj4jdL3EfgJkG09iAXg3Og=s288-mo-c-c0xffffffff-rj-k-no this one]) is the largest resolution available, especially since the URLs don't seem to follow a specific format. [[User:Systwi|Systwi]] ([[User talk:Systwi|talk]]) 18:38, 21 July 2019 (UTC)</div>Systwi