Difference between revisions of "RapidShare"

From Archiveteam
Jump to navigation Jump to search
(added info on discovery)
Line 30: Line 30:


=== Discovery ideas ===
=== Discovery ideas ===
*Try scraping [http://rapid-search-engine.com rapid-search-engine.com] for download links or ask its owners for the database
*Try scraping [http://rapid-search-engine.com rapid-search-engine.com] for download links or ask its owners for the database '''In progress'''
*Get a list of valid download links [http://www.reddit.com/domain/rapidshare.com from reddit.]
*Get a list of valid download links [http://www.reddit.com/domain/rapidshare.com from reddit.] '''Done'''
*Scrape Google with the keyword "intext:rapidshare.com/files".
*Scrape Google with the keyword "intext:rapidshare.com/files".
*Scrape URLs from the Wayback Machine.
*Scrape URLs from the Wayback Machine. '''In progress'''
*Search in the expanded URLs derived from URLTeam?
*Search in the expanded URLs derived from URLTeam? '''URLs have been grabbed from the torrent, URLTeam 2 results haven't been searched yet'''


== How can I help? ==
== How can I help? ==

Revision as of 17:42, 30 March 2015

RapidShare
RapidShare logo
RapidShare-homepage.png
URL https://www.rapidshare.com
Status Closing
Archiving status In progress...
Archiving type Unknown
Project source rapidshare-discovery, rapidshare-grab
Project tracker rapidsharedisco, rapidshare
IRC channel #rapidscare (on hackint)

RapidShare is a file sharing website, shutting down on March 31, 2015.

Shutdown notice

Notice

Dear RapidShare Customers

Kindly note that RapidShare will stop the active service on March 31st, 2015. Extensions of STANDARD PLUS and PREMIUM will be possible until February 28th, 2015.

We strongly recommend all customers to secure their data. After March 31st, 2015 all accounts will no longer be accessible and will be deleted automatically.

If you have any questions, please do not hesitate to send us an e-mail to support@rapidshare.com

Thank you for many years of trust.

Your RapidShare Team

Site structure

Discovery ideas

  • Try scraping rapid-search-engine.com for download links or ask its owners for the database In progress
  • Get a list of valid download links from reddit. Done
  • Scrape Google with the keyword "intext:rapidshare.com/files".
  • Scrape URLs from the Wayback Machine. In progress
  • Search in the expanded URLs derived from URLTeam? URLs have been grabbed from the torrent, URLTeam 2 results haven't been searched yet

How can I help?

Discovery

Running a Warrior

You can start up a Warrior and there select RapidShare Discovery. (If you don't really care what you are archiving, select ArchiveTeam's Choice instead, as at some points ArchiveTeam may priorize another project.)

Running the script manually

If you use Linux and you're a bit familiar with it, you can try running the script directly.

The instructions can be found at github.com/ArchiveTeam/rapidshare-discovery. Using concurrency of more than 1 (or even just 1) may result in a permanent ban from rapid-search-engine.com!

Some additional information
Don't forget to replace YOURNICKHERE with your nickname.

If you want to stop the script, please do it gracefully if possible. To do so, create an empty file named STOP in the folder of the script (terminal command: touch STOP). The script finishes the current item(s) and stops only after that. (If you kill the script immediately, the items get broken, and they will need to be reassigned to another user.) – Before starting the script again, don't forget to remove the STOP file.

If you see "Project code is out of date", kill the script, go to its folder (cd rapidshare-discovery) and issue git pull https://github.com/ArchiveTeam/rapidshare-discovery. After the updating has finished, re-launch the script.

Content grab

Running a Warrior

You can start up a Warrior and there select RapidShare. (If you don't really care what you are archiving, select ArchiveTeam's Choice instead, as at some points ArchiveTeam may priorize another project.)

Running the script manually

If you use Linux and you're a bit familiar with it, you can try running the script directly.

The instructions can be found at github.com/ArchiveTeam/rapidshare-grab. Using concurrency of more than 10 may result in a ban from RapidShare!

Some additional information
Don't forget to replace YOURNICKHERE with your nickname.

The number after --concurrent determines how many threads run at the same time. You can increase this number if your resources (RAM, CPU, bandwidth) are sufficient. However, if you constantly see messages about rate limiting, there is no need to increase the concurrency. When archiving RapidShare, don't increase the concurrency above 10 on one IP, otherwise you'll be banned. (10 seems to be okay, but the safe number might be lower.)

If you want to stop the script, please do it gracefully if possible. To do so, create an empty file named STOP in the folder of the script (terminal command: touch STOP). The script finishes the current item(s) and stops only after that. (If you kill the script immediately, the items get broken, and they will need to be reassigned to another user.) – Before starting the script again, don't forget to remove the STOP file.

If you see "Project code is out of date", kill the script, go to its folder (cd rapidshare-grab) and issue git pull https://github.com/ArchiveTeam/rapidshare-grab. After the updating has finished, re-launch the script.

Donating to the Internet Archive

Content downloaded by the ArchiveTeam will be uploaded to the Internet Archive, where it will be stored and be available – hopefully – forever. However, storing it costs thousands of dollars in the long run. So, if you can afford, please consider donating to the Internet Archive, so that this piece of history can be kept for us all. http://archive.org/donate

Do you like our cause?

If you want to help in other projects, want to learn more about ArchiveTeam, or even help in development in general, navigate to the Main Page of this wiki, from there you can reach a lot of information. The Team consists of volunteers working on the projects in their free time, so helping hands (and resources) are always welcome.