Difference between revisions of "FTP"
(Undo revision 19109 by Dashcloud1 (talk)) |
Dashcloud1 (talk | contribs) (Add both choices to the page again) |
||
Line 30: | Line 30: | ||
Check the size of the site before you start to make sure you have the space to hold the site and tar afterwards, also account for large files on the site when using <code>tar --remove-files</code> | Check the size of the site before you start to make sure you have the space to hold the site and tar afterwards, also account for large files on the site when using <code>tar --remove-files</code> | ||
lftp -c 'set ftp:use-feat no; du - | lftp ftp://site.com -e 'du -h' | ||
An alternate to try if the above does not work correctly (happens more often on old servers): | |||
lftp -c 'set ftp:use-feat no; du -h ftp://site' | |||
Now zip/tar it up and [[Internet_Archive#Uploading_to_archive.org|send to the spacious Internet Archive]]![https://archive.org/details/ftpsites] (If you're short on space: <code>tar --remove-files</code> deletes the files shortly after adding them to the tar, not waiting for it to be complete, unlike <code>zip -rm</code>.) | Now zip/tar it up and [[Internet_Archive#Uploading_to_archive.org|send to the spacious Internet Archive]]![https://archive.org/details/ftpsites] (If you're short on space: <code>tar --remove-files</code> deletes the files shortly after adding them to the tar, not waiting for it to be complete, unlike <code>zip -rm</code>.) |
Revision as of 00:38, 18 June 2014
FTP | |
Status | Online! |
Archiving status | Not saved yet |
Archiving type | Unknown |
Project source | https://github.com/ArchiveTeam/ftp-nab |
IRC channel | #effteepee (on hackint) |
Archiving a whole public FTP host/mirror is easy:
SketchCow> I use wget -r -l 0 -np -nc ftp://ftp.underscorporn.com tar cvf 2014.01.ftp.underscorporn.com.tar ftp.underscorporn.com tar tvf 2014.01.ftp.underscorporn.com.tar > 2014.01.ftp.underscorporn.com.tar.txt
OR, use this handy dandy function to put in your .bashrc file, you can also remove the first and last line to turn it into a fancy bash script. Made by SN4T14
ftp-grab(){ target="$1" wget -r -l 0 -np -nc "$target" if "$target" =~ ^ftp://.*$ then target="$(echo "$target" | cut -d '/' -f 3)" echo "ftp" echo "$target" fi tar cvf $(date +%Y).$(date +%m)."$target".tar "$target" tar tvf $(date +%Y).$(date +%m)."$target".tar > $(date +%Y).$(date +%m)."$target".tar.txt }
Check the size of the site before you start to make sure you have the space to hold the site and tar afterwards, also account for large files on the site when using tar --remove-files
lftp ftp://site.com -e 'du -h'
An alternate to try if the above does not work correctly (happens more often on old servers):
lftp -c 'set ftp:use-feat no; du -h ftp://site'
Now zip/tar it up and send to the spacious Internet Archive![1] (If you're short on space: tar --remove-files
deletes the files shortly after adding them to the tar, not waiting for it to be complete, unlike zip -rm
.)
The Project
- We're currently listing all FTP sites on the internet to download them all.
- We're auding a list of some select FTP sites manually: https://www.piratepad.ca/p/old-ftp-list
Midas | ftp.tu-chemnitz.de |
Midas | ftp.uni-muenster.de |
Midas | gatekeeper.dec.com |
Midas | ftp.uni-erlangen.de |
Midas | ftp.warwick.ac.uk |
Uni FTP's are massive, currently only grabbing DEC and Sweex.