Project Noahs Ark

@Movation I’m not sure these will work, plus it will take for ages no matter if it works or not. It will also put everything under a lot of stress and cost a lot of traffic. Also lots of stuff isnt worth saving to be honest so its not that much work when you selectiviely do it the way described above. I already don’t know what to get next and it not that I DL much :slight_smile: If you could get the actual db dump it would be awsome of course, but just a big amount of HTML without search capabilities doesn’t serve us well.
@Juddymuddy Is this an export of your stuff? Looks like it. If not, how did you generate this?

1 Like

Used a web scraper for it

not any of my stuff

Without pictures its useless im afraid.

I might be able to setup pictures but will take me some time

1 Like

Don’t get me wrong, I’m a computer scientist, I love automatic routines for doing stupid stuff like that but I’m not sure its worth it unless we could reproduce the whole forum for import to this one or create a new installation with the data. All those grabbers produce are static HTML pages, lots of them, or like in your case a giant unreadable spreadsheet, so you have the info that is actually worth saving somewhere buried in a big mess. Then we could just do as I described and save selected threads but in a whole and have a catalogue page up front so you can at least find it quick.

2 Likes

@longhairedboy and @jamie are leaders on the builder forum maybe they would have access to an API?

1 Like

Couldn’t we use archive.org as a temporary solution to back up the entire website? If someone makes a snapshot of it…

or if someone is willing to risk it just buy the site from Enertion… Domain, hosting, data, everything.

Would be the most ideal but most complex solution

i’m still a leader over there? i thought i would have been removed by now. You know, because i stole the community like fucking Carmen San Diego.

9 Likes

grab-site using cookie file option to WARC, compatible with discourse, maintains a lot of the JS functionality and references all URLs, can be rehosted using a nginx server

Probably 1-3Tb, really stretching limits of the s/w and your IP will get blocked before it finishes.

I gave up and restarting the grab requires further investigation

4 Likes

Might have to use a proxy to finish it or if you already have a good vpn maybe use that

save whiteponies build logs. pure eskate porn. :eggplant::sweat_drops:

3 Likes

And bring him back too!

1 Like

If you want it saved, do it and post it here like I described above :wink:

1 Like
7 Likes
Summary

[SRO]

2 Likes

I have a considerable amount of storage and bandwidth available and will attempt to mirror the entire site in its current form.

9 Likes

Jason loves you

1 Like

another option for site backup and then usage after backup is httrack.

I use this all the time. for instance I have the entire documentation site for ardupilot on my laptop for use offline in the field.

the interface is simple but powerful and the end result works just like the original website.

there is an android app as well.

it grabs pictures and html links to videos (yt) as well.

it’s free and open source

https://www.httrack.com

1 Like