| Click here to select a new forum. |
| restoring 68kmla from google and archive.com web caches |
Posted by: Dog Cow on 2010-01-28 13:32:49 I've written a script which spiders phpBB forums and extracts all posts, users, and forums into database tables. It even does a so-so job at converting the HTML back into BBCode. It then has a front-end for the database, and it can be searched too using MySQL Fulltext index.
|
Posted by: johnklos on 2010-01-28 13:46:24
I've written a script which spiders phpBB forums and extracts all posts, users, and forums into database tables. It even does a so-so job at converting the HTML back into BBCode. It then has a front-end for the database, and it can be searched too using MySQL Fulltext index. Would you like to run that into a database on one of my machines so we can see about creating something off-site, and if it looks good and clean, we can see about how to get it into 68kmla?
John
|
Posted by: Dog Cow on 2010-01-28 15:31:30 Sure. Where would the HTML pages come from? Are they in a local directory, or somewhere on the web?
|
Posted by: Mars478 on 2010-01-28 15:38:23 Make a second archive. Truly, all that data is invaluable. I would be willing to set up an extra server at my house in NJ which we go on only the weekends. The server would be a BW G3.
|
Posted by: Osgeld on 2010-01-28 15:40:34 and if you cant get it back into form, ive got quite a few scripts that will redo all the links in a static html spider dump so you can rehost it anywhere you want (including locally, which is not really a requirement here...) , I did this quite a lot back when I was mirroring the SecondLife scripting wiki (every 3 months)
http://www.cheesefactory.us/lslwm
http://www.cheesefactory.us/slwm
|
Posted by: Mars478 on 2010-01-28 16:05:37 Kinda feel awesome that I resparked this movement. 🙂
|
| < 2 |