What's the best way to scrape part of a website for offline reading? I'm on GNU+Linux so preferably something that isn't too complicated for a relative noob. I want to have all of Warosu's /g/ saved on my computer.
here comes the plane, open your mouth~wget >>60818943
Some kid actually turned this in for a grade. The fact that we have become this outrageously stupid as a species makes me furious.
>>60818943
php file_get_contents(), $dom = new DOMDocument(); $xpath = new DOMXpath()
easy as shit
>>60818986
Some kid actually thought these pictures are real. The fact that we have become this outrageously stupid as a species makes me furious.
>>60818997
Even if you're only pretending to be retarded, you're still being retarded.
>>60818990
How do I translate this into a command? I want this (https://warosu.org/g/) with every thumbnail, image, post, thread, etc saved on my hard drive. I'm an Ubuntu user so I'm not too familiar with the CLI lingo.
>>60819006
He's right though. These pictures are old as hell and fake as fuck. They're still funny though.
>>60818943
>>60819072
>I want to have all of Warosu's /g/ saved on my computer.
Not happening. Don't bother.
>>60819221
Too large? I have a lot of storage space. I wonder if I should just find some neet on /r9k/ to pay and have him manually save every thread one by one.
>>60819072
build a crawler with Python
>>60819072
can you write js? python? ruby?
1. grab the pages
2. parse it
3. ???
4. profit!