-
Notifications
You must be signed in to change notification settings - Fork 130
[Feature Request] Parse the regular website instead of using the tumblr api #33
Comments
For people only interested in photos and videos, here we go scraping the website. Might be faster since the scan connections can be cranked up in the settings as it does not depend on the Tumblr api anymore.
|
That's definitely preferable for me. I only want pictures, video, and gifs. I assume photos includes gifs? FWIW, 1.0.4.31 never successfully downloaded anything for me coming from my 1.0.4.18 index files, although that doesn't matter much if the non-api verison works. I'll test in a bit. |
I've rebased it again to the current master release (the one that uses the v1 api) @bb20189. Nothing is tested at all. I too think that's the way to go because we could potentially access private blogs, no connection limits. It's probably even worth considering to just use RegEx and filter all jpg/png/gifs instead of using html nodes. It's much less a hassle to write the code and even if tumblr changes their layout, nothing should be broken if we just scan the whole html for specific tags. I think it does not detect photosets nor inline images yet. |
1.0.5.2 and 1.0.4.35 cannot load my index files. "Error 1: Could not load file in library:" when I launch the program I also just realized that 1.0.4.18 must have messed up some of my index files last month when I ran it....I have blank "Downloaded Files"(But "Number of Downloads" is still populated) for some long running blogs, and the index files are very small. When I try to download again with 1.0.4.18, sure enough, it downloads every post ever instead of just the ones from the last month. I made sure to never expose said index files to any of the newer versions, I keep a backup, and a good number of the other blogs seem fine. Argh, not sure how to deal with this.... |
1.0.5.1 seems to work with those broken index files though....I assume because it's only grabbing the posts since the last completion date? While that idea is great in theory, I worry about bugs, like if the program crashes mid crawl or something. Is that version safe to use for now? If I use that, will my index files become incompatible moving forward? |
So, we could also use a WebDriver, which essentially is a chrome/firefox without the UI. That would also simplify the private blog access. It probably uses lots of memory and fattens the application a lot, but it sure is the most simple implementation for now. |
Hello, i do not seen any settings or things about this no-use of API, |
The rate limiter is not included in this release, thus it's not limited and just connects to the normal website depending on the remaining connection settings. You can adjust the scan connections with the scan connection settings and if you want to include the scanning connections in the bandwidth throttler. |
ok tks i understand now ;) |
I've uploaded a branch here where my first steps are visible. If someone wants to participate, I think the hard thing is already done: How to parallelize it?
By using the archive page it's possible to determine the lifetime of the blog. Now we can access the archive page by using _archive?before_time= where the before time is unix time. So we could start the crawl at several times during the blogs lifetime. Say xxx crawls per month and stop the crawlers once it hit a post of a different crawler started at an earlier time.
I've uploaded a functional commit for image crawling without accessing the tumblr api. Still needs optimization and a code rebase to the current master branch.
The text was updated successfully, but these errors were encountered: