-
-
Notifications
You must be signed in to change notification settings - Fork 995
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Question] What will happen if there is an update during download? #368
Comments
I wish it could be better like set |
I'm having a problem:
Edit:
Edit 2: |
Tumblr's API has a rate limit of 1000 (?) requests per hour and there is even some special logic in place to wait until this limit has recovered, but I've managed to "break" it by changing stuff elsewhere. Sorry. I'll fix this ASAP.
Yes, you can use the timestamp of the last post you managed to scrape as Regarding your questions
No, if anything you will get duplicate content. All posts get moved "one ahead" and the last post in a list of 50 will reappear as the first one in the next list of 50 if a new one gets added at the beginning.
50 posts per API request is the maximum. Setting a higher number has no effect and still only returns data for 50 posts. |
Exceeded error doesn't return exit code. I was planning to a loop and break if fail: TUMBLOG=kwwwsk; \
for YEAR in {2019..2004}; do \
gallery-dl -o date-max=${YEAR}-01-01T00:00:00 -o date-min=$((YEAR - 1))-01-01T00:00:00 -vj https://${TUMBLOG}.tumblr.com > ${TUMBLOG}-${YEAR}.json || break; \
done It doesn't work. |
Let's say, I'm scrapping the JSON data from a tumblr blog that has 500k+ posts.
Eventually there will be an update, because of post schedule and scrapping took hours.
Is there will be a skipped content? A content that skipped and didn't count on the API. Because what I see is tumblr using offset and it's limit.
The text was updated successfully, but these errors were encountered: