You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
If a url has been already crawled and it was a 20x or 30x then we know it was good, and we are simply checking for regressions.
If it was a 40x then it is either a regression we want detected which might possibly be fixed on the other end, but more likely it is just the link itself which is broken and needs to be fixed.
50x on the other hand are temporary and a clear signal that we should try again and get a different response.
So suggesting that the previous return code should be a soft weighting factor in the queue prioritization so that 50x are done earlier to try and clear them.
The text was updated successfully, but these errors were encountered:
If a url has been already crawled and it was a 20x or 30x then we know it was good, and we are simply checking for regressions.
If it was a 40x then it is either a regression we want detected which might possibly be fixed on the other end, but more likely it is just the link itself which is broken and needs to be fixed.
50x on the other hand are temporary and a clear signal that we should try again and get a different response.
So suggesting that the previous return code should be a soft weighting factor in the queue prioritization so that 50x are done earlier to try and clear them.
The text was updated successfully, but these errors were encountered: