Weāre setting up a load of redirects in an āinterestingā way and I donāt want to do check 100+ links manually
Anyone know of a handy standalone tool or easiest/quickest process that I can use for the followingā¦
get child URLs by crawling a domain (e.g. A .com/stuff[n])
visit each URL in turn (e.g. /stuff[1], /stuff[2], etc.)
record the resulting URL loaded (after redirection) and final status code (e.g. A .com/stuff[1] loads A .com/thing[1] with 200 response)
Thereās quite a few tools I can find for generating site maps or scraping links. But not as easy to find stuff that will ease the checking of the redirection. Iāve found HEADMasterSEO.com which looks potentially helpful - but would love advice.
Oh and just because Iām fussyā¦ if the tool or method takes longer to install/setup/run than manually going through the list twice (once on staging & once on live) then itās not going to help as this is a non-repeating activity and Iām trying to save time/brain power
Hmmm that works for your example, but not for my situation.
When I do curl -I <my_URL>
I get a 200 OK and no alternative location.
As I mentioned, the redirects are being done in an āInterestingā way
Are we still talking about HTTP redirects or URL proxying/Route mapping? Those are two separate things. For the latter I donāt have a run-of-the-mill solution.
Iām not entirely sure - I know itās been set up at the AWS S3 level as there are limits in place preventing it being done as a web server redirect.
So this is why HeadmasterSEO is winning so far - as Iām approaching this very much as an end user sticking in URLs, and get a report back on final loaded URL and status code
Oh, gotcha. Thatās more of a spidering tool, similar to Screaming Frog SEO Spider. We use it (or command line tool wget) to generate a list of URLās we can then feed into our little script I sampled earlier. This way we can automate the testing part.
Bizarrely Iāve been asked to do this exact same thing this week by our business usersā¦
I have setup a basic TESTNG framework based on something I saw at https://www.swtestacademy.com/data-driven-excel-selenium/. The business provide a excel sheet ( they lurve excel) and this will run through the sheet chcking the url , the expected return code and the redirect (from Location header).
I switched to using restAssured rather than selenium as its a bit more suited to this job.
This is now run from our jenkins every time a content change or category rejig is done. This same spreadsheet is provided to the webmasters.
The weird synchronicity of the tech world
Like the sound of your solution - will bear it in mind if there is ever the need to do similar here (there has been talk about it after changing of content titles etc.)