Membership Level: Demonstration Site
Not Logged In: Demo ONLY
We often have large lists of URL's that have extra text either in front of, or after, the physical web address.
These sometimes come from systems outputs, where the PageRank has been suffixed, or there is link/anchor information following the http://address. Often they're submission system or account creation exports that are part of a larger report, and they can't be seperated out to just leave the URL on it's own.
So... We created this tool to handle all the situations when we need to either strip characters from the front or the end of a line. And it doesn't just have to be just URL's; it can be any text lists which need processing.
Remove the first characters and/or the last characters from each line... OR...
Strip everything up to (and including) the first instance of , and everything after (and including) the last instance of .
The URLs must be the ONLY thing on the line for the following options to work; so clean off any other chars first...
Make sure http://... IS at the start IS NOT at the start Trim URL's to root domain De-Duplicate List
Note: With URL's we add an extra unique function - where we trim down to the 'root' domain. This simply means we remove the final part of the URL after the domain. This is useful in a lot of situations where we want to take a quantity of URL's and drill down to the unique root domains in that list. We can also 'de-dupe' or remove the duplicates to cull the list down. (This aspect works in a similar way to ScrapeBox.)
This tool is also useful for getting URL lists 'in-line' with http:// syntax. If some of your URL's have it, and others don't, then you can drop them in here and choose if you want the http:// at the start or not. Generally, for URL processing, you want the http:// there. Standard members can process 500 URL's/Lines at one time; Premium members can process 2,500 at one time.
Click HERE To Display/Hide NOTES Below...
When performing any of the bottom line of URL functions, make sure you've stripped out any extra text first, so that each line is purely a URL. The tool has no way of telling if there are 'spelling' or syntax mistakes in a URL, so if you have odd characters at the start, and ask it to ensure that it has http:// at the start, then it will add it!
You can also run the bottom 3 options simultaneously; so you could drop in 500 URL's, ensure they're all http:// encoded, trim them all to root domains and then de-dupe them... All at once.
It performs the functions in the order they're shown on the page: Top-down. So you can theoretically strip start/end chars and perform the bottom URL functions all at once; but it's probably easier to do things one/two steps at a time. You can simply copy the output from this page back into this page afterwards and re-process in another way.
If you use de-dupe and trim to root at the same time, it will trim to root first and then de-dupe.
If you're using the http:// function, you have to first click the checkbox to 'use' it - and thenalso select which option you want (IS there/Is NOT there.)
You can also use SPACES as delimiters in the 'Strip everything up to (and including)...' function.
The De-dupe function also works on any plain text list as well. This can be useful if you have a long list of keywords from a keyword tool (1 per line) and want to remove all the duplicates. Most keyword software has this functionality, but if you're doing it manually - you can do it here!