We are a LauncHUB funded Startup that focuses on big data research of domain names and Whois information. SEO experts can use our unique system for finding expired domains with authority and backlinks. Investigators can use the same database to find information for domain owners and hosts with Reverse Whois and Reverse IP tools. We can also offer brands trademark monitoring in the vast pool of ccTLD domain names.
When the domain drops, the minute it drops, literally, they will try to get it and if you can get it, it’s yours. So, that’s it for expired domains. You can see it’s a little bit complicated and it takes some leg work, but if you’re interested in building a publisher network or a private blog network, this is a great way to find sites that have a great link profile without you having to actually build any links. So, that’s it for this video. I’ll see you in the next one.
What I like to do is sort by DP which stands for domain pop, and this is basically the number of linking root domains. So, BL is the number of back links. As you know that can be somewhat misleading if they have a lot of site wide links or multiple links from the same domain. I like to sort by domain pop. What that does is it brings up the sites with the most amount of referring domains.
Hi, I purchased a .org domain in a Godaddy auction, actually the closeout auction so it had been renewable and then available as a closeout after 36 days. I purchased the closeout domain because I was thinking of using it as a catchy anagram. I instead offered it to the Trademarked company to purchase for a nice sum of money, and now they are threatening to file legal action against me for “Domain Squatting” saying that I purchased it in bad faith. etc… The way I see it, is it was available for them to renew for 26 days, and if it was not owned by them they had a window of 10 days to purchase the name before I did in the regular auction. What’s your thoughts on this?
So, let's start off with the simple website list crawl. So settings for this is covered by the general crawl settings and these apply to all the other types of crawl as well, such as the Search crawl and the Endless cruel, so pretty simple really. Delay between each request to a website. One second, this is in seconds. Secondly, concurrent websites crawl how many websites you want to crawl at any one point in time and then how many threads will concurrently crawl per website. So that's ten, a crawl of ten websites at once, and each of those websites there's three different threads crawling. That's 30 concurrent connections you've got going.
Okay, so let's run it. I'm just going to copy those in there to give it a bit of a kick start, but I'm on modern fiber UK fiber, so that might actually be too slow still, but yeah just trial and error with the amoount of crawlng threads, because it depends what kind of resources your PC has, and what kind of connection you're on. Okay, so we'll come back in a minute, oh by the way also, you can get the DA and the PA for each result for any of the crawl types. As long as you put in your MOZ key in the settings, so I just paused the video because I want to play around with the settings, because I haven't used this on my home fibre and I found on my UK fibre, these things worked pretty well but, like I said it will vary from connection to connection and machines to machine as well. So because I want some results to show you pretty quickly. I swapped it for website list and we're going to crawl some directories because they're always really good for finding tons of expired domains.
Let’s just say for example we wanted this site, crystalgiftsworld.com. It looked good based on our analysis and we’d head over to Snap Names, Nameja is another one and what these service do is they have special technology on their site that allows them to try to register a domain on your behalf over and over again. Okay? If you tried to do that your IP would get banned, but they have some system where they know how to do it just enough to get the domain, but not enough to get blacklisted.
The whitelist filter for wanted characters is now split into three filters. The old whitelist is now named whitelist (only). When used, the domain has to consist only of the defined characters or numbers. No other characters are allowed. The new whitelist (any) filter finds domains that have at least one of the defined characters or numbers. The new whitelist (all) filter finds domains that have all defined characters or numbers at least once.