Although I have one query (this answer does answer it, but I have a doubt)… I just bought a domain which had the creation date as 2010 but after acquisition upon checking I found that it has been reset to 2014. So, from what I understand from your article, this could not be undone anyways because this particular domain’s status said as ‘Expired’ and I got it for under $10 from godaddy?
If you are pursuing option 1, then you will want to use an automated service for the reasons mentioned in the article. Follow the process in the article to see which registrar the domain name is located at. If it’s at a Auction House like GoDaddy, NameJet or SnapNames, I believe they publicly share if there is a backorder on a domain name but not who placed that backorder. (Any interest in a domain name may be enough for someone else to place a backorder, thereby driving interest in the domain name that may be visible to the registrant.) So if the domain is at a registrar partnered with one of these companies, then you’ll have to monitor it manually (via calendar entries based on the whois expiration data) to maintain your distance.
Today, the story is different. Domain name registrars realized that they could auction expired domain names to the highest bidder and generate additional revenue. If no one wanted the domain names in an auction, the domains would then drop and become available for anyone to register. Much of the time, however, domain names are successfully auctioned.
So, let's start off with the simple website list crawl. So settings for this is covered by the general crawl settings and these apply to all the other types of crawl as well, such as the Search crawl and the Endless cruel, so pretty simple really. Delay between each request to a website. One second, this is in seconds. Secondly, concurrent websites crawl how many websites you want to crawl at any one point in time and then how many threads will concurrently crawl per website. So that's ten, a crawl of ten websites at once, and each of those websites there's three different threads crawling. That's 30 concurrent connections you've got going.
It’s easier to rank higher with expired domains: Depending upon your domain age, page authority, trust flow, expired domains tend to rank really well when compared to new domains. Most new domains don’t have any DA, page authority or trust flow so it’s really harder to rank well in Google. This won’t happen with the expired domains as they are already established well and might be getting decent amount of traffic from search engines.
When searching from keyword DHG will take your list of keywords and search each one, the pages will then be crawled to find unique domains to check availability of. The process is relatively simple but very time consuming when done manually or semi automated as was necessary in the past. Luckily now the job can be set and left to work while you do something else.
I am interested in a domain name that is set to expire in April 2016. The domain is registered with Domain.com, so I have placed a backorder on Snapnames.com (thanks for confirming I at least did that part right). I also placed backorders with NameJet.com and Pool.com, based on previous advice I received. Now I’m wondering if that might be overkill since Snapnames is the designated auction house. Will I ultimately end up bidding against myself?
Also note the Expired Domain Plugin is a lifetime “per user” license, so if you have multiple ScrapeBox licenses registered to the one email address purchasing the Expired Domain once will activate the plugin for all your ScrapeBox licenses with the same email for life. The plugin is a one-time payment and is not a monthly or yearly subscription, all updates are free.