Saturday, December 02, 2017

Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

Posted By: BP Media - Saturday, December 02, 2017

Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

Google harus membenci ScrapeBox. Sudah lama menjadi alat favorit SEO topi hitam. Tapi sekarang alat yang terkenal ini menemukan kehidupan baru sebagai penghemat waktu yang sangat baik untuk SEO topi putih. Dalam tutorial yang mengungkapkan ini, XPRTL LRT dan studi kasus yang luar biasa, Bartosz Góralewicz, menunjukkan kepada kita bagaimana menggunakan alat 'hitam' ini untuk membuat Google sangat senang dengan situs Anda. Ditching Excel hanyalah salah satu dari banyak alasan bagus untuk menggunakan ScrapeBox untuk melakukan pengangkatan berat pada audit tautan berikutnya. Jika Anda masih belum mengaudit situs Anda, mohon bantuan kewaspadaan Anda dan audit tautan Anda secepatnya. berikut cara penggunaannya : Introduction
The SEO community seems to be divided into two groups: fans of manual work and link processing with e.g. Excel, and fans of tools made to speed up the process, like Link Research Tools and ScrapeBox, created specifically to speed up working with backlinks. Obviously, I am not a huge fan of Excel myself. ScrapeBox was one of first automated Black Hat tools. Nowadays I doubt that anyone is still using ScrapeBox for mass blog commenting. Not because comment links are not that powerful anymore, but mostly because there are many more advanced Black Hat SEO tools. Now, after 5 years, ScrapeBox is making up for all the SPAM issues from the past by helping with link audits, White Hat link building and on-page SEO work. Why ScrapeBox? Scrapebox is really cheap – it costs only $57 ($40 discout price) when bought with the BlackHatWorld discount here: http://www.scrapebox.com/bhw It is a one-time-off payment for life It is easy to use Scrapebox overview As you see, the tool itself looks quite simple when opened. Now let me explain the most important fields step by step. Scrapebox overview explained You can do most of the list processing from the main window above. For some more advanced actions, we need to go to the “Addons” tab. Processing the link lists Let’s start with the basics. As a good example, each one of you has been in a situation where you were just flooded with link reports from an ex SEO agency, exports from Google Webmaster Tools, MajesticSEO, Ahrefs etc. I bet each and every one of you that having this organized quickly has a huge value, but only when you can do it within 5 minutes. Let’s see what we can do with such a situation step by step. 1. Removing duplicate links Instructions: Open ScrapeBox Click “Import URL list”
Scrapebox add url Click Paste/Add from Clipboard (or “Import and add to current list” if you want to import from a TXT file). Repeat with each of the lists you have After you’ve pasted all the URLs, click Remove/Filter Scrapebox remove duplicates Click Remove Duplicate URLs Scrapebox remove duplicates Now you’ve got a list of unique URLs that you can save, and use only 1 file for all your reports. 2. Removing duplicate domains Instructions: Open ScrapeBox Click “Import URL list” import urls Click Paste/Add from Clipboard After you’ve pasted all the URLs, click Remove/Filter remove duplicate domains Click Remove Duplicate Domains remove duplicate domains results Now you can see the list of unique domains. ScrapeBox will show you the popup window with a report about the amount of duplicate domains removed. Personally – I like to keep stuff tidy. I would rather see all of those domains in a top domain format only. So Domain.com instead of domain.com/page 3. Trimming URLs to root With any list or URLs uploaded to the harvester’s field, just click the “Trim To Root” button.
 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO


After that we should see only top domains with no subpages on our list. trim to root results Now our list is de-duplicated and trimmed to root. This format is usually used when working with disavow files (for example to see the percentage of disavowed domains vs. “alive” domains linking to our website). 


4. Exporting the lists If you would like to export this (or any previous) list, you can of course do that as well. There are also some options to choose from while exporting lists. 
To export any link list, simply click “Export URL List” and choose the option that suits you. 

export urls Export as Text (.txt) 

Simply exports all the URLs to a .txt formatted file with the URLs listed one after another. 

Export as Text (.txt) and split list This option is really helpful when working on really large link lists. URLs will be exported into multiple .txt files with a selected amount of URLs in each of those files. 

Tip: 
I personally use it when I want to use the Link Research Tools – Link Juice Tool, that accepts up to 10,000 URLs. If I want to analyze a larger list, I split it into the chunks that I then paste into the Link Juice Tool. 
Export as Text (.txt) and randomize list 
This option is pretty self-explanatory. All the exported URLs will be randomly sorted in the TXT file. 

Export as Unicode/UTF-8 Text (.txt) I have never used this option, but you can simply change the URL’s formatting to Unicode or UTF-8. 

Export as HTML (.html) This is an interesting option, used for indexing backlinks in the past. It exports all the URLs to the HTML list of links (example below).

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO


Export as Excel (.xlsx) Exports all the URLs to Excel format (.xlsx) (example below).

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO
 
Export as RSS XML List Creates an RSS feed from the link list and exports it in XML format (example below). 
export rss xml 
Export as Sitemap XML list 
Creates a sitemap-format XML file from the link list. Really useful when used with the Sitemap Scraper addon. 

Add to existing list 
Simply adds the URLs on the list to existing the TXT file. 

5. Checking for dead links/domains 
After we’ve gathered all the backlinks, we need to find the alive ones and filter out the dead backlinks. We can do this with the ScrapeBox Alive Check addon. 

Instructions: 
Open ScrapeBox 
Go to Addons and click ScrapeBox Alive Check 

ScrapeBox alive check Load the URLs that you would like to check 

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

Now, all you need to do is click Start.

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

After the check, we should see a window similar to the one above. You will find the stats of the check in the bottom of the window. Now all you have to do is save your alive and dead links.
Click Save/Transfer and pick the option that suits you.

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

6. Remove duplicate URLs or domains in a huge list (up to 160 millions) This is a feature that I wish I’s known about while working on my Expedia.com case study. With it, we can merge, de-duplicate, randomize or split huge lists (up to 180 million links). To open this ScrapeBox Addon, we need to go to Addons and then click ScrapeBox DupRemove.
 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

Now you can see a new window with the tool’s overview

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

Using the tool is really intuitive. All you got to do is load source and target files with each part of the tool. For target files, I recommend using new and empty TXT files. 7. Scraping Google This is probably the most popular use of ScrapeBox. In 2009 it was a feature allowing you to harvest more blogs for posting comments. I personally use the ScrapeBox scraping feature for: Scraping a website’s index in Google (for on-page SEO purposes) Scraping backlinks’ footprints (SEO directories with duplicate content, footprintable link networks, footer links from templates etc.) Looking for link building opportunities A word about proxies To start scraping, we need some proxies, otherwise our IP will be blocked by Google after just a few minutes. Usually the best way to find a Google proxy is to use the built-in ScrapeBox Proxy Harvester. Using it right is quite complicated though, and I will not cover the whole process here. As an easier way for SEOs starting with ScrapeBox, I recommend going to any large SEO forum. There are always a few “public proxy threads”. As an example, you can go to one of the posts listed below and simply use the proxies listed there daily. http://www.blackhatworld.com/blackhat-seo/f103-proxy-lists/ Tip:
The average lifetime of a Google proxy is 1 to 8 hours. For more advanced scrapes you’ve got to use either a lot of private proxy IPs or simply use more advanced scraping software (feel free to contact me for info).



Google scraping workflow
1. Find and test Google proxies We’ve already got a few proxy sources (pasted above). Let’s see if we can get any working Google proxies from those lists. Go to http://www.blackhatworld.com/blackhat-seo/f103-proxy-lists/ Find at least 2000 – 3000 proxies and paste them into the Proxy tab in ScrapeBox, then click on “Manage” and start the test. We are obviously looking for a Google Proxy. This is how proxy testing looks: proxy check You will see the number of Google-passed proxies at the bottom of the screen. Google proxy will also be shown on the list in green. After the test is finished, we need to filter out the list. To do that, click Filter, and “Keep Google proxy”.
 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

Now we’ve got only Google proxies on the list. We can save them to ScrapeBox and start scraping.

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

Note:
Remember to use proxies straight away, as they will usually not be alive for more than 1-3 hours.
2. Setup the desired keywords Now that we’ve got the proxies, all we need to start scraping are our desired keywords or footprints. To show the scrape in a “real life” example, I will scrape the footprint used for Expedia’s WordPress Theme. For those of you that didn’t read Expedia.com case study, it is a WordPress theme, with footer links. Pretty easy to footprint. http://derabwesendeheld.de/ expedia footer links As you can see on the screenshot above, our footprint to scrape is “Designed by Expedia Rental Cars Team.” Copy the footprint mentioned above and paste it into ScrapeBox. scrapebox settings settings To setup your scraping, follow the screenshot above. Paste your desired footprint to the top right field. Then add as many keywords as possible (I only used 3, as this is just an example), to generate more results. Yahoo, Bing, AOL I personally don’t like using them. In my opinion, scrapes done with them are not as precise as the ones done with Google. On the other hand, I know that many of my SEO colleagues use those search engines quite successfully. I leave the choice to you. You can run some benchmarks yourself and decide for yourself. Why should we add extra keywords? Each Google search is 10 – 1000 results (depending on the setup). If we want to scrape, for example, 20,000 results, we need to use extra keywords, so our footprint will look like: “Designed by Expedia Rental Cars Team.” Cars “Designed by Expedia Rental Cars Team.” Travel “Designed by Expedia Rental Cars Team.” hotels etc. This way we can cover much more “ground” and dig much deeper. Tip:
Before scraping, Google your footprint manually. With that you can have a clear idea of what you want to accomplish, and then benchmark your results.
For the footprint we’ve got Google shows ~180 unique results.

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO


Of course, having 180 unique pages scraped is a perfect score for us, but it is not always possible. Let’s see if we will be lucky enough to get close to 180 pages. All we’ve got to do now is press “Start Harvesting”.

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

Now we can watch ScrapeBox doing what it does best. Scraping Google. harvester completed OK, the search is finished, we’ve got 226 results. This is not epic, but pretty good for only 3 keywords. After clicking OK, ScrapeBox will show us the good (with results) and bad (no results) keywords.
 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

The stats above are really helpful, as with more complex searches you can be much more effective by filtering the keywords. Unfortunately, we are not finished yet. The results we see are coming from different, unique searches, therefore they are almost always heavily duplicated. Fortunately all we need to do is click “Remove/Filter” and “Remove duplicate URLs”. remove duplicate urls Let’s see how close we are to our desired 180 results: unique domains We’ve got 145 unique results. With only 3 keywords used, this is a really great result. There are probably ~35 more pages that we missed out there, but I’m sure that we’ve got all the unique domains on the list. remove duplicate domains Now let’s see the results:.

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

With ~ 15 minutes of work, we’ve got the whole footprint scraped. 145 unique URLs with 21 domains. In my opinion, scraping is a skill that is really important to anyone dealing with link audits. There are some SEO actions that you cannot do without scraping. The Orca Technique is a great example. It is not possible to implement it fully without scraping Google. Scraping and de-duplicating is not all you can do though. Imagine that you want to see which of the domains above have already been disavowed. We can do that just by a really simple filtering. 8. Filtering the results This is my favorite part. This is something that is really complicated (at least for me) to do with e.g. Excel. ScrapeBox couldn’t make it any easier. Let’s go back to the example of one of my customers – www.extremetacticaldynamics.com. This situation was also described in my “Squeeze more juice out of Google Webmaster Tools” case study. Working on their backlinks is quite complex, as the disavow file is huge. Of course Link Detox is doing all the work for us, but I want to show you how you can filter the disavowed links out of the link list with ScrapeBox. First, we need to load a large link list to ScrapeBox. A report from Google Webmaster Tools will be a good example here. All we have to do to start is copy the backlinks from the Google Webmaster Tools CSV export to ScrapeBox. GWT to notepad After importing the URLs, let’s remove duplicates (yes – there are duplicate URLs in GWT exports).

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

As you can see, there were 240 duplicated URLs. Now what I would like to do is to filter out all the disavowed URLs. To do that, we are going to use the “Remove/Filter” options.

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

As you can see, when we utilize the options listed above wisely, we can filter out almost everything. To filter out all the disavowed links, we are going to use the “Remove URLs Containing” option. Workflow: We need the disavow file. It is best to download it from Google Webmaster Tools using this link: https://www.google.com/webmasters/tools/disavow-links-main Copy the CSV content to TXT (Notepad or Notepad ++) file (to get out of Excel as quickly as possible) ☺ Use the “Replace all” option in Notepad to replace “domain:” with nothing. This way we get only a list of single disavowed URLs and domains. Save the TXT file to your hard drive as e.g. “Disavow-extremetacticaldynamics.txt” Go to “Remove/Filter” and click “Remove URLs Containing entries from…” and select your saved disavow file (without “domain:”).

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

Now we can see the filtered results. What we can see here is ~11k links removed from the list.

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

We are done. Our list is filtered by our disavow file. I think that the example above should give you an idea about the possibilities of processing your URL lists with ScrapeBox. This is a basic tutorial, but with just a few hours of playing with ScrapeBox you can become an advanced user and forget about Excel. 9. Scraping Google Suggest This is one of my favorite uses of ScrapeBox. I think that Google Suggest holds a lot of interesting keywords that we can use for scraping, post ideas or keyword research. It is also one of the easiest features to use. Workflow: Open ScrapeBox Click Scrape

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

Enter your source keywords

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

In this example, let’s use Link Research Tools as our source keyword. Usually we should use many more than just 1, but this is only to show you how this tool works. Select your Keyword Scraper Sources

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO

As you see on the screenshot above, you can also choose Bing, YouTube and many other engines. In this example, I will only use Google Suggest, though. Click Scrape

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO
And we’re done. You can see the Google Suggest scrape on the right. To get more results, you obviously need more keywords, or you can play with Level (1-4) of scrape. Tip:
I often use it to monitor all the brand-related searches. You would be surprised how much info you can get with just 1 brand keyword (example below).

 Cara menggunakan ScrapeBox untuk link audit, link building, dan on-page SEO
Just with typing in 1 keyword – “Expedia” and setting the level to 3, I got 321 brand related keywords searched in Google. Conclusion Link Research Tools is making a lot of the link audit workflow quite easy already. You would be surprised how many processes described here are actually happening in the background the moment you start any report in Link Research Tools. Unfortunately, there are unusual cases, when you need to do a lot of manual work. ScrapeBox is a good tool to organize and speed up your manual searches and link processing.

3 komentar:

  1. Looking To Sell Your Car For Cash or removal Service Is Free and We Remove Truck Van 4wds SUVs. Scrap Or Old Running Or Non Runners Call Us. 0429 794 488
    used car buyer brisbane

    ReplyDelete

Design by: Templaztezy | Copyright hapydownload