Hi guys welcome to part 5 of my ultimate guide to tiered link building In the last video we setup tiers 2 and 3 to post forever, in this video I’m going to show you how to export and organise all of your links How to get every single you have created indexed on auto pilot and most importantly how to send the search spiders ripping through the millions of unique paths that lead to your site Now if you remember from the last video we created this link tracker spreadsheet and I created some additional tabs here What we need to do is populate these tabs with tier 2 and 3 links and in the case of link wheel bandit tier 4 links that we have created And once we’ve got that list together we can move on to getting them all indexed and things like that So first of all lets jump over into link wheel bandit so exporting the report from linkwheelbandit is really easy, just select tools, generate report and that gives us this window here and then we can select our link wheel that we want to export the report from You can select date ranges; I always go for the beginning of time to make sure we get all of the links here In options we can select which tiers we are going to export, remember this is actually tier 2, this is tier 3, this is tier 4 because we built the link wheels out to our tier 1 links So you will need to in order do tier 1 posted article urls and then save that as Tier 1 And this is going to say there’s no data for my options because I didn’t actually build this link wheel out but yours will actually export a list of URLS So you need to go through and do a report for each of the tiers as so, generate report and again yours will actually export some data here and then tier 3 and generate report here, ok So if you come to C link wheel bandit reports here you’ll get the text files of the reports you should have 3 here, tier 1, tier 2 and tier 3 So what we need to do is open this up and select it all and come to our excel spreadsheet and remember tier 1 in linkwheelbandit is actually tier 2, paste them in here and then go through and do the same for the tier 2 text file which goes on the tier 3 tab and the tier 3 text file goes on the tier 4 tab So I’m just going to paste in some dummy content here to show you how it all pieces together Just before I show you how to export the URL’s from GSA I just wanted to show you the value of Captcha Sniper since my last tutorial Last time the total solved was around 100,000 and since then it’s done another 86,000 captcha solves for me That an immense number and would cost about the same cost as Captcha Sniper to solve and that is in just a weekís time So again I highly recommend captcha sniper it is essential for posting out on mass with GSA so just make sure you pick it up it will save you a lot of money in the long run So let’s just minimise that Exporting our tier 2 links first here is the project we created last time, I didn’t let it run for very long just enough to build a few links so I can export them and show you how this works So if we just right click on the project, show urls, verified and in this window here click and press ctrl + a to select all, right click copy URL, come to our spreadsheet and paste it down into the tier 2 tab here And then we just need to come back into GSA and find our tier 3 project and the same process again, right click, show URL’s, verified, select all, right click, copy urls and then go to tier 3 of our spreadsheet and drop them in there Then at this point you want to be saving a copy of all of your exported urls Now we just need to export all of these URL’s into their own individual text file So to do that select all of the tier 1 links, come into your text editor, paste them, save tier 1 dot txt and then tier 2 tab, come up to the top here, select them all and tier 2 dot txt and then come here to tier 3 links, save as tier 3 dot txt and then here the tier

4 links And again these are just sample ones I’ve tweaked all the URL’s so none of these are actually live URLs and your list will be an awful lot longer here so copy these, come across tier 4 dot txt And now we can cross reference all of these text files against each other in Scrapebox to make sure all of the links are still alive and none of them have been deleted before we move onto the next step of getting them all indexed and boosted even further So in Scrapebox the first thing that we are going to do is make sure that all of our tier 2 links are still all live and linking to our tier 1 links So to do that we are going to use the check links function in the websites field here we are going to load our list of tier 1 urls In the blog lists here we are going to load our list of tier 2 urls and then click on check links And what that will do is this is a list of our tier 2 links here and it will check each of these and see if they are linking to any of our tier 1 links Now this tier 2 list is only 62 URL’s here but your tier 2 lists is going to be much bigger So once you’ve got that loaded in like this just click start and Iím going to pause the video and resume it when itís finished Ok that finished so we just need to go to export links, export links found and we are going to call this tier 2 found and press save Then we can close this next in this field here we need to open our tier 2 links and this time down here we are going to open tier 3 Click check links again and this time this is a list of our tier 3 links and its going to check each of these to see if they link to any of our 62 tier 2 links down here so again click start I’m going to pause the video I’ll be back when itís done Ok thatís done so again export links, export links found this oneís tier 3 found and click save and close And we just need to do it once more this time in here a list of our tier 3 links and in here a list of our tier 4 links click check links click start and Iíll resume the video when thatís done Ok thatís done you can see that it only found 7 entries here this is because I didnít actually build out the full campaign and build all of the tiers in link wheel bandit and GSA I just let it build a couple of URLS before I stopped it this is purely for demonstration purposes So again thatís why youíre not seeing all the URLs, you’ll have a much bigger list here when youíre doing this final check but anyway export links, export links found and that is tier 4 found If we come back over to our link tracker spreadsheet we can actually go through and delete everything out of tier 2, 3 and 4 And then what we need to do is open our tier 2 found, tier 3 found and tier 4 found files that we have just created and then copy those into the relevant tabs And once you’ve done that what you will now have is a nice organised spreadsheet of all your live links split into tiers This will be really handy moving forward so make sure you get everything organised like this and now we can start indexing everything up and sending all the search spiders through all of these links to eventually find our site Just before we move on to the next step I wanted to show you a bit of software called Inspyder Backlink Monitor I’ve only just started playing with it properly I donít actually use it as part of my main process yet but itís a really nice way of managing a tiered link building campaign Now the process I’ve just shown you in excel while it works it is a bit of a ball ache and it does take a bit of time to put together but with Inspyder Backlink Monitor you can just setup a project and paste all of your backlinks into this one box and thats all of your tier 1, 2, 3 and 4 and thereís no need to separate them out in anyway just one

big long list of all the links you’ve built in your campaign Once youíve done you can just hit go here and it goes out and checks various things for each page If the link is still alive, how many outbound links there are, if the links no follow, the domain page rank, the page rank, whether or not itís indexed in Google, the anchor text and the IP address And when the run is completed if you come over to this link hierarchy tab you can see here earlier on I think in video 3 I created bookmark links to Google.co.uk and the rest of the links to Yourdomain.com as examples But you can actually see here we can double click here and this is our money site URL, these are all tier 1 links this is a tier 2 link pointing to the tier 1 link and we can see another example there a tier 2 link pointing to a tier 1 link and if we just open this one up here, here again we’ve got a tier 2 linking to our tier 1 link If you’d have done the campaigns properly, oh here we go so here we’ve got a tier 2 link, tier 3 link and all the additional tiers under that If I’d have built this campaign out fully there would be a lot more here for you to see but this is a really nice way of organising your tiered link building campaign as it automatically puts the link hierarchy and the structure together so you can easily see exactly whatís going on, the PR, the links whether things are indexed or not and then as you build new links for your campaign all you need to do is paste them into this box click Go again and it will go out and update all of these and update the data for your backlinks and recreate the link hierarchy and things like that There’s also the selection tool here so you can select things based on criteria You can say I just want tier 1 links that are not indexed or you might want just tier 3 links that are not indexed and you can see this supports up to 10 tiers of links which is far beyond anything I’m ever going to build But I just wanted to show you this tool itís not something I’m using heavily right now I’ve only just started testing it properly this last week or two it might be a better solution for some people out there so I just wanted to go over that and show you how it all works So letís take a look at how we are going to index everything up It is really important that we get all of the tiered links we have created indexed in Google We also need to make sure we send the search spiders on a path of discovery throughout the tiers and as the spiders move through the tiers they are going to discover millions of unique and relevant paths to find your money site The way we are going to do that is by building more backlinks obviously, this is tiered link building and the best way to index anything is to build more links to it But thereís a clever way to go about it, we are only going to build links to pages that are already indexed in Google and get regular spider visits already This means that we donít have to worry about building links to index links and then we need to build more links to index the links we have created above that If we just build links that are already indexed by Google and get regular spider visits then it will index everything underneath it automatically, we don’t need to worry about indexing any more So there’s two ways that we can do this one is to use auto approve blog comments and we can use software like scrapebox to achieve that quite easily and the second one is guestbook submissions Personally I use xrumer to do these just because itís much faster but I appreciate that not everyoneís going to be able to afford the costs of xrumer nor the server that you’ll need to run it on so I’m going to show you how to do it with the Scrapebox learning mode poster and also GSA search engine ranker so you’ve got 2 ways to go about doing the Guestbook submissions So just before we jump to scrapebox we need to make a master list of our tier 2, 3 and 4 links So to do that if you just select all of your tier 2 links and bring them to a new tab here, all of your tier 3 links and the tier 4 links which is actually tier 3 from linkwheelbandit if you remember correctly and drag that here

And if we just rename that Master List and give that a save Next we need to go and get a big long list of related keywords so if we fire up Scrapebox And once Scrapebox is open go to scrape, keyword scraper and just enter a few of your root keywords into here and hit scrape and Iím just going to pause the video while that completes Ok thatís completed so hit ok, remove duplicates and we actually need a really really long list of keywords so take those scraped keywords and we are going to add them to the search box here and hit scrape again and that will find even more keywords for us so I’m just going to pause the video Ok that’s done and then if we just remove duplicate keywords, transfer results to main keyword list, close and then we’ll just save that as keywords Ok and what I should have done before with the master list from excel is if you just take a copy of the master list we created and save that as a text file as well and that can be called websites, save So then we need to generate a list of names and e-mail addresses and we can do that up here and I’m just going to generate a couple thousand names and we are not actually going to use the names for posting we are going to use the keywords we just saved for posting so we can skip the names files and just come across here and generate some e-mails and save those Ok So close that and then we just need to setup the actual comment posting project Make sure your using proxies this is important in this example I’m not going to both but you should be using proxies here and setting the project up for names we are actually going to select our keywords file E-mail our E-mails file Websites our websites file which is a master list of our tier 2, 3 and 4 links, Ok Your comments file, I have a pre-prepared spun comments file, I’ll show you what that looks like quickly This is something you should prepare in advance but itís just a list of generic type comments that are suitable for any site And finally your blog lists or target lists of sites that we are going to post to and this is a huge auto approve list that I’ve created over time If you need to create your own auto approve list then have a look around some forums and things like that you can usually find some lists going about Spend some time collecting lots of different ones and you can merge them into one list As you can see here mines over 180MB in size I think thereís over 3 million entryís in it at the moment but just go around the web, collect your own list and make sure you remove any that constantly fail So open And that might take a while to load because of the size of my list so I’ll pause it while it’s having a think Ok and once thatís loaded all you need to do is press start posting and that will go out and start submitting a load of blog comments to your targets URL’s and if I just press Ok there and we can see that’s working through the list already So we can just leave this running now in the background and this will build lots of auto approve blog comment links to our tier 2, 3 and 4 properties and next we can move over to guestbook submission So with guestbook submissions I normally use xrumer but itís pretty expensive and you need a dedicated server or at least a VPS to run it really so it’s probably going to be out of most peopleís price range So we are going to look at two different ways we can post to guestbookís without a huge cost The first one is going to be with the Scrapebox learning mode, now you can install the learning

mode in the show available add-ons toolbox and once you’ve got that installed just open it up And once it’s open you can come across to the learning tab here Now you can pretty much teach Scrapebox any platform you want with the learning mode and it’s really easy to do So Iím going to teach it the Lazarus guestbook platform here and I’ve already got a list of Lazarus guestbookís as you can see here so if we go to load urls from file and select our list of Lazarus URLS and just open it up and here it will say detected forms, name = book that looks like a guestbook form to us and click select and then it’s just a case of clicking in these fields and it will come up with a list of variables to assign, so in the name field we are going to want username, Ok For e-mail field user email, Ok Location it isn’t a required field so we can skip that Homepage userurl, ok Your message, comment, ok And Captchas this is a text based question captcha so we can just click here and it says select the part before the question that will identify the captcha which it has already done for us there Select the part after the question which will identify the captcha and it’s already done that for us there as well so just click Ok And in here we have got a text captcha result and press OK, and once youíre happy with those we can move onto the next URL up here Detected the form, select and as you can see it has already filled out these fields for us based on what we filled out for the last one So letís take a look at one more and check they are all the same Oh that oneís got an image based captcha so we’ll skip that you can’t actually mix image and text based captchas in the learning mode you have to separate them out so letís try this one instead Thatís an image based one, here we go another text based captcha so select username, useremail, user URL and comment Itís already looking pretty trained to me lets double check all of this should be filled out and thatís it, so that is scrapebox we’ve taught it the Lazarus platform here you can go through and teach it lots of other different platforms but once you’ve done that you just need to save the learned form data, Lazarus posting config and hit save and let’s quickly have a look at how we post out to all of these guestbookís So submitting to the guestbookís is really easy if we just come across to poster here and the definition file is the file we have just saved the posting config open that up and then just fill out the rest of these quickly Name and E-mail, comments thatís it there and load URL list this is our list of Guestbookís we want to target which is there and that is pretty much it So we taught it the Lazarus guestbook platform, exported that config, loaded it all into the poster here and then you’ve just got to click on start posting and that will go out and submit your link to all of the guestbookís This is kind of a slow and sluggish way to achieve guestbook posting as you’ll have to go through each guestbook platform and teach it to scrapebox which is going to take you a while and it’s a bit of a ball ache managing different lists for different platforms and things like that Thereís a much easier way to do it with GSA Search Engine Ranker if you’ve got it but

if not Scrapebox, you can do it with So let me show you how to do it with GSA because it’s much easier So if you open GSA and click new, untick everything here except for guestbook we want that ticked and you can see how many guestbook platforms are supported there, thereís quite a few So letís fill out this form quickly our target URL’s which are our tier 2, 3 and 4 links and we are going to import our master text list of those urls which is here Our keywords is a comma separated list of keywords and anchor text is a spun list of our anchor text Tick these 3 boxes here it comes with some default text inserted there but I like to use my own custom written and hand spun comments so if you grab that and chuck it in there and also replace the German one and the guestbook titles which you can leave at default Come across to options the only thing that you want to untick is that so we only post guestbook links otherwise we end up getting some of the other types come through and right click check all with English language and click on Ok and this is going to ask us for a name so I’m going to call it video tutorial guestbook spam and hit ok And as you can see our project has been created there now what GSA will actually do is it will go out and scrape its own target list of guestbookís before it submits to them What we can actually do is use Scrapebox to scrape a huge list of all the different types of Guestbookís and then import that URL list into GSA so that this project only has to focus on the submission of our URL’s and itís just a really nice way to speed things up So underneath this video on my blog you will find a list of guestbookís that I have previously posted to and used you can download that and import as target list here free But what you should do is use Scrapebox to scrape your own targets, now I’ve included a list of footprints as well under this video that you can use for scraping so letís just head over to Scrapebox and Iíll show you how to use those very quickly one moment So in Scrapebox to complete the guestbook scrape you need 2 files, one is the list of footprints for all of the guestbookís that are supported in GSA this file you can get underneath this video on my blog and then we also need the merge list of common words which I made available in video 2 I believe it was So with those 2 files in place here we can import from file and thatís going to be our actual footprints list here and then if we click the merge button we can merge that with our merge list and that creates a list of things to go out and scrape for You will need to use proxies here to complete the scraping process I’m not going to just for this example I’m just going to do a very brief harvest here and just see what we can get so I’m just going to pause the video while that completes and we’ve got a couple of thousands results in now so I’m just going to stop harvesting, remove duplicates and then just export that URL list as GSA Guestbook Targets Now obviously that will take a lot longer to run as thereís nearly 10,000 keywords here and youíre going to end up with a huge list but once you’ve got a list of guestbookís exported you can head back over to GSA Once youíre in GSA to import that list of Guestbookís you can just right click on the project, import target URLs from file and select that target list and GSA will automatically parse that list and submit to them all as soon as you press start

So letís quickly sum up what we have done in this video First of all we learnt how to extract and organise all of our tiered backlinks with scrapebox and a spreadsheet Don’t forget to take a look at inspyder backlink monitor as well Then I showed you how to setup an auto approve comments blast with Scrapebox Next we looked at 2 different ways to post to guestbookís, one way was with the Scrapebox learning mode and the other was with GSA search engine ranker The end result is you now know how to index all of your tiered links and most important get the attention of the search spiders All of the comment and guestbookís links have been placed on pages already indexed in Google so now all we have to do is wait for Google to find our links and send it’s spiders through all of our tiers In the next video I’m going to reveal my personal advanced tier 1 tactics to help throw you up the rankings in no time at all Youíre also going to learn how to automate natural social signals in just a few clicks And I’m going to show you how to buy and use new or aged domains in your campaign to create the perfect tier 1 link Well that wraps up part 5 of the ultimate guide to tiered link building tutorial If you have enjoyed it and want to see more like this then please subscribe to my blog at www.matthewwoodward.co.uk See ya!