A site check-up allows one to find opportunities that might have otherwise been missed and shines a spotlight on where one should focus their online efforts for the next year.
Austin, TX (PRWEB) December 16, 2014
Here Shweiki teams up once again with expert Alicia Lawrence of WebpageFX to present a must-watch webinar on performing an SEO and inbound marketing audit to determine if one has a healthy site.
Internet marketing has been listed as one of the top 8 fun careers, so although this webinar focuses on how to do an SEO and inbound marketing audit, one shouldn't think of it as a monotonous task, but as an annual checkup to see how healthy one's site is.
In previous webinars, there's been a lot of talk of content marketing and other various SEO tactics to help boost rankings and conversions, but these tactics won’t be very effective if the site has a few obvious flaws. A site check-up allows one to find opportunities that might have otherwise been missed and shines a spotlight on where one should focus their online efforts for the next year.
Tools To Use
As in a real-life check-up, the first step towards a cure is to go to the doctors--which in this analogy, are tools like Screaming Frog, SEMrush, Google Webmaster Tools, and Search Metrics--to see how well the site has been performing:
-Screaming Frog: This is where one can scan their website (errors, meta data, rel canonicals) to fix any page errors, most of which will be things like open tags, stray elements and attribute errors.
-SEMrush: This enables one to get an overview of site traffic (whether it's steady or seasonal) and discover how it ranks compared to competitors
-Google Webmaster Tools: This should be set up (using the gmail connected to one's site) to track impressions and technical data.
Things To Look At
-Traffic: One should check to see if there are sudden drops, a steady incline, etc. Webmaster Tools gives one the ability to look at impressions one's site would have had in the search index--not necessarily ones that made it onto the actual site, just those that could have.
-Penalties: One would find this under "Manual Action" in Google and will need to figure out on which page, link or keyword the penalty is on. (A description of various penalties and how one can get their site back in Google's good graces will be found in a future webinar.)
-Site Architecture: One should try to keep all pages to a single subdomain as it will result in higher rankings for the pages. A single subdomain like a blog is fine, though the subject is up for debate whether subdomains pass value to the rest of the site.
-Navigation: Sites should be less than three layers deep to make them easy to navigate and to make articles accessible to searchers.
-Keywords and Links: One can try this tool from WebpageFX or utilize the "Links to Your Site" section of Webmaster tools. Google doesn't like site-wide links.
-Anchor Text: One's anchor text should be broken up in the following manner:
70% branded, URL, Brand+keyword, white noise (non-targeted anchor text)
20% partial, phrase and broad match
10% exact anchor text match
-No Follows Links: These should be considered for links under DA 30 or any links to off-topic sites (which one really shouldn’t be linking to at all).
-Where Links Are Coming From: This needs to be well-rounded.
-URLs: One should avoid URLs that contain certain characters (?, &, $, =, +) and should be static, less than 100 characters and user-friendly.
-Robots.txt: This restricts access to pages on a site, but one should make sure they aren’t inadvertently blocking pages that they want Google to index. One should upload a file via FileZilla in a folder to tell Google which pages to crawl in that section. Pages to add a robots.txt file would be like “wp-content/plugins” since many plugins have links to the creator’s site.
-NOINDEX tags: NOINDEX is another way to tell Google not to index a page in the search engines. One will want to use a NOINDEX, FOLLOW tag on pages like wp-includes on a page-by-page basis (such as for /tag, /category or /author to make sure they aren’t being crawled when duplicate. By allowing them to be FOLLOW, Google will still allow links to pass. One should also keep in mind that when NOINDEXing pages, it's important to want to remove those from one's sitemap as well.
Then one should use the Moz SEO Toolbar or look at the source code to see if the pages use canonical tags.
-URL Redirects: In almost all situations, a 301 Redirect should be used. One can, however, use a 302 redirect for temporary pages because of a promotional campaign or something of the sort. Any page for which the URL is changed after publishing or is deleted should be redirected to the correct or similar page.
-Duplicate Content: It's crucial to ensure that there is no duplicate content. If one is using https and http and the same page appears on both, that’s duplicate content.
Shweiki Media's mission has always been to help publishers improve by providing the most profitable, hassle-free printing experience possible. This includes guaranteeing the highest quality product, exceptional customer service, world-class communication, an on-time guarantee, and no surprises--whether printing magazines, postcards, flyers or anything else.
As a printer and publisher, Shweiki Media also believes that this hassle-free experience includes making their clients better. Utilizing relationships with industry experts, Shweiki Media strives to educate clients and help them thrive in the exciting world of publishing--while having lots of fun along the way!
For more great info from and about Shweiki Media, please check out our blog at shweiki.com/blog (and sign up for our FREE weekly expert webinars!), and subscribe to our Youtube Channel at youtube.com/shweikimedia. You can also follow us @ShweikiMedia and "like" us at Facebook.com/shweikimedia.