By relying so much on factors such as keyword density which were exclusively within a webmaster's control, early search engines suffered from abuse and ranking manipulation. To provide better results to their users, search engines had to adapt to ensure their results pages showed the most relevant search results, rather than unrelated pages stuffed with numerous keywords by unscrupulous webmasters. This meant moving away from heavy reliance on term density to a more holistic process for scoring semantic signals. Since the success and popularity of a search engine is determined by its ability to produce the most relevant results to any given search, poor quality or irrelevant search results could lead users to find other search sources. Search engines responded by developing more complex ranking algorithms, taking into account additional factors that were more difficult for webmasters to manipulate. In 2005, an annual conference, AIRWeb, Adversarial Information Retrieval on the Web was created to bring together practitioners and researchers concerned with search engine optimization and related topics.
Organic content marketing is a potential boon for your business, as it involves providing the material that online users are looking for. But it’s a slow process that involves a long-term commitment to producing high-quality material while constantly interacting with customers through various online platforms. It isn’t enough to post a few articles a month. If you want to benefit from potentially over three times as much traffic, you need to publish at least 16 articles a month. If you handle your campaign in the right manner, there’s a chance to see sustainable growth in your brand and a deeper level of engagement with your customers. But this resource-intensive approach may require that you outsource work to stay relevant and get the results you need.
I would also advise to continue doing what works. If something you have rolled out generates great traffic and links bring out a new version of the content, for example the 2012 version worked effectively bring out the 2013 version of the content. Another effective strategy is to make the piece of content into an evergreen article which you add to over time so it is always up to date.
Since there is an obvious barrier of entry for anyone trying to beat you once you’re established, you won’t have to worry about having competitors “buying” their way to the top. Their only option is pay per click ads, but then again, it isn’t the same as getting a higher position on the SERPs. Again, this is assuming that you took the right steps and were patient enough to solidify your place in the top search results.
This refers to your ability to rank for certain keywords. For instance, say you sell shoes online, you will need to optimize your site, backlinks, website speed and much more so that you can “rank” high for certain keywords that are highly relevant to your business. Relevant keywords may include “buy shoes,” “shoe sale,” “where to buy shoes,” and so on. Once you can rank high (top page) for these keywords, you will enjoy an increase in traffic and business as a result.
And then on the flip side with partners, it is a little bit different. Again, you can do that education piece, but ultimately offering reciprocation is going to be your best bet. So finding out who the social media manager is for your partner, talking with them, figuring out, ‘Hey, we’re going to retweet you this many times this month, or is there any events that you have coming up that we can help promote?’ And that reciprocation is very natural. It also builds credibility for both organizations. I think there’s a lot of opportunity to engage with your partners as well as amplify their content.
Beyond organic and direct traffic, you must understand the difference between all of your traffic sources and how traffic is classified. Most web analytics platforms, like Google Analytics, utilize an algorithm and flow chart based on the referring website or parameters set within the URL that determine the source of traffic. Here is a breakdown of all sources:
You can then use the Data Cube to uncover topics relevant to prospects later in the buyer’s journey, such as your brand versus a competitor or case studies illustrating the value you provide for customers. Since your organic search research will help you understand what people are interested in reading, your emails will be more targeted and helpful, boosting your success rates.
For example, we have seen retailers who produce seasonal items, such as ornaments, based upon data that they are able to find through the Data Cube. This information tells them about the themes and ideas that are popular throughout the year and during their peak selling seasons, helping them to know which items will be the most successful and the optimal names to select.
I'm having a problem that I suspect many marketers share. Quite simply … SEO or just buy the traffic. I noticed that you switched to SEO because you like the passive income component. But when I consider ALL the work and ongoing moving parts to SEO .. visions of the hamster on a treadmill appear in place of couch potato cash. Have you noticed that there is always something new to do … now it's Google+ ect. and "more to do" is surly on it's way. It's reached the point where it's mind numbing.
For our client: We took the top PPC terms based on conversion and worked these keywords into existing pages on the website. We also created new high-quality content-based pages from these conversion terms. This type of strategy can work very well in assisting overall conversions on the website and driving more revenue. We also conducted a large-scale keyword research project for the client which yielded in uncovering many areas of opportunity for content development and targeting.
To avoid undesirable content in the search indexes, webmasters can instruct spiders not to crawl certain files or directories through the standard robots.txt file in the root directory of the domain. Additionally, a page can be explicitly excluded from a search engine's database by using a meta tag specific to robots. When a search engine visits a site, the robots.txt located in the root directory is the first file crawled. The robots.txt file is then parsed and will instruct the robot as to which pages are not to be crawled. As a search engine crawler may keep a cached copy of this file, it may on occasion crawl pages a webmaster does not wish crawled. Pages typically prevented from being crawled include login specific pages such as shopping carts and user-specific content such as search results from internal searches. In March 2007, Google warned webmasters that they should prevent indexing of internal search results because those pages are considered search spam.
Facebook ads contain links back to your business’s page. Even if the goal of your ads is to get people to click on a link that takes them off of Facebook, there’s a chance they’ll go directly to your Facebook page to learn more about you. If your page is empty or outdated, that’s where their curiosity ends. If you’re spending the time and money to advertise on Facebook, make sure you follow through with an up-to-date Facebook page.
Google claims their users click (organic) search results more often than ads, essentially rebutting the research cited above. A 2012 Google study found that 81% of ad impressions and 66% of ad clicks happen when there is no associated organic search result on the first page. Research has shown that searchers may have a bias against ads, unless the ads are relevant to the searcher's need or intent