Spider-driven search engines such as Google®, Yahoo!® and MSN® use "robots" or "crawlers" to score websites across the Internet. Robots "spider/crawl" each site and "score" pages based on how relevant they are. A website's score or placement within a spider driven search engine is derived from hundreds of variables such as link popularity, density and frequency of keywords in page content, HTML code, site themes and more. You will want to focus many criteria in your SEO strategy to position yourself well among the major search engines. Here are two of the most influential factors:
Also make sure that your blog posts are consistent with one another and that each post has the same-sized images, headings and font. Always ensure that your blog post titles don’t lead your visitors astray.  This may seem obvious, but it happens more often than you’d think. For example, if your blog post is titled “The Top 10 Places to Hike in Southern California” but the post itself talks about hiking spots all throughout the entire state of California, you’re probably going to lose visitors. After all, it’s not what they had signed on for!
Firstly, a disclaimer – don’t spam Reddit and other similar sites hoping to “hit the jackpot” of referral traffic, because it’s not going to happen. Members of communities like Reddit are extraordinarily savvy to spam disguised as legitimate links, but every now and again, it doesn’t hurt to submit links that these audiences will find genuinely useful. Choose a relevant subreddit, submit your content, then watch the traffic pour in.
Thanks Brain, these tips are useful. The key thing with most of the tips that you provided is that it will take time and most people want to have more traffic, but they do not want to do the work and put in the time. However, if you put in the word and you do a quality job then it will work out. I think that is the overall strategies that a lot of SEOs have to do today is just to take the time and figure out quality strategies.
Do not be fooled by those traffic sellers promising thousands of hits an hour. What they really do is load up your URL in a program, along with a list of proxies. Then they run the program for a few hours. It looks like someone is on your site because your logs show visitors from thousands of different IPs. What happens in reality is your website is just pinged by the proxy, no one really sees your site. It is a waste of money.

When Googlebot crawls a page, it should see the page the same way an average user does15. For optimal rendering and indexing, always allow Googlebot access to the JavaScript, CSS, and image files used by your website. If your site's robots.txt file disallows crawling of these assets, it directly harms how well our algorithms render and index your content. This can result in suboptimal rankings.
This one is so obvious, we’re going to look at it first. Paid search, social media advertising and display advertising (try our Smart Ads Creator!) are all excellent ways of attracting visitors, building your brand and getting your site in front of people. Adjust your paid strategies to suit your goals – do you just want more traffic, or are you looking to increase conversions, too? Each paid channel has its pros and cons, so think carefully about your objectives before you reach for your credit card.
If you're looking to upload an image to a blog post, for example, examine the file for its file size first. If it's anywhere in megabyte (MB) territory, even just 1 MB, it's a good idea to use an image compression tool to reduce the file size before uploading it to your blog. Sites like TinyPNG make it easy to compress images in bulk, while Google's very own Squoosh has been known to shrink image file sizes to microscopic levels.