Regarding RankBain, my own assumption is that user signals are part of the training data RankBrain gets (even though Paul Haahr does not confirm that in the talk at SMX or the discussion afterwards). If that is true, then RankBrain will see your high CTR and maybe TOS, might try to figure out what pattern causes them and MIGHT try to change it’s own algorithm in a way that ranks results LIKE YOURS higher.
SEO.com is a certified Google Partner, which means the search engine has certified our knowledge and experience of Google applications and tools. In addition, we have access to specialized Google training and products, and a dedicated Google representative. Whether you are in ecommerce, technology, retail, dentistry, or B2B, you will benefit working with SEO.com.
Damn, i recently just made a powerpage about linkbuilding, and ranked top 3 with my content in SERP real quick here in Denmark, but i could see that most people weren’t ready for the content, so i had a bounce rate on nearly 70% (sad face) I think it was both that my above the fold content is very borring, compared to a few scrolls down the page where the infographics starts. I might just put a video in my above the fold content now for higher dwell time.
Engagement – Google is increasingly weighting engagement and user experience metrics more heavily. You can impact this by making sure your content answers the questions searchers are asking so that they’re likely to stay on your page and engage with your content. Make sure your pages load quickly and don’t have design elements (such as overly aggressive ads above the content) that would be likely to turn searchers off and send them away.
Robots.txt is not an appropriate or effective way of blocking sensitive or confidential material. It only instructs well-behaved crawlers that the pages are not for them, but it does not prevent your server from delivering those pages to a browser that requests them. One reason is that search engines could still reference the URLs you block (showing just the URL, no title or snippet) if there happen to be links to those URLs somewhere on the Internet (like referrer logs). Also, non-compliant or rogue search engines that don't acknowledge the Robots Exclusion Standard could disobey the instructions of your robots.txt. Finally, a curious user could examine the directories or subdirectories in your robots.txt file and guess the URL of the content that you don't want seen.
Brian, I recently found your blog by following OKDork.com. Just want to say you’re really amazing with the content you put out here. It’s so helpful, especially for someone like me who is just starting out. I’m currently writing posts for a blog I plan to launch later this year. I think my niche is a little too broad and I have to figure out how to narrow it down. I essentially want to write about my current journey of overcoming my fears to start accomplishing the dreams i have for blogging, business, and travel. In doing so, I will share the best tips, tools, and tactics I can find, as well as what worked, what didn’t and why.