Redirecting is the act of sending a user to a different URL than the one initially requested. There are many good reasons to redirect from one URL to another, for example, when a website moves to a new address. However, some redirects are designed to deceive search engines and users. These are a very poor user experience, and users may feel tricked or confused. We will call these “sneaky redirects.” Sneaky redirects are deceptive and should be rated Lowest.
It is important you spread all that real ‘PageRank’ – or link equity – to your sales keyword / phrase rich sales pages, and as much remains to the rest of the site pages, so Google does not ‘demote’ pages into oblivion –  or ‘supplemental results’ as we old timers knew them back in the day. Again – this is slightly old school – but it gets me by, even today.
Your site’s URL structure can be important both from a tracking perspective (you can more easily segment data in reports using a segmented, logical URL structure), and a shareability standpoint (shorter, descriptive URLs are easier to copy and paste and tend to get mistakenly cut off less frequently). Again: don’t work to cram in as many keywords as possible; create a short, descriptive URL.
Optimizing the page for keywords is quite simple if we follow the basics mentioned in this article. As said above, use keywords in URL, meta description, Title and headings. I think it doesn’t matter how many times we use keyword in a page, but where we use keywords matters. In the blog content we can use LSI keywords which are probably the best keyword ranking approach in 2018.
When optimising a title, you are looking to rank for as many terms as possible, without keyword stuffing your title. Often, the best bet is to optimise for a particular phrase (or phrases) – and take a more long-tail approach. Note that too many page titles and not enough actual page text per page could lead to doorway page type situations. A highly relevant unique page title is no longer enough to float a page with thin content. Google cares WAY too much about the page text content these days to let a good title hold up a thin page on most sites.
To avoid undesirable content in the search indexes, webmasters can instruct spiders not to crawl certain files or directories through the standard robots.txt file in the root directory of the domain. Additionally, a page can be explicitly excluded from a search engine's database by using a meta tag specific to robots (usually ). When a search engine visits a site, the robots.txt located in the root directory is the first file crawled. The robots.txt file is then parsed and will instruct the robot as to which pages are not to be crawled. As a search engine crawler may keep a cached copy of this file, it may on occasion crawl pages a webmaster does not wish crawled. Pages typically prevented from being crawled include login specific pages such as shopping carts and user-specific content such as search results from internal searches. In March 2007, Google warned webmasters that they should prevent indexing of internal search results because those pages are considered search spam.[47]
The more important concept to keep in mind is that you want to choose keywords that best relate to the content present on a web page and on a website; if you don't have a dog food comparison matrix, then don't bother including comparison-related keywords; you are misleading your users, and certainly not fooling Google. So in an ideal world, you do have a comparison section, a reviews section, and a rankings section, housed on different pages or sections of your site, with each one tagged with the appropriate keywords. Correspondingly, your SEO and PPC search engine marketing efforts should that content by driving review keywords to the review pages and so on.
At the moment, I don’t know you, your business, your website, your resources, your competition or your product. Even with all that knowledge, calculating ROI is extremely difficult because ultimately Google decides on who ranks where in its results – sometimes that’s ranking better sites, and sometimes (often) it is ranking sites breaking the rules above yours.

An SEO meta description is a brief description of content found on a web page. The meta description is shown to users in search engine results pages to help them decide which search result to click on. Meta descriptions are not a ranking factor in the eyes of a search engine, but they can influence how many people click on a result -- which directly affects the result's organic ranking.

Writing blog posts is especially effective for providing different opportunities to land on page one of search engines -- for instance, maybe your eyeglass store’s website is on page three of Google for “eyeglasses,” but your “Best Sunglasses of 2018” blog post is on page one, pulling in an impressive amount of traffic (over time, that blog post could also boost your overall website to page one).
Baseline ranking assessment. You need to understand where you are now in order to accurately assess your future rankings. Keep a simple Excel sheet to start the process. Check weekly to begin. As you get more comfortable, check every 30 to 45 days. You should see improvements in website traffic, a key indicator of progress for your keywords. Some optimizers will say that rankings are dead. Yes, traffic and conversions are more important, but we use rankings as an indicator.