Many are not even aware about SEO. While some beginners struggle with it. The question is - From where to start the process of search engine optimization? Today, I'll let you know some basics of the initial SEO work that you can upload on your new website in order to get the high ranking's on search engines result pages. Following is a list to get started!
1. Domain Authority
>>You can use bulk da checker(http://www.bulkdachecker.com/) to check the current domain authority score of your website. It may increase or decrease with time.
2. Page Authority
>>Just like you checked the domain authority you can check the page authority. The difference between them is of an entire domain and a web-page.
3. Back-links Profile
>>Back-links have always been a constant and important factor according to Google guidelines. You can analyse your back-links with the help of tools like Ahrefs, Google search console etc.
4. Link Saturation
>>Didn’t understood? Don’t worry! This term refers to the numbers of web-pages of your website that is indexed by search engines.
5. Domain Age / Domain Creation
>>Search engines provide first preference to old domains which are registered for a long tenure. So domain age also matters up to some extent to get the higher rankings on search engines.
6. Domain Expiry
>>We can register domains up to 10 years. So it is always expected that domain that you book is registered for higher number of years.
7. Google Analytic
>>Google has provided a great tool to analyse the traffic and various ways/sources of referral. You can install the provided code in <head> section of all your web-pages.
8. Hidden Text / Links
>>Having a hidden text or hidden links on the website is not recommended and is considered as black hat technique. It is always on safe side to make sure that no hidden texts or links are present on website or any of the web-pages.
10. Website Load Time
>>Also known as page speed time. It is recommended that website or web-page should render in no time or take minimum time to load on browsers.
11. Mirror Sites
>>Website with unique appeal and content attracts more visitor.
12. Check Server Status/htaccess
13. Similar Sites on Current Domain
14. Website Content
>>Check all the web-pages content including home page and other service pages.
15. Duplicate Content/Plagiarism
>>Having unique/fresh content always helps and it is a basic factor to rank on search engine.
16. Custom 404 Error Page
>>This helps in helping a user to navigate in a better way to the website. If user does not find anything then with can denote that this page does not exist. This custom page hold the status code 404.
17. 301 Redirect
>>301 redirect is a permanent redirection of URLs.
18. Google Webmaster Tools/Search Console Verification
>>Google webmaster or Google search console is one of the best tools made and provided by Google to numerous webmasters.
19. Xml Sitemap
>>It is a website map in XML(extensible markup language) file format which is installed in root directory of a web-server for search engine crawlers.
20. Simple/HTML Sitemap
>>It is a website map in HTML(hyper text markup language) file format which is installed in root directory of a web-server.
>>Webmasters use to avoid duplication between two pages of website. i.e. home page and index page
22. Social Widgets
>>Social widgets are always recommended to enhance the user experience. We can install widgets like Google+1 button, Facebook like button and posts, Twitter tweets etc. Which also helps us to gain new likes and sometimes traffic also!
23. Social Profile Icons
>>We should always link our social profiles with our website which helps search engines to find the correct profile of your brand.
24. Social Profile Schema
>>By doing social profile schema, we can inform Google about our social profiles. This also helps in branding. The search for this shows in knowledge graph panel of result pages.
25. Schema Tag
>>Google uses microdata format which involves schema markup for reviews/testimonials, addresses etc.
26. Robots.txt File
>> Robots.txt is an instructive file which instrcuts the search engine robots to crawl or not to crawl the website.
27. Noindex – Nofollow Tag
>>This comes as similar concept to robots.txt file. The only difference between them is of file and tag. And both serves for the same indexing and following purpose by crawlers. Below is the tag which can be included in <head> section of a home page.
<meta name=”robots” content=”noindex, nofollow”>
28. Broken Link Checker/Dead Links
>>The links which direct us to pages or links which does not exist is a broken link. We need to rectify them as soon as possible because it hurts the structure of the website.
29. Integration of internal Blog
>>internal blog helps to increase the website volume and it also drives the traffic to site through fresh content.
30 External/Outgoing Links
31. Dynamic Pages
>>It is always beneficial to have static web-pages keeping search friendly aspect. Dynamic pages are not actual pages which exist on a web-server. So it is always recommended to have a static web-pages.
>>Robots can not crawl frames/iFrames located on a web-pages, so it is recommended to have them externalized or removed if possible.
34. Contact Form and Tracking
>>To analyse the web traffic that came through contact form submissions, we need to implement form tracking in Google Analytics.
35. Duplicate Meta Tags
>>Unique meta tags that are concise always helps to increase CTR. So it is good that we have a unique web-page based meta tags that clearly explain the web-page intent.
That’s all for now ;) You can also use this check-list to ensure that all the basics of you website are covered.