Wednesday, October 14, 2015

Deprecating our AJAX crawling scheme

Deprecating our AJAX crawling scheme

Deprecating our AJAX crawling scheme
Wednesday, October 14, 2015
Deprecating our AJAX crawling scheme - Hallo TRENDING NEWS, In the article that you are reading this time with the title Deprecating our AJAX crawling scheme, we have prepared this article well for you to read and take the information in it. hopefully the contents of the post Article advanced, Article crawling and indexing, what we write you can understand. all right, have a nice reading.

Title : Deprecating our AJAX crawling scheme
link : Deprecating our AJAX crawling scheme

Related Post


Deprecating our AJAX crawling scheme

tl;dr: We are no longer recommending the AJAX crawling proposal we made back in 2009.

In 2009, we made a proposal to make AJAX pages crawlable. Back then, our systems were not able to render and understand pages that use JavaScript to present content to users. Because "crawlers … [were] not able to see any content … created dynamically," we proposed a set of practices that webmasters can follow in order to ensure that their AJAX-based applications are indexed by search engines.

Times have changed. Today, as long as you're not blocking Googlebot from crawling your JavaScript or CSS files, we are generally able to render and understand your web pages like modern browsers. To reflect this improvement, we recently updated our technical Webmaster Guidelines to recommend against disallowing Googlebot from crawling your site's CSS or JS files.

Since the assumptions for our 2009 proposal are no longer valid, we recommend following the principles of progressive enhancement. For example, you can use the History API pushState() to ensure accessibility for a wider range of browsers (and our systems).

Questions and answers

Q: My site currently follows your recommendation and supports _escaped_fragment_. Would my site stop getting indexed now that you've deprecated your recommendation?
A: No, the site would still be indexed. In general, however, we recommend you implement industry best practices when you're making the next update for your site. Instead of the _escaped_fragment_ URLs, we'll generally crawl, render, and index the #! URLs.

Q: Is moving away from the AJAX crawling proposal to industry best practices considered a site move? Do I need to implement redirects?
A: If your current setup is working fine, you should not have to immediately change anything. If you're building a new site or restructuring an already existing site, simply avoid introducing _escaped_fragment_ urls. .

Q: I use a JavaScript framework and my webserver serves a pre-rendered page. Is that still ok?
A: In general, websites shouldn't pre-render pages only for Google -- we expect that you might pre-render pages for performance benefits for users and that you would follow progressive enhancement guidelines. If you pre-render pages, make sure that the content served to Googlebot matches the user's experience, both how it looks and how it interacts. Serving Googlebot different content than a normal user would see is considered cloaking, and would be against our Webmaster Guidelines.

If you have any questions, feel free to post them here, or in the webmaster help forum.




Thus the article Deprecating our AJAX crawling scheme

That's the article Deprecating our AJAX crawling scheme this time, hopefully it can be of benefit to all of you. well, see you in another article post.

You are now reading the article Deprecating our AJAX crawling scheme with the link address https://poccoll.blogspot.com/2015/10/deprecating-our-ajax-crawling-scheme.html
Deprecating our AJAX crawling scheme
4/ 5
Oleh

No comments: