Will Googlebot click a button/link to reveal AJAX content?

my client has to inform it’s customers about some new regulations that the Googlebot should NOT crawl. It is not possible to place this information on a separate page and disallow Google to crawl it. So the idea is to place a button/link on the page, that will AJAX-load the corresponding information only when the user clicks it. My assumption is, that Google is unable to click the link and crawl that specific AJAX content.

Am I right? And if yes, is there an official documentation that proofs my point on this?

How would Google treat a GitHub-Pages site, that is based on the same content as the GitHub repository itself?

I’ve just created a GitHub Pages "site", out of one of my repositories on Github. Basically it’s just a landing page, that pretty-prints my readme.md file(that is found in the repo). GitHub allows you to create such a site automatically.

I don’t know too much about SEO, but i’ve heard that google takes into account "duplicate" content, as a "devaluation" factor.

So i was wondering: If google finds a Github repo page, that outputs some readme.md file, and then finds a Github.io(GitHub Pages) page, that has almost the identical content- would it treat it as some duplicate, and apply some kind of penalty to it?

Content available vary from site to site

I thought that only the SRD content was supposed to be offered for free but on some website I can found content from the basic rules document or even the xanathar’s guide to everything book.

I usually classify sites that offer such content as piracy but there is some sites that I think to be legitimate. Here is two sites for instance :

  • dndbeyond with basic rules content on their character builder (subraces, bakgrounds…) or their data (wood elf from the basic rules : https://www.dndbeyond.com/races/elf#WoodElf)
  • aidedd.org with spells/feats/invocations from XGTE but not magic items nor monsters (eg : https://www.aidedd.org/dnd-filters/feats.php)

Do you know why we can found such content on theses website ? should I consider theses website legitimate ? How can I know ?

Sitemap: Should I dynamically update sitemap for dynamic content or create a page containing all the dynamic links

Say i have the following route http://<my-domain>/{category}/subjects/{id}/Sitemap: Should I dynamically update sitemap for dynamic content or create a page containing all the dynamic links

the ones in the brackets are dynamic, I’m struggling with what is better or any better way to let google crawl through all these dynamic links

Approach 1: manually doing the job by removing or adding the record to site map and updating <lastmod>

Approach 2: create a page that includes all those links and reference that page in sitemap.xml

The second approach can be generated as a plain html file which is generated from the server app. Or, a simple webform aspx page that dynamically generates those links without having to create an html file.

Show/hide page content based on cookie from url parameter

I am looking to show some content on a page based on the on the parameter in a link.

If a link is given to a user https://www.examplesite.com/example-page?feeback=1 then they will see the content of the page, if not using the link, then users will not see the content.

Additionally, I need the users of the link to be able to look on other pages and return to the page where the content is hidden/shown and still see the content.

I have set a cookie in functions.php, that will expire in 30days

add_action('init', 'set_feedback_cookie');   function set_feedback_cookie () {     if (isset($  _GET['feedback'])) {     $  name = 'client_feedback';     $  value=$  _GET['feedback'];     setcookie($  name, $  value, time()+60*60*24*30, "/example-page/", "examplesite.com" , "true" ); } 

The cookie is loaded on to the page (as seen in chrome dev tools),

Now i just need to hide the content on the page. The section to hide has an id #form__feedback.

What code do i need to add to get the content to hide, is this code added to functions.php or to a page template?

Is php or jQuery the best option for doing this?

Does Google penalise overwriting of content at render time with JS?

I have found out from Google’s Javascript SEO video, that Goodle indexes webpages in 2 waves and the second wave which involves content generated with Javascript is slower and less frequent.

I have an idea to tackle this issue:

  1. I will call/incule the API (I will get the Parameters by changing it to pretty URL) in the current page’s backend and print the data in plain text inside a div. This will make google crawlers get its hands on the content in the first wave itself.
  2. Then once JS is loaded, I will call that particular API again and this time I will replace the previously created div and render all the data recieved from the API once again.

My questions:

A. Is it SEO friendly? Or Google will penalise this practice? B. Will it solve the late indexing issue or am I being an idiot? C. Are there more pros than cons or vice versa?

Duplicate content, canonical url and json-ld Article

If I have a duplicate content, which URLs should I use in json-ld Article schema? To be precise, let’s say I want to publish an article from my blog.example.xyz on another site e.g. popular.example.com. When I publish the article on popular.example.com I go and update canonical URL on my blog.example.xyz, and og:url property (both should be same as implied here):

<!-- this is on blog.example.xyz --> <link rel="canonical" href="https://popular.example.com/my-duplicate-article/"> <meta property="og:url" content="https://popular.example.com/my-duplicate-article/" /> 

I also use http://schema.org/Article to describe the Article entity, however, usually all URLs within it point to my own blog.example.xyz

<!-- this is on blog.example.xyz --> <script type="application/ld+json">   {     "@context": "http://schema.org",     "@type": "Article",     "author": {       "name" : "Gitnik"     },     "headline": "My Duplicate Article",     "description" : "Uhm, how do I deal with Article entity in case of duplicate content?",     "inLanguage" : "en",     "wordCount":  600 ,     "datePublished" : "2020-05-18T21:00:00",     "dateModified" : "2020-05-18T21:00:00",     "image" : "http:\/\/blog.example.xyz\/img\/logo.png",     "mainEntityOfPage" : "http:\/\/blog.example.xyz\/2020\/05\/my-duplicate-article\/",     "publisher" : {       "@type": "Organization",       "name" : "http:\/\/blog.example.xyz\/",       "logo" : {           "@type" : "ImageObject",           "url" : "http:\/\/blog.example.xyz\/img\/logo.png",           "height" :  60 ,           "width" :  60       }     }   }   </script> 

but Google states that mainEntityOfPage should be same as canonical URL. Now, it’s easy to update mainEntityOfPage to the new url. But what about image url and organization? Can image url point to a location other than canonical url? And should I change Organization entity to the popular.example.com and change the url there also? How would it affect SEO if I left them unchanged? How would it affect SEO if I left mainEntityOfPage unchanged?

Does the re-use of HTML email newsletter content for blog posts have SEO (or other) repercusions?

I am (in parallel) building both a newsletter subscriber list (using Mailchimp) and also a blog website (using WordPress). I want to be as efficient as possible with the use of my content / copy. I would like to send out newsletters to subscribers first and then re-use the content for time delayed blog postings. Using a WordPress plugin such as ‘postie’ I can easily send an email directly to my website that is then entered into my blog post stream automatically. If I prepare the styling of my email appropriately in Mailchimp, the posting should look like native content on my website.

However, my current understanding is that robust HTML emails tend to use A LOT of tables to force the email client to render the email correctly. Also, they tend to use A LOT of inline CSS and the like. If I simply insert this into my blog stream the website HTML is going to get ‘ugly’ and likely slow to load.

But, does this actually matter? Will webcrawlers & search engines have a tantrum?

If I need to avoid this scenario, should I search for a solution (hopefully not manual re-coding) that creates two versions of the HTML: one for newsletter use; one for blog stream insertion?

I really want to avoid the manual effort of having to create two versions of the HTML content from scratch each time.

Many thanks in advance for any thoughts, advice and suggestions you can offer.

[ If it is of any interest / relevance, you can visit my fledgling blog site at https://charlesgull.mobi ]

GamesBOB.com Old games site since 2005, perfect domain + content selling for $2k

Why are you selling this site?
Had it for a long time and I've stopped maintening it

How is it monetized?
Google Adsense, last payment was £73 in June 2019

Does this site come with any social media accounts?
No

How much time does this site take to run?
Left abandoned but has potential to be a great blog for a gamer or tip/cheats for games

What challenges are there with running this site?
I made it when I was 18, now I'm old at 34