High Number of Crawl Anomaly Errors

since April 9, i am receiving a steadily increasing number of Crawl Anomaly Errors. The number of crawl anomaly pages is currently over 22000.

  • Each page in the error list can be reached by live url testing from the search console and by curl or webbrowser.
  • The live test rendering is also successful.
  • Server logs don’t even have records of Google visiting the crawl anomaly pages.
  • There are no javascript errors or 404 resources.
  • The load time of most of the crawl anomaly flagged pages is <1.5s till initial paint and <2.5s till final paint ( tested by several sites)
  • Server logs show no signs of downtime or server side errors in this time period
  • Other pages are visited by Google, but the number of indexed pages is not increasing
  • Some of the flagged pages were crawled normally by Yandex and Bing
  • i have tried clicking on check for fix of the errors, but the check has failed multiple times
  • When i do the live test on a single crawl anomaly url, they become marked as crawled, but not indexed
  • Pages have some Rich Data warnings, but no errors
  • There are a couple of soft 404 errors, about 122
  • the number of indexed pages is steadily declining
  • switched server from HTTP 2 to HTTP 1.1, no change

Here is the link to the site: https://www.partbeaver.com

Here are some crawl anomaly urls

https://www.partbeaver.com/ru/articles/weidmueller/1336720000-sm-37-5-17-mc-ne-gr/116172/

https://www.partbeaver.com/en/articles/rockwell-automation/800t-n32kf4b-/377033/

https://www.partbeaver.com/ru/articles/abb/2cds271103r0218-s201m-z1na/252490/

https://www.partbeaver.com/pl/articles/dold/0056769-mk9056-12-100-ac400hz-100-127v/632134/

Thank you very much

Using SharePoint search API to crawl document with c#

My scenario :

I have a SharePoint 2019 server that contain some documents with different file extension (ppt,doc,docx,xls,xlsx,pdf etc). I have manually forced indexing and crawling to make sure my search result are correct.

I was able to query SharePoint using the search API and what i got back was the file path of files containing the queried word and i can get the Hit-highlighted summary that is only a maximum of 10000 character.

What i have done is read that file and extract the paragraph where that queried word is found. I have been able to do it using file stream for docx,pdf,and txt file types.There 50 more file extension that i need to cater for.

My question : Is there another way to query the content of the search result returned by the search API instead of opening the individual file returned and reading the content?.

           Microsoft.SharePoint.Client.ClientResult<System.IO.Stream> stream = null;              KeywordQuery keywordQuery = new KeywordQuery(clientContext);              keywordQuery.QueryText = "SharePoint";              keywordQuery.EnablePhonetic = true;              keywordQuery.EnableOrderingHitHighlightedProperty = true;             //keywordQuery.SummaryLength = 500;               SearchExecutor searchExecutor = new SearchExecutor(clientContext);              ClientResult<ResultTableCollection> results = searchExecutor.ExecuteQuery(keywordQuery);              clientContext.ExecuteQuery(); 

Deadlock error messages in crawl server and search not working in SharePoint web application

In SharePoint 2010 Search Service Application when i remove content source and run again full crawl on same content source

I found below errors and users unable to search documents

A database error occurred. Source: Microsoft SQL Server Native Client 10.0 Code: 1205 occurred 1 time(s) Description: Transaction (Process ID 329) was deadlocked on lock resources with another process and has been chosen as the deadlock

CGatherer::LoadTransactionsFromCrawlInternal fails Error 0x80004005, CrawlID 1741342 [gatherobj.cxx:4663] d:\office\source\search\native\gather\server\gatherobj.cxx

SQL call failed (0x80004005: Unspecified error )
[atldbext.cxx:1290] d:\office\source\search\libs\utild\atldbext.cxx

A database error occurred. Source: Microsoft SQL Server Native Client 10.0 Code: 1205 occurred 1 time(s) Description: Transaction (Process ID 329) was deadlocked on lock resources with another process and has been chosen as the deadlock victim. Rerun the transaction. Context: Application ”

error 0: Transaction (Process ID 329) was deadlocked on lock resources with another process and has been chosen as the deadlock victim. Rerun the transaction. (source Microsoft SQL Server Native Client 10.0, 0x80004005, code 1205: Unspecified error ) [atldbext.cxx:1386] d:\office\source\search\libs\utild\atldbext.cxxa

Crawl your links fast by Rocket Fast Indexer (2000 Link’s crawling plan) for $19

Rocket Fast Indexer – Index your 1000 links faster! For $ 19 With the help of 15 years+ experience of SEO, Link building & Online marketing. I find that indexing of the backlinks has been always a big challenge for all of us – The Link builders. It’s, even more, harder nowadays after recent updates, as we all know. I have made my private “link crawler cum indexer”, Here are its features: – We don’t make any form of links for your backlinks. – The uniquely designed approach by me – Technically it Softly asks the G-bot to crawl.No extra investment of VPS or proxies or aged G accounts.Generally seen links starts indexing within 2-3 days of processing, most of the time even earlier too. Have recorded in our live tests up to 80% links indexing noticed in 1-2 months. Introductory Offer: 15% Off Coupon Code: 15off Give it a try with a minimum small plan and see the difference yourself before ordering any more.

by: nidhim
Created: —
Category: Link Building
Viewed: 263


Can you reduce yourself, crawl into the Tarrasque’s airway, and try to collapse its lung by enlarging inside of it?

A player tried to use the spell enlarge/reduce on his PC while in the mouth of the Tarrasque to shrink himself, then go deeper into its airway and to try to collapse its lung by enlarging himself inside of it. I imagine there’s no real rule for this but I felt silly just denying it, because it seemed like a cool idea that you could only probably do with a creature as large as the Tarrasque.

Can you crawl into a gargantuan creature’s mouth or into its airway via reduce/enlarge and attack its brain or internal organs?

SharePoint 2010 Search crawl gatherer’s temp folder got disappeared

I have 1 application server with SharePoint 2010 SP1 installed and another for Database SQL server 2008 R2. Suddenly the Temp folder for search crawl got disappeared and due to this search crawling got failed. It was fixed by manually creating those temp folders.

But after few months temp folders again got disappeared. We are not able to find root cause and permanent fix for this issue. It seems search service got corrupted and need to rebuild again. Does anyone has faced this issue. I don’t prefer rebuilding the search application as this issue is on Production server. Please let me know your views.

SharePoint 2013 Crawl error?

I want to enable search in my SharePoint 2013

when I click on crawl rule it give blow error:

 System.InvalidOperationException: The search service is not able to connect  to the machine that hosts the administration component. Verify that the  administration component '84caf2fc-6524-4f55-b090-07db8450fb0f' in search  application 'SPSearch Service Application' is in a good state and try again.    at   Microsoft.Office.Server.Search.Administration.SearchApi..ctor(SearchServiceApplication application)    at Microsoft.Office.Server.Search.Administration.SearchServiceApplication.get_SearchApi()    at Microsoft.Office.Server.Search.Administration.CrawlRuleCollection.InitCache()    at Microsoft.Office.Server.Search.Administration.CrawlRuleCollection..ctor(Content parent)    at Microsoft.Office.Server.Search.Administration.Content.get_CrawlRules()    at Microsoft.Office.Server.Search.Internal.UI.WebControls.ManageCrawlRulesControl.OnLoad(EventArgs e)    at System.Web.UI.Control.LoadRecursive()    at System.Web.UI.Control.LoadRecursive()    at System.Web.UI.Control.LoadRecursive()    at System.Web.UI.Control.LoadRecursive()    at System.Web.UI.Control.LoadRecursive()    at System.Web.UI.Control.LoadRecursive()    at System.Web.UI.Control.LoadRecursive()    at System.Web.UI.Page.ProcessRequestMain(Boolean  includeStagesBeforeAsyncPoint, Boolean includeStagesAfterAsyncPoint)