The item was aborted because the pipeline did not respond within the appropriate time. This item will be retried in the next crawl

We are facing two issues related to SharePoint search

Issue 1:

The item was aborted because the pipeline did not respond within the appropriate time. This item will be retried in the next crawl.

above error in SharePoint 2013 search crawl.

We have tried following,

  1. Index Reset in Search Service Application.
  2. Stop Timer Service – > Configuration Cache clear. -> Start Timer Service.
  3. Restart Search Host Control Service.
  4. Created New Search Service Application.
  5. Set Search performance level to Reduced and Increased Time Out value in Farm Search Administration.

Error is majority is on DispForm.aspx page. Solution suggested in following is also not working

Pipeline did not respond within the appropriate time

Anything which we are mission on?

Issue 2

We have created DateTime type managed property “ReceivedDate” in search schema which is searchable, Queryable, Retrievable.

But when we try to search only using ReceivedDate>01/01/2015 we are not getting any results. But if we concatenate with another text type managed property we can see the filtered results.

SharePoint 2016 unable to crawl new version PDF files

I am setting up a SP2016 on-premise farm. Some pdf files with newer version cannot be crawled. Here is my findings:

  1. Those pdf files are v1.7 (Acrobat 8.x, Acrobat 9.x)
  2. Search only return those pdf by search with file name. Search with pdf content will not return result.
  3. In the search result page, those files preview show

“If this message is not eventually replaced by the proper contents of the document, your PDF viewer may not be able to display this type of document … You can upgrade to the latest …”

  1. In the crawl log, those files are shown up as warning. Error is “Document was partially processed.”

Should I install pdf viewer on the crawling server to solve the problem?

Is the Gritty Realism variant incompatible with the classic dungeon crawl?

The variant Gritty Realism rule presented in the DMG (see p. 267) changes short rests from 1 hour to 8 hours, and long rests from 8 hours to 7 days. Its description suggests, casually and without elaboration, that

[t]his approach encourages the characters to spend time out of the dungeon.

In other words, 5e’s designers evidently thought the rule’s ramifications would increase the difficulty of dungeons enough to warrant comment, yet unhelpfully declined to articulate why. Given that dungeons are a namesake of the game, this is a regrettable oversight, even for a variant rule. DMs and players considering the rule are left guessing as to what pitfalls they should expect — e.g., what game mechanics are made more complicated by the design principles informing an archetypal dungeon-based adventure. For 5e veterans, that might not be a heavy lift, but those with less (or no) experience face potential frustration.

A number of Q&As here on RPG.SE have discussed Gritty Realism, e.g.:

  • What game mechanics may be inadvertently broken by changing the time required for resting?
  • How can I reduce the number of encounters per day without throwing off game balance?
  • How do the activity limitations for a long rest work in the Gritty Realism variant?

However, none have meaningfully examined why — or even whether — dungeons might be especially problematic in a game using Gritty Realism.

Given the game mechanics implicated, is the DMG‘s observation that Gritty Realism discourages dungeon crawling really accurate? Might it be overstated? Might it be understated, such that groups primarily interested in dungeon-delving should absolutely eschew Gritty Realism?

The item was deleted because it was excluded by a crawl rule

My search crawl is not working for my local site. On my Crawl Log I see a warning:

Item not crawled due to one of the following reasons: Preventive crawl rule; Specified content source hops/depth exceeded; URL has query string parameter; Required protocol handler not found; Preventive robots directive. ( This item was deleted because it was excluded by a crawl rule. )

Important information:

  • I just moved all my content databases to a newly installed Sharepoint on a new server
  • I am running Sharepoint Server 2019 Standard
  • When I change the address in the content source to full FQDN, the search works partially with many 401 warnings
  • The address in question is https
  • Same address with sps3 gets crawled successfully with no issues
  • There is no robots.txt file in the site

I tried and checked everything I could think of, including:

  • I checked that the default access account has access to the website
  • I browsed the website ( using a browser ) while being logging in using the default access account
  • I made sure there are no crawl rules
  • I created a new crawl rule to include the affected URL
  • I did “index reset”
  • I disabled loopbackCheck
  • I used Fiddler to monitor the traffic, the crawler is not even accessing the site

If you have any suggestions, please advice.

High Number of Crawl Anomaly Errors

since April 9, i am receiving a steadily increasing number of Crawl Anomaly Errors. The number of crawl anomaly pages is currently over 22000.

  • Each page in the error list can be reached by live url testing from the search console and by curl or webbrowser.
  • The live test rendering is also successful.
  • Server logs don’t even have records of Google visiting the crawl anomaly pages.
  • There are no javascript errors or 404 resources.
  • The load time of most of the crawl anomaly flagged pages is <1.5s till initial paint and <2.5s till final paint ( tested by several sites)
  • Server logs show no signs of downtime or server side errors in this time period
  • Other pages are visited by Google, but the number of indexed pages is not increasing
  • Some of the flagged pages were crawled normally by Yandex and Bing
  • i have tried clicking on check for fix of the errors, but the check has failed multiple times
  • When i do the live test on a single crawl anomaly url, they become marked as crawled, but not indexed
  • Pages have some Rich Data warnings, but no errors
  • There are a couple of soft 404 errors, about 122
  • the number of indexed pages is steadily declining
  • switched server from HTTP 2 to HTTP 1.1, no change

Here is the link to the site: https://www.partbeaver.com

Here are some crawl anomaly urls

https://www.partbeaver.com/ru/articles/weidmueller/1336720000-sm-37-5-17-mc-ne-gr/116172/

https://www.partbeaver.com/en/articles/rockwell-automation/800t-n32kf4b-/377033/

https://www.partbeaver.com/ru/articles/abb/2cds271103r0218-s201m-z1na/252490/

https://www.partbeaver.com/pl/articles/dold/0056769-mk9056-12-100-ac400hz-100-127v/632134/

Thank you very much

Using SharePoint search API to crawl document with c#

My scenario :

I have a SharePoint 2019 server that contain some documents with different file extension (ppt,doc,docx,xls,xlsx,pdf etc). I have manually forced indexing and crawling to make sure my search result are correct.

I was able to query SharePoint using the search API and what i got back was the file path of files containing the queried word and i can get the Hit-highlighted summary that is only a maximum of 10000 character.

What i have done is read that file and extract the paragraph where that queried word is found. I have been able to do it using file stream for docx,pdf,and txt file types.There 50 more file extension that i need to cater for.

My question : Is there another way to query the content of the search result returned by the search API instead of opening the individual file returned and reading the content?.

           Microsoft.SharePoint.Client.ClientResult<System.IO.Stream> stream = null;              KeywordQuery keywordQuery = new KeywordQuery(clientContext);              keywordQuery.QueryText = "SharePoint";              keywordQuery.EnablePhonetic = true;              keywordQuery.EnableOrderingHitHighlightedProperty = true;             //keywordQuery.SummaryLength = 500;               SearchExecutor searchExecutor = new SearchExecutor(clientContext);              ClientResult<ResultTableCollection> results = searchExecutor.ExecuteQuery(keywordQuery);              clientContext.ExecuteQuery(); 

Deadlock error messages in crawl server and search not working in SharePoint web application

In SharePoint 2010 Search Service Application when i remove content source and run again full crawl on same content source

I found below errors and users unable to search documents

A database error occurred. Source: Microsoft SQL Server Native Client 10.0 Code: 1205 occurred 1 time(s) Description: Transaction (Process ID 329) was deadlocked on lock resources with another process and has been chosen as the deadlock

CGatherer::LoadTransactionsFromCrawlInternal fails Error 0x80004005, CrawlID 1741342 [gatherobj.cxx:4663] d:\office\source\search\native\gather\server\gatherobj.cxx

SQL call failed (0x80004005: Unspecified error )
[atldbext.cxx:1290] d:\office\source\search\libs\utild\atldbext.cxx

A database error occurred. Source: Microsoft SQL Server Native Client 10.0 Code: 1205 occurred 1 time(s) Description: Transaction (Process ID 329) was deadlocked on lock resources with another process and has been chosen as the deadlock victim. Rerun the transaction. Context: Application ”

error 0: Transaction (Process ID 329) was deadlocked on lock resources with another process and has been chosen as the deadlock victim. Rerun the transaction. (source Microsoft SQL Server Native Client 10.0, 0x80004005, code 1205: Unspecified error ) [atldbext.cxx:1386] d:\office\source\search\libs\utild\atldbext.cxxa