- The search crawler lets you implement federated search without developer resources. You can set up multiple crawlers in your help center to crawl and index different content in the same or different websites.
- The Federated Search API is available if you have developer resources. This REST API that lets you ingest records of your external content into the Zendesk search indexes to implement federated search in your help center.
After you set up federated search, you need to configure the content that you want to include in help center search results (see Including external content in your help center search results).
Setting up the search crawler for federated search for your help center
The search crawler lets you implement federated search in your help center without developer resources. You can set up multiple crawlers in your help center to crawl and index different content in the same or different websites.
To set up the search crawler for federated search
-
In Guide, click the Settings icon (
) in the sidebar, then click Search settings.
- Under Crawlers, click Manage.
- Click Add Crawler.
- In Name this crawler, enter the following:
- Name that you want to assign to the crawler. This is an internal name that identifies your search crawler on the crawler management list.
-
Owner who is the Guide admin responsible for crawler maintenance and troubleshooting. By default, the crawler owner is the user creating the crawler, however you can change this name to any Guide admin.
Crawler owners receive email notifications both when the crawler runs successfully and when there are error notifications, such as problems with domain verification, processing the sitemap, or crawling pages.
- In Add the website you want to crawl, verify ownership of your domain by configuring the following:
- Website URL - Enter the URL of the website that you want to crawl.
-
Domain ownership verification - Click Copy to copy the HTML tag onto your clipboard, then paste the tag into the <head> section in the HTML code of your site's non-authenticated home page. You can do this after you complete the crawler setup and you can always find the verification tag on the edit crawler page. See Managing search crawlers.
Note: Do not remove the tag once it is in place, as the crawler needs to complete successful domain verification each time it runs.
- In Add a sitemap, in Sitemap URL, enter the URL for the sitemap you want the crawler to use when crawling your site.
The sitemap must follow the sitemaps XML protocol and contain a list of all pages within the site that you want to crawl. The sitemap can be the standard sitemap containing all the pages of the site or it can be a dedicated sitemap that lists the pages that you want to crawl. All sitemaps must be hosted on the domain that the crawler is configured to crawl. The search crawler does not support sitemap indexes.
You can set up multiple crawlers on the same site that each use different sitemaps defining the pages you want the search crawler to crawl.
- In Add filters to help people find this content, configure the source and type filters used to filter search results by your end users. Source refers to the origin of the external content, such as a forum, issue tracker, or learning management system. Type refers to the kind of content, such as blog post, tech note, or bug report.
- Source - Click the arrow, then select a source from the list or select + Create new source to add a name that describes where this content lives.
- Type - Click the arrow, then select a type from the list or select + Create new type to add a name that describes what kind of content this is.
- Click Finish.
The search crawler is created and pending. Within 24 hours the crawler will verify ownership of the domain and then fetch and parse the specified sitemap. Once the sitemap processing succeeds, the crawler begins to crawl the pages and index its content. If the crawler fails either during domain verification or while processing the sitemap, the crawler owner will receive an email notification with troubleshooting tips to help resolve the issue. The crawler will try again in 24 hours.Note: Zendesk/External-Content is the user agent for the search crawler. To prevent the crawler from failing due to a firewall blocking requests, whitelist (or allow-list) Zendesk/External-Content.
After you set up the search crawler, you need to select the content that you want to include and exclude in help center search results. See Including external content in your help center search results.
You can also include external content in search results in the knowledge section of the context panel for agents. See Configuring Knowledge in the context panel.
Using the API to configure federated search for your help center
Your developers can set up federated search in your help center using the Federated Search API. This method requires that your developers build and maintain a middleware layer to integrate the service or site that hosts the external content and the help center.
- Build your own integration with the Zendesk REST API then ingest the content you want to show up in your search results. See the Federated Search API reference documentation.
After you set up federated search, you need to select the content that you want to include and exclude in help center search results. See Including external content in your help center search results.
You can also include external content in search results in the knowledge section of the context panel for agents. See Configuring Knowledge in the context panel.
22 comments
Abdullah Ali
well-documented content 👍
0
Julien SERVILLAT
Hello, do you know if the search crawler can be used to index content in a Jira Confluence site ? Many thanks !
0
Jeff C
Hello Julien,
Yes! The crawler can be used to index content from your Confluence site. If you are running into issues setting it up, please reach out to us directly for support.
0
Matt Farrington-Smith
Hey Julien SERVILLAT did you manage to get this set up and working?
@... we were told this wasn't possible, so am getting conflicting information...
0
Julien SERVILLAT
Hello Matt Farrington-Smith, yes we managed to setup the crawler with Confluence. Indexing of the article works, we are just finalizing the upgrade of the theme to include the correct placeholders to return the results of federated search.
0
Cesar
hello, the content of the crawler, after a search, is presented embedded or it is just a link that pops-out a window or tab to the external content?
I ask this because I am wondering if it make sense to host HTML content in a s3 bucket, without custom domain. if the crawler embeds the content, it doesn't matter not having a custom domain. different to the case if it pops-it out.
I suspect not using the custom domain may create problems like the domain-verification rule..
I hope my question makes sense!
0
Dave Dyson
When users perform a search, relevant external content discovered by the crawler is ranked and presented on the search results page, where users can filter the results and click the links to view the external content link in another browser tab.
For more information, see About Zendesk Federated Search
0
Marcus Ko
Hi,
Do we have guide on deploying search crawlers on MS Sharepoint?
0
Sarah Rebecca Miller
Hello, Is the crawled external visibility set to Everyone? Is there a way to control the visibility settings of the crawled external content?
We'd like to limit the crawled external content to agents and admins for one of our use cases but based on what's available in the settings, I assume that it's not possible.
0
Norm Van Eeden Petersman
I'm having difficulties getting external content to show up for everyone in search results. I have enabled it and it is verified that the crawler is working. Am I missing something?
0
mfg
When configuring the crawler, if you designate a site while including a subdomain, will the crawler cover the entire domain or just the subdomain?
Is the crawler's verification tag persistent across crawlers or is a new one generated for each crawler? For instance, if I start configuring a crawler, copy the tag, but don't save the crawler because I don't have the sitemap URL, will that tag be the same when I go back to finish creating the crawler?
1
Dane
This is visible to Everyone and it is not possible to restrict it the same way as user segment for Help Center articles.
Hi mfg,
For every domain that you will designate for the crawler, it will have a different verification tag. Same goes if you will create another crawler for the same domain that has already been verified.
0
Chuck Mikuzis
Has anyone attempted this with Madcap/Flare? We're testing currently and and we're unable to verify the domain in our POC for a <companyname>.mcoutput.com domain. I'm guessing it is because we don't own the Madcap domain of mcoutput.com, am I right?
1
System
We are using an eccommerce platform that has their site index as storename/xmlsitemap.php
looks like this requires it to be .xml? google crawler has no issue with our sitemap. What can i do?
0
Håvard Skare
Hi.
I am trying to set up a search crawler for our help desk. I have embedded the meta tag in the header of the target webpage, the url for the site and for the sitemap are both correct, but It still gives me a warning saying that the domain cannot be verified and that it cannot find the sitemap. What can I do?
1
Rajan
Hey Elizabeth Williams
I want to initiate re-indexing URL's for google. Will this help me with that too?
0
Destiny
Thank you for getting in touch.
It seems you're interested in whether reindexing URLs with Google will enhance the visibility of your pages in search results. Have I understood that correctly?
It's important to note that the process of reindexing URLs through Google Search Console is separate from setting up Zendesk's search crawler. They serve distinct functions. Could you clarify your specific requirements or the goal you're aiming to achieve? This will enable me to offer more tailored advice.
0
Anne-Flore Caire
Hello. We have a load problem on our own sites that we index with Zendesk crawlers, and it's apparently not possible to customize/modify the frequency or speed of the crawlers. Is an evolution possible or planned?
0
Shawna James
0
Sebastiaan Cools
is it possible to have the crawler search a SharePoint site with authentication?
0
Robert Zahn
unfortunatly it is not possible to use script generated XMLs like /zendesk/sitemap.xml?path=/topic/*&template=tutorial-article. Can you say why there are restrictions?
0
Destiny
Thank you for your inquiry regarding the use of script-generated XMLs, such as /zendesk/sitemap.xml?path=/topic/*&template=tutorial-article. I understand that these restrictions can be frustrating, and I’d like to provide some insight into why they are in place.
0