显示下一次爬网的开始日期和时间。
显示爬网开始时表的架构版本。
Displays the schema version of table at time the crawl started.
“否”,则表示爬网正在进行。
显示最近一次爬网是否已完成。
重新对内容源进行爬网。
显示在当前或最近一次爬网中更新的行数。
Displays the number of rows updated in the current or most recent crawl.
在配置搜索设置页面中,点内容源和爬网计划。
On the Configure Search Settings page, click Content sources and crawl schedules.
显示当前或最近一次爬网的开始日期和时间。
Displays the date and time that the current or most recent crawl started.
因此,已定义的爬网程序和维护处理可能不会运行。
Therefore, the defined crawlers and the maintenance process might not be running.
包含有返回到顶层页面的链接的页面将有助于爬网。
Pages with links back to the top-level pages will help with crawling.
在大列表中还有要考虑的是爬网时间和爬网超时。
Another consideration with very large lists is crawl time and crawl time-outs.
为了更容易地爬网,请确保页面含有返回到顶层页面的链接。
Ensure that your pages have links back to top-level pages for easy crawling.
在爬网程序运行时,不同搜索源的内容可能会稍微失去同步。
The content of the different search collections might be slightly out-of-sync while crawlers are running.
点击我们内容源上的菜单,选开始完全爬网或开始增量爬网。
Click the menu for your content source and click either Start Full Crawl or Start Incremental Crawl.
在搜索管理Portlet中,创建集合,并配置爬网程序和分类。
In the search admin portlet, create your collection, and configure your crawlers and taxonomies.
它使用一个调度程序来发起周期性的事件,比如爬网程序执行和全文索引维护。
It USES a scheduler to initiate periodic events, such as crawler executions and full-text index maintenance.
注意,我们可以现在就对业务数据进行爬网,也可以在我们创建好范围后再爬网。
Note that you can either crawl Business Data Catalog content now, or you can crawl it after you create a search scope.
返回可供全文爬网或全文爬网范围的全文收集器使用的共享内存池的有关信息。
Returns information about the Shared memory pools available to the full-text gatherer for a full-text crawl or a full-text crawl range.
每一个服务器上都运行爬网程序和建立维护操作的索引,这将增加系统资源的消耗。
The crawlers and index maintenance operations run on every server which increases system resources consumption.
在我们创建完爬网计划后,就可以开始为我们的BDC应用程序定义一个搜索范围了。
After you create your crawl schedule, define a search scope for your Business Data Catalog application.
为搜索引擎爬网程序提供入口点,以使爬网程序轻松地跟踪您的Web页面内的链接。
Provides an entry point for the search engine crawler to easily follow the links within your Web pages.
如果有多于一个BDC应用程序存在,我们可以在这里选择是否对多个应用程序进行爬网。
If more than one Business Data Catalog application exists, you can choose to have multiple applications crawled in the applications section.
在每个服务器上,调度程序将启动爬网程序,并且将对集合进行更新以使其具有相同的内容。
On each server, the scheduler will start the crawlers, and the collections will be updated to have the same content.
站点地图对于爬网深度是非常有用的,它使爬行器能够轻松地访问并建立深嵌于该站点内的页面的索引。
A site map is useful to facilitate deep crawling so that crawlers can easily visit and index pages deeply embedded within the site.
高效站点地图将会构建为站点中不同页面的简要概述,并且会在尽可能少的页面中为爬网程序呈现大量的链接的列表。
An effective site map is built as a high level overview of the different pages in the site and presents a large list of links to the crawler in as few pages as possible.
因此,聚合的方式非常方便,但是却难以实现,因为它需要通过爬网的方式处理所有可用的数据来源,并且扩展搜索引擎。
Thus, aggregation is handy but difficult to achieve, since it requires crawling through all the available data sources and scaling the search engine.
此方法会创建一个高效的站点地图,以供搜索引擎爬网程序在您的整个站点中遍历尽可能多的URL并针对这些URL建立索引。
This method creates an effective site map for search engine spiders to traverse and index as many URLs as possible in your entire site.
为了能够在一个搜索结果WebPart中使用我们刚刚创建的范围,我们需要返回到搜索设置页面,并对我们的内容源重新爬网。
To use the scope you just created in a Search Results page, click Search Settings in the breadcrumb trail.
如前所述,PortalSearch包括用来启动爬网程序和索引维护处理的调度程序,其中索引维护处理可以删除过时的文档和断开的链接。
As described above, Portal Search includes a scheduler to start the crawlers and the index maintenance process that removes outdated documents and broken links.
使用PortalSearch,管理员可以定义内容源爬网程序,爬网程序使用HTTP协议来抓取Web站点或内容存储库并为其建立索引。
Using Portal Search, administrators can define content source crawlers which use the HTTP protocol to crawl and index Web sites or content repositories.
应用推荐