There are also a number of non-traditional uses for crawlers, such as archiving content or generating statistics. 爬虫也有一些非传统的用途,例如对内容进行归档或生成统计数据。
One such set of tools are the robots, spiders, and Web crawlers that aggregate information for Internet and intranet search engines. 这种工具有robot、spider和为Internet和Intranet搜索引擎搜集信息的Web爬行榜。
E-mail harvesting crawlers search Web sites for e-mail addresses that are then used to generate the mass of spam that we all deal with each day. 这种爬虫会在Web站点上搜索e-mail地址,然后生成我们每天不得不处理的大量垃圾邮件。
However, crawlers don't care about these; they ignore ( and should!) 然而,爬虫程序却对此漠不关心;他们忽略了(不应该忽略!)
If this pattern continues, crawlers are squeezed between the dual pressures of maintaining acceptable crawling performance while gathering constantly-increasing volumes of information. 如果这种模式继续下去的话,爬虫程序会受到双重的压力,在维持可接受的爬行性能的同时又要聚集不断增加的信息量。
The final step is to allow this XML document to be made accessible to crawlers through a Web server. 最后一步需要使爬虫程序能通过Web服务器访问这个XML文档。
Once the Sitemap XML is completed, you can make it available to Web crawlers by submitting a URL to the sitemap to those sites that support the protocol. 完成SitemapXML文件后,便可以向支持这一协议的站点提交站点地图的URL,这样web爬虫程序便可以使用该文件了。
Helps to optimize the entire structure of the site for a crawler by providing an alternate set of Web pages so that crawlers can quickly access and index a large number of embedded pages. 有助于为爬网程序优化站点的整个结构,通过提供备用的Web页面组,从而使爬网程序能够快速地访问大量的嵌入式页面并针对这些页面建立索引。
It also provides crawlers that can ingest content from databases and XML files, as well as Web pages. 它还提供了可以从数据库、XML文件和Web页面中收集内容的信息收集器(crawler)。
These crawlers look for missing content, validate all links, and ensure that your Hypertext Markup Language ( HTML) is valid. 这些爬虫会查找缺少的内容、验证所有的链接,并会确保超文本标记语言(HTML)是有效的。
Making the sitemap accessible to crawlers 使站点地图能被爬虫程序访问
Other useful specialized crawlers include Web site checkers. 其他有用的专用爬虫包括Web站点检查器。
These new semantics are important for non-visual users of the page ( including anyone using assistive technologies like screen readers) but also for computer programs like search engine crawlers. 这些新语义对于页面的非视觉用户(包括使用屏幕阅读器这样的辅助技术的用户)和搜索引擎爬虫这样的计算机程序很重要。
But convincing Americans to eat anything containing creepy crawlers is not an easy sell. 但要说服美国人吃任何一种以令人毛骨悚然的爬虫为原料的食物,可不是一件容易做到的事情。
The crawlers will search Microsoft products such as SQL Server, Excel and SharePoint and extra metadata. 抓取工具将搜索微软的产品,如SQLServer、Excel、SharePoint和额外的数据。
Google itself offers general advice on how best to present material so that it is easily digestible by the crawlers that scan the web. 就如何最好地呈现内容,以便扫描网页的爬虫能够轻松领悟,谷歌提出了自己的一般性建议。
Searchable, but not always found: In recent years, search engines have introduced the ability to index swf files into their respective Web crawlers. 在最近几年,搜索引擎已经能在各个页面上索引.swf文件。
This helps customers and search engine crawlers. 该页面对客户和搜索引擎都是有所帮助的。
Automation detection-Detecting bots, crawlers, scanners and other surface malicious activity. 自动检测&对机器人、爬虫、扫描及其他恶意行为进行检测。
Some people use the creepy crawlers to catch fish. 有些人利用令人毛骨悚然的爬虫捕鱼。
This equipment adsorption on the steel wall surface with permanent magnetic, and drive through the trailing wheel in the double crawlers. 该爬行器采用永磁铁吸附在钢板表面上,通过双履带后轮驱动,可以在船体壁面上连续爬行,并有一定的越障能力。
XML sitemap is a file that gives instructions to web crawlers about the website. XML网站地图可给予网络检索器关于你的网站的指导。
The walkers moved faster than the crawlers but meandered& a good strategy for exploring new territory. 步行者比爬行者运动的要快,但是曲折前进&一种探索新领域的好策略。
Site-based Web Crawler is crawlers using site as its basic unit. 基于站点的采集器是一类有针对性的采集器,它以站点作为采集的单位。
Topic crawler based on dynamic topic base was proposed by studying on topic crawlers which filter URLs based on different strategies. 通过对基于不同策略过滤URL的主题爬虫的研究,提出了一种基于动态主题库的主题爬虫。
Research on Parallel Crawlers Based on Multi-Task and Multi-Channel Techniques 面向多任务、多通道并行爬虫的技术研究
Our crawlers are well designed and professionally operated, providing an invaluable service that is in accordance with search industry standards. 我们的爬虫很好被设计并且职业地操作了,提供根据搜索工业标准的一种无价的服务。
The crawlers are innocuous and cause no harm to an owner's site or servers. 爬虫是无害的并且不引起伤害到一位主人的地点或服务者。
Bless the land and allow the land to bless you in return along with each kingdom associated of plant, animal, mineral fish or sea creatures, dolphin or whale and creepy crawlers. 祝福土地,允许土地以及每个王国,包括植物,动物,矿石,鱼类,海洋生物,海豚鲸鱼以及爬行生物回报给你祝福。
So if you're a small business that relies on local search traffic, it's a good idea to make sure search engine crawlers have a good sense of where you're located. 所以,如果你经营着一家依赖本地搜索流量的小企业,那么,明智的办法就是设法确保搜索引擎爬虫知道你在哪里。