A Search engine uses a crawler (or a spider) to retrieve and index web pages. A “general crawler” crawls across the Internet by following only the hyperlinks directly provided by web pages. On the other hand, a “deep crawler” has the capability of retrieving all contents of web pages whether the hyperlinks are available or not. The process of a crawler’s work can be split into two major steps: the “retrieving” step and the “storage and indexing” step. As far as the retrieving step is concerned, even though a deep crawler can generate all URLs by itself, there should not be any unauthorized access at all. However, a crawler, either a general or deep one, will be charged with trespassing to chattel if it interferes in the website, whereas, in the “storage and indexing” step, owners’ copyright of web pages will in principle not be infringed upon by a crawler. Nevertheless, there are controversies about “cache”, which is provided by certain popular search engines, such as Google and Yahoo. Search engines may claim their fair use defense of “cache”, because one of the main goals of search engines is to improve access to information on the internet. Actually, we must consider many factors before we decide whether “cache” is fair or not, for instance, whether there is competition between the search engine and the retrieved web sites; and what percentage of the contents is retrieved by the crawler, etc. To resolve all possible legal conflicts between crawlers and web pages, among several economic resolutions, a newer and complete robot exclusion standard is required. In addition, a self-detecting mechanism adopted by web pages may be even a more powerful one.