Sometimes people find they have been indexed by an indexing robot, or that a resource discovery robot has visited part of a site that for some reason shouldn't be visited by robots. In recognition of this problem, many Web Robots offer facilities for Web site administrators and content providers to limit what the robot does. This is achieved through two mechanisms:
A Web site administrator can indicate which parts of the site should not be vistsed by a robot, by providing a specially formatted file on their site, in http://.../robots.txt.웹사이트 관리자는 로봇의 정보수집을 원치않는다면 특별한 형식의 파일을 넣음으로써 해결할 수 있습니다. http://.../robots.txt
A Web author can indicate if a page may or may not be indexed, or analysed for links, through the use of a special HTML META tag.웹 페이지 제작자는 메타 테그를 이용해서 로봇의 접근을 차단할 수 있습니다.
The remainder of this pages provides full details on these facilities. Note that these methods rely on cooperation from the Robot, and are by no means guaranteed to work for every Robot. If you need stronger protection from robots and other agents, you should use alternative methods such as password protection.아래에는 위 두 가지 방법의 자세한 설명이 덧붙여져 있습니다.
User-agent: * Disallow: / | 모든 검색엔진이 긁어가는 것 모두 막기 |
User-agent: * Disallow: | 모두 허용하기 |
User-agent: * Disallow: /cgi-bin/ Disallow: /tmp/ Disallow: /private/ | cgi-bin 디렉토리, tmp 디렉토리, private 디렉토리 긁어가는 것만 막기 |
User-agent: empas Disallow: / | 엠파스 검색로봇만 긁어가기 제외 |
User-agent: webCrawler Disallow: | 웹크롤러 검색로봇만 긁어가기 허락 |