head	1.2;
access;
symbols
	RELEASE_8_3_0:1.1
	RELEASE_9_0_0:1.1
	RELEASE_7_4_0:1.1
	RELEASE_8_2_0:1.1
	RELEASE_6_EOL:1.1
	RELEASE_8_1_0:1.1
	RELEASE_7_3_0:1.1
	RELEASE_8_0_0:1.1
	RELEASE_7_2_0:1.1
	RELEASE_7_1_0:1.1
	RELEASE_6_4_0:1.1
	RELEASE_5_EOL:1.1;
locks; strict;
comment	@# @;


1.2
date	2012.11.17.06.02.55;	author svnexp;	state Exp;
branches;
next	1.1;

1.1
date	2008.02.16.13.52.51;	author miwi;	state Exp;
branches;
next	;


desc
@@


1.2
log
@Switch exporter over
@
text
@# New ports collection makefile for:	p5-WWW-RobotRules-Parser
# Date created:		2008-01-30
# Whom:			Masahiro Teramoto <markun@@onohara.to>
#
# $FreeBSD: head/www/p5-WWW-RobotRules-Parser/Makefile 300897 2012-07-14 14:29:18Z beat $
#

PORTNAME=	WWW-RobotRules-Parser
PORTVERSION=	0.04001
CATEGORIES=	www perl5
MASTER_SITES=	CPAN
PKGNAMEPREFIX=	p5-

MAINTAINER=	markun@@onohara.to
COMMENT=	Just Parse robots.txt

BUILD_DEPENDS=	p5-libwww>=0:${PORTSDIR}/www/p5-libwww \
		p5-URI>=0:${PORTSDIR}/net/p5-URI
RUN_DEPENDS=	p5-libwww>=0:${PORTSDIR}/www/p5-libwww \
		p5-URI>=0:${PORTSDIR}/net/p5-URI

PERL_CONFIGURE=	yes

MAN3=		WWW::RobotRules::Parser.3

.include <bsd.port.mk>
@


1.1
log
@WWW::RobotRules::Parser allows you to simply parse robots.txt files
as described in http://www.robotstxt.org/wc/norobots.html.

WWW:	http://http://search.cpan.org/dist/WWW-RobotRules-Parser/

PR:		ports/120160
Submitted by:	Masahiro Teramoto <markun at onohara.to>
@
text
@d5 1
a5 1
# $FreeBSD$
@

