/* ** (c) COPYRIGHT MIT 1995. ** Please first read the full copyright statement in the file COPYRIGH. */
This module parses a robots.txt exclusion file which nice robots are expected to honor. Together with the robot META tags, the webbot should now behave itself on the Internet.
#ifndef ROBOTTXT_H
#define ROBOTTXT_H
typedef struct _user_agent_ {
char * name;
HTList * disallow;
} UserAgent;
extern char * skip_comments(char *ptr);
extern UserAgent * new_user_agent(void);
extern BOOL set_name_user_agent(UserAgent *ua,char *name);
extern char * get_name_user_agent(UserAgent *ua);
extern BOOL add_disallow_user_agent(UserAgent *ua, char *disallow);
extern HTList * get_disallow_user_agent(UserAgent *ua);
extern BOOL delete_user_agent(UserAgent *ua);
extern BOOL delete_all_user_agents(HTList *user_agents);
extern void print_user_agent(UserAgent *ua);
extern void print_all_user_agents(HTList *user_agents);
extern HTList * get_all_user_agents(char *rob_str);
extern BOOL get_user_agents(char * ptr, HTList *user_agents);
extern BOOL put_string_disallow(HTChunk *ch, UserAgent *ua);
extern char * get_regular_expression(HTList* user_agents, char *name_robot);
extern char * scan_robots_txt(char *rob_str, char *name_robot);
#endif