This is a web software project; the main logic part consists of PHP pages which are hosted on a Linux server. The web server must have PHP version 4.0 or above installed on it. The server should also support MySQL database as all the tables are developed in MySQL. The client part is a simple HTML page which contains some javascript as well.
Installation:
Database and Tables Creation:
i. Create a database in MySQL, I have named it as opalsyst_se. ii. Create a user in the database with all the access opalsyst_seusr1. iii. Create below tables and setup their relationships:
Counter
robotUrl urlDescription urlKeyword iv. Create the tables using below create command: create table counter (pos int(11) default 1); create table urlKeyword (url varchar(255) not null, keyword varchar(245) not null, rank int(11), primary key (url, keyword)); create table urlDescription (url varchar(255) not null, description varchar(245) not null); create table robotUrl (url varchar(255) not null, pos int(6) not null auto_increment, primary key (pos));
Database Setup:
For initial setup of the tables following commands will need to be run: insert into robotUrl values('www.cnn.com', 1); insert into robotUrl values('www.hollywood.com', 2); insert into counter values(1);
Above statements will initialize the counter table with value 1, and will add two URLs in the url table which will be the starting point of the robot. Robot will pick first URL and copy all the URLs into the robotURL table and go from there.
Server Setup:
Once the database is setup then copy all the PHP and other files onto the server in one directory using any FTP tool. Maintenance:
The robot can be called using the command line call to the robot.php file. Once started it starts adding websites to the table and keep spreading recursively over the web. If the start was an error then the process will need to be killed from command line itself. To make a fresh start