The program is a Hit Bot which mimics a web surfer and the "visit" is logged on the other site as a normal page visit with no trace of a computer program being used. The Hit Bot eats cookies and leaves a footprint exactly that of a normal human web surfer.
The program uses a list of proxies for each hit, and this is where a minor modification is required: If you've used proxies with your web browser, you will know that many proxy servers automatically display a "welcome" page upon first connecting, and you must refresh the page before you can begin surfing. Unfortunately, the bot was not coded with this in mind, and therefore when it comes across a "welcome" page, the bot thinks that it has made a successful hit and stops there.
What we need to do is to create a simple procedure where upon first connecting to a proxy server, the bot will search a list of predefined keywords or whole strings to see if it's reached a "welcome" page. These keywords/strings can later be defined by me. If the bot does detect a "welcome" page, it should simply repeat the hit using that proxy, which has the same effect as refreshing the page in a web browser.
I have attached a snippet of the procedure from the perl script dealing with the botting mechanism. Please note that most of the variables, including $proxy are defined elsewhere in the script. Also bear in mind that the script works from a MySQL database and supports "deep botting" which is the purpose of the grab_links subroutine.
I'm looking for an experienced programmer who would be able to undertake this task within a matter of days! :)