I have a web site for scraping purposes made by a developer who couldnt finish the project and I have to deliver it in less than 8 hours.
the idea is to find illegal hotels in 3 touristic portals against a csv file uploaded by the user.
here is how it should work:
1) the user logs in and copies/paste a url from 3 touristinc portals. They are booking.com, despegar.com.ar and welcomeargentina.com). The URLs will be MANUALLY ADDED (no hard coding)
2) The user uploads his csv file and a string (another field) for statistic purposes.
3) The user clicks on BEGIN and a task is added to the cron. The cron runs within some hours. No rush here.
4) Once the scrap is done, the site should email to him/her and the results should be shown in the reports panel
the scrap should work based on geolocation first (postal address) and hotel name second.
in order to avoid false positives, the idea is to "convert" the postal addresses uploaded from the csv file against the url's map location (see booking.com as an example where you can see the property in the map with the address. If that matches, we are good! got the idea?)
Reports panel: allow each user (depending of their permissions) to generate statistics and graphs. Examples: amount of analysis per city, % of illegal hotels, amount of published hotels per site, etc
NEED IT BY TODAY. MSG ME ASAP. THANK YOU