You've landed at the right place. oDesk is now Upwork. Learn about the new platform.

Web Scraping Jobs

161 were found based on your criteria

show all
show all
only
only
only
show all
only
only
only
only
only
show all
only
only
only
Fixed-Price - Est. Budget: $ 200 Posted
Looking for someone to recreate a section of Amazon.com bestsellers for our site. This will recreate the six tab directory seen here and populate it with our affiliate code on images and links. Each Category and subcategory has 100 items displayed across 10 pages each. There are roughly 30 Categories, each with five pages of products. with up to 5 subcategories each. Project will be done utilizing Fresh Store Builder. I will send you a link with the section to be recreated. Prefer candidates with strong Amazon Web Services, API and Fresh Store Builder experience. You must adhere to deadlines and have good English communication skills. Job Summary 1)Set up a Wordpress CMS install our our VPS 2)Make that installation a Multisite Install 3)Set up primary and secondary domain (the top 100 will be under our primary domain. 4)Set up an index page with Amazon Top Rated, categories and subcategories. Along with Index tabs across the top of the page. (we will provide this link, LOOK...
Hourly - Est. Time: Less than 1 month, 10-30 hrs/week - Posted
I have a database of approximately 15,000 email addresses and school names. I would like to add the following columns to the database, based on the email address: - Contact name (if it is a personal email address) - School Name (provided for 95%, but preferable to double check name for accuracy) - First Line Address - Second Line Address - City - Post code / Zip code - Country This will need to be done either my manual scraping or some automated scraping solution.
Fixed-Price - Est. Budget: $ 60 Posted
Write 4 scripts to extract data from dinamic webs, Use web scraping to get data from productos from some web stores: The webs to scrap are: 1.- Wallmart http://www.walmart.com.mx/super/Busqueda.aspx?Departamento=d-Carnes-y-Pescados&Familia=f-Carne-de-res&Linea=l-Cortes 2.- Superama. http://www.superama.com.mx/superama/inicio.aspx 3.- Comercial Mexicana https://www.lacomer.com.mx/lacomer/doHome.action?key=Lomas-Anahuac&succId=14&succFmt=100 4.- Soriana http://www1.soriana.com/site/default.aspx?p=12118&temprefer=25134415 You will get such data for México. Superama and Wallmart have the same price and products for every store on México. Comercial Mexicana and Soriana have diferent prices and products for each store on México, so you have to get the data for each store. –You will get for every web (if apply) the following information: (State, city, store, department, category subcategory, product name, price, amount/measure, presentation, branch and price.) –Only...
Hourly - Est. Time: Less than 1 week, 10-30 hrs/week - Posted
I am looking for someone who is expert in Web research & data entry both. Your main duty contact info finding from web and input into Excel spreadsheet. I will provide company name and city, state where you will find them. Must have Skype and fluent in English speaking. Filipino is preferred. Thanks
Hourly - Est. Time: Less than 1 week, Less than 10 hrs/week - Posted
I have a list of companies, as well as phone numbers and addresses, that I am looking to find President/CEO name and email address. I am looking for someone who has the technology to find these addresses through a crawler. I am NOT looking for someone to manually find them. I would like to try someone out first for 3-5 hours and see how they do before committing to more databases.
Hourly - Est. Time: Less than 1 month, Less than 10 hrs/week - Posted
We are soliciting for the DESIGN of a META SEARCH engine and dynamic reporting system which will search The VERIS COMMUNITY data and Framework (Public Domain) to produce search results on any search topic; allow users to add content back in to our data repository in the same formats and data input fields as VERIS; and allow for adHoc reporting. In essence we are asking for the design of a dynamic Security Breach application and reporting system. The winning design will have a search, membership sign up, reporting, data analytics dashboard and community feel to the product which will be provisioned as a website and web application with mobile extensions (ability to be seen on mobile devices) and alerting capabilities for any new data that is meaningful. The winning design will also take in to account the ability to search additional data sources both locally stored in the application/website and from third party resources (a service integration layer to connect to other data repositories). Ideas,...
Hourly - Est. Time: Less than 1 month, Less than 10 hrs/week - Posted
We are soliciting for the DESIGN of a META SEARCH engine which will search multiple data HETEROGENOUS repositories of genomic data across multiple locations that are web accessible and be able to search future repositories that area publicly accessible in Google's cloud computing environment and in AWS. The design can be either realtime and not store any data in our own repository (ex. Kayak search) or can electronically transfer data in batches to a central cloud repository and then be searchable. In fact, we expect a description of the tradeoffs between both in any design proposal. The list of sources for the INITIAL set of data can be found: https://gds.nih.gov/02dr2.html Concepts like downloading the repositories or transferring these repositories to a single central repository are acceptable but only if Practical and LEGAL as part of the terms and conditions of the sources. A realtime meta search that only indexes the repositories is preferred. The UI design and UX is not a...
Hourly - Est. Time: 1 to 3 months, Less than 10 hrs/week - Posted
Overview: SEO Quotient (www.SEOQ.com) is a website that seeks to answer the big question ... why does his website rank higher than mine? Requirements: To do this, I need to crawl a website looking for specific information and then bring that back and present in a meaningful way. I would be looking for things like keywords in the title tag, what CMS is being used, how fast or slow is the site, does it pass Google's mobile friendly test, etc. Skills: The front-end of the system was built in WordPress, PHP, Bootstrap, HTML, CSS and the back-end of the system was built in Python, Django, PHP, AngularJS, etc. Some folks who have done work like this in the past thing that PERL is the method to use. I'm sure there are plenty of ways to crawl and get the data but, ultimately, it needs to live in our Python/Django/MySQL back-end. Probably, you will also need to know Git to help contribute to the repository. Knowing how to get data to/from APIs and perhaps how to build APIs...