You've landed at the right place. oDesk is now Upwork. Learn about the new platform.

Data Scraping Jobs

300 were found based on your criteria {{ paging.total | number:0 }} were found based on your criteria

show all
  • Hourly ({{ jobTypeController.getFacetCount("hourly") | number:0}})
  • Fixed Price ({{ jobTypeController.getFacetCount("fixed") | number:0}})
show all
only
only
only
show all
only
only
only
only
only
show all
only
only
only
Looking for the Team App?
Download the New Upwork Team App
Fixed-Price - Intermediate ($$) - Est. Budget: $50 - Posted
Hi I need a freelancer who can extract leads information from the group which I will provide to you. Group have a 30,000 Members and If you know the data mining and Data Extraction than its easy for you. LinkedIn group only show the 1500 members and we require 10000 leads from 30,000 for now. if you have a trick than it can be easy for you. I need it to be done ASAP Thanks
Skills: Data scraping Data Entry Data mining
Fixed-Price - Expert ($$$) - Est. Budget: $500 - Posted
We have a list of 1000 websites and we need a data-building expert to find contacts from that company. The requirements are 1. Find contact name, email address, phone number, designation and website monthly traffic from SimilarWeb and/or ClearWebStats 2. Contacts should be from Digital/Online Marketing, Ecommerce, Product Management, UI/UX, etc 3. We need minimum 3 contacts from each company of manager and above designation If interested, please respond with the following details: 1. Relevant experience 2. Estimated time to complete 1000 websites 3. Number of people in your team who'll work on the project 4. How soon can you start? 5. Data accuracy percentage
Skills: Data scraping Data Entry Data mining Internet research
Hourly - Entry Level ($) - Est. Time: 3 to 6 months, Less than 10 hrs/week - Posted
We need to insert ads in our portal by scraping them from internet and insert in our import tool. We use import.io but other tools JSON compatible to our import tool are accepted. Data entry should take care about English text and focused on keywords. A job on customer service and assistant for administrative task can be available at the end of this task
Skills: Data scraping Data Entry json
Hourly - Entry Level ($) - Est. Time: Less than 1 month, Less than 10 hrs/week - Posted
We have recently won a grant to develop our new product, Pikhaya, which uses open data to provide free market intelligence to help entrepreneurs find viable business premises. There are 350 local authorities in England and Wales, of which 70 currently publish compliant data we need to include in our model. We will need, on a quarterly basis, to extract data from their websites, transform it into a common, machine-readable (CSV) format, and then upload it into our database. This is the first time we are running this process and we are aiming to hire three different freelancers to do this. After the first round of ETL, the most accurate freelancer would then work with us on a regular basis to produce the data we require. Scope of work: 70 current local authorities, most with only one dataset (usually Excel or CSV) to download and transform. An example includes Chorley (http://chorley.gov.uk/Pages/AtoZ/Information.aspx). We are specifically interested in the empty business premises which is not always available (sometimes only ‘all’, sometimes only ‘occupied’). On the example page, you’ll need to click on Freedom of Information and then the file-name: ‘Chorley - Ratepayer account data February 16.csv’ You will need to be comfortable with transforming messy, inconsistently-structured tabular data into standardised machine-readable format, and with a high level of data-accuracy. We’ll provide you with the details in Google Sheets for each local authority site where you need to do a download. We are also in the process of filing Freedom of Information requests for the other 280 local authorities (so we may have more shortly). Please respond with: 1 The time you think it will take you (and we’d like to have it complete by early March); 2 Examples of work which reflect the approach you would take with our brief (i.e. something relatively similar); 3 How you will ensure data quality; 4 An estimated price for the full job (although you can bill by the hour).
  • Number of freelancers needed: 3
Skills: Data scraping Data Entry Data mining Google Spreadsheets
Fixed-Price - Intermediate ($$) - Est. Budget: $800 - Posted
List of URLs: lakorn.guchill.com www.seriesubthai.tv www.kodhit.mobi cuptv.com www.startclip.com cn.upyim.com www.jengmak.com www.songdee.com th.hao123.com www.friv.com newsupdate.todayza.com diply.com tvshow.guchill.com www.subthaiseries.com www.tunwalai.com www.yumzap.com www2.adintrend.com devian.tubemate.home lakorn.guchill.com www.kodhit.mobi cuptv.com www.jengmak.com Step 1 Review the list of URL’s above, mine the top 1000 pages from that URL by using a crawler. Step 2 Extract terms from each of the these web pages that are mined Step 3 Determine if the terms from these web pages, match terms from the entertainment list or match elements from the news list. (Fill in the attached worksheet) Step 4 For each site create a new worksheet and repeat process
Skills: Data scraping Data Analytics Machine learning Web scraping
Fixed-Price - Expert ($$$) - Est. Budget: $50 - Posted
I'm going to Angola, a Portuguese speaking country, and need someone to conduct some research (within financial services/ banking and telecommunications) on this country, some important individuals and other aspects of these industries. A lot of this information will be in Portuguese and I need someone with excellent research skills - web research/ deep mining - and native (or near native) Portuguese: written, reading and speaking. Ideally the individual will have lived or live in Angola - though this is not completely necessary. I'll provide you with the list of names (approx. 10) , companies/ societies (approx. 10) and specific details on the aforementioned industries. I'd need this work completed by: 16th February. It should be presented in a table either using MS Word or Excel. All to be written in English. Please ask any needed questions. Price given is negotiable depending on skill, expertise, and speed of delivery.
Skills: Data scraping Internet research Market research Portuguese
Hourly - Entry Level ($) - Est. Time: Less than 1 week, Less than 10 hrs/week - Posted
I need an internet scrape done, with all of a specific type of business. I would need the following pieces of inofrmation on each business: 1. Name 2. Location (Address with Zip Code) 3. Industry 4. Contact Phone Number (if possible) 5. Contact Email (if possible) 6. Graphic (if possible) I then need all the information put into excel format. I will make hiring decision on Feb. 12, 2016 Please answer the questions below in your proposal.
Skills: Data scraping Data mining Internet research Microsoft Excel
Fixed-Price - Expert ($$$) - Est. Budget: $100 - Posted
I would like to scrape data from the following website: https://www.courts.mo.gov/casenet/cases/searchCases.do?searchType=caseNumber The website requires you to enter a case number and then provides data on the case. I have a list of 1,048,575 case numbers that I would like data on. For an example, enter: 0516-CV05476. For a given case there are multiple tabs. I would like to scrape only the data from the "Case Header" tab and the "Civil Judgments" tab (when available). Output should be in tab separated format (TSV) with UTF-8 character encoding.
Skills: Data scraping Web scraping
Fixed-Price - Expert ($$$) - Est. Budget: $250 - Posted
I would like to scrape data from this USPTO website: http://portal.uspto.gov/pair/PublicPair It has a CAPTCHA, but only seems to require it to be filled out once per session. (I haven't verified how many queries it takes to trigger it to reappear). After the CAPTCHA you can look up a patent application by "Publication Number." I have a list of 3,996,534 publication numbers. An example of a publication number would be: 20120257316. For each publication number, I want to scrape the basic bibliographic data that shows up in the "Application Data" tab. I want the scraped data in tab separated values (tsv) format with UTF-8 character encoding. On the USPTO website, when you look up a publication number, there is also usually a tab called "Image File Wrapper", which contains pdf files associated with the patent application. For each publication number that has the "Image File Wrapper" tab present, I want to download the pdf files with the document code "OATH" and "ADS." Note, even when the Image File Wrapper tab is present, both of these pdfs may not be present. In particular the ADS file is often missing. Thus the project has two deliverables: - A TSV file with one row for each publication number and one column for each piece of bibliographic data in the application data sheet. - A folder with one directory for each publication number (the directories should be named by publication number) containing 0-2 pdf files corresponding to the OATH and ADS documents when present.
Skills: Data scraping Web scraping
Looking for the Team App?
Download the New Upwork Team App
Fixed Price Budget - ${{ job.amount.amount | number:0 }} to ${{ job.maxAmount.amount | number:0 }} Fixed-Price - Est. Budget: ${{ job.amount.amount | number:0 }} Open to Suggestion Hourly - Est. Time: {{ [job.duration, job.engagement].join(', ') }} - Posted
Skills: {{ skill.prettyName }}
Looking for the Team App?
Download the New Upwork Team App