You've landed at the right place. oDesk is now Upwork. Learn about the new platform.

Data Scraping Jobs

254 were found based on your criteria {{ paging.total | number:0 }} were found based on your criteria

show all
  • Hourly ({{ jobTypeController.getFacetCount("hourly") | number:0}})
  • Fixed Price ({{ jobTypeController.getFacetCount("fixed") | number:0}})
show all
only
only
only
show all
only
only
only
only
only
show all
only
only
only
Hourly - Est. Time: Less than 1 month, 30+ hrs/week - Posted
We need web scraping of one specific website into CSV. We require the: Title Price Category Image Description (if applicable) Scrapped into columns of a CSV. Happy for you to use a combination of manual and automatic scrapping. Images need to be renamed and saved locally in one folder and referenced from the CSV by file/image name. I will give you the site details on application.
Fixed-Price - Est. Budget: $ 500 Posted
Have a series of sites I want to scrape and automate. Need a series of scripts done: 1. Populate (scrape) Basically take the information on the site and return it in a pre-defined format. 2. Register(profile) Automate the process of registering on the site 3. Login(profile) Automate the process of logging in 4. Apply(url, profile) Automate the process, given a url for a posting, to apply using a particular profile. 5. Update(profile) Using a particular profile. Update the information held on the site. Primarily uploading a particular file(provided within the profile) to a certain place within the site. Detailed doc available here. https://docs.google.com/document/d/1Tafuxp08oz7gXe4DCmvyfn-cohpHADCCO2LP2HYMljw/edit?usp=sharing
Fixed-Price - Est. Budget: $ 20 Posted
OVERVIEW: We will give you email lists that we would like cleaned, so we don't have bouncing emails when we send. We want to know how many emails you can clean for $20 (Think of this as your first potential contract of many). NOTE: EVEN IF YOU ARE SUPERMAN, and the most amazing email cleaner in the planet, TOO MUCH MONEY IS STILL TOO MUCH MONEY. So... Be smart and not greedy :) TO APPLY PLEASE ANSWER: A) How many email files can you clean for $20? B) What program will you be using to clean the emails? C) How long will it take you to clean them?
Fixed-Price - Est. Budget: $ 40 Posted
OVERVIEW: We would like to purchase $40 worth of USA Business Email Data Lists (You will be bidding on how many files you can sell for $40) REQUIREMENTS: -No .edu emails -No .gov emails -No non-USA emails -Must Contain at least: Email Address, Company Name, Phone Number, City & State (We prefer more, but this is bare minimum) -Must guarantee replacement of any bounced/bad emails TO APPLY PLEASE ANSWER THE FOLLOWING: A) How many email files will we be receiving for $40? B) Can we choose which states and/or industries we would like? And if so, which can we choose (If not both). C) Explain how your list was collected? Be specific (For instance: If it was scraped, where was it scraped? If it was collected from a website, which website?) D) When was this list collected? (We'll know if you're lying when we test the bounce rate, so just be honest) E) Roughly how many % of these emails will likely bounce? (Again, we will be testing this, just be honest) F)...
Hourly - Est. Time: More than 6 months, 30+ hrs/week - Posted
We only want to hear from people that are aspiring for greatness. If this isn't you, please don't read this advertisement, and don't apply for this job. Working with us, you will be in the right environment for you to meet and exceed your full potential. This job is to work alongside one of our most dedicated crew members who has an amazing work ethic and is a fantastic leader. It is an opportunity to learn from one of our best. Your new career with us will involve working full-time hours for us during Australian Eastern Standard Time (AEST) standard workday hours of 9:00am to 5:30pm. Candidates located in the Philippines would be ideal. You will need to be on the ball and responsive at all times. This schedule is in place to remain in sync with our team here in Australia and your team leader who is based in Sydney. In order to be suitable for this role, you will need: - Excellent written and verbal communication skills (and reply to an email to demonstrate this before...
Hourly - Est. Time: Less than 1 week, 10-30 hrs/week - Posted
We would love your help to better understand our users using publicly available information. Specifically, our database of active users includes first name, last name, email address, and location, but does not include professional (job) information. We would like your help to identify each user’s company, role, department/function, and LinkedIn profile link (URL). First, we will need you to sign our Non-Disclosure Agreement (NDA), Consulting Agreement, and SOW. Although this does seem like an unnecessarily number of documents, it is important to do so in order to protect the privacy of our users and uphold our Privacy Policy. Our current policy only allows us to share user information with contractors who have completed these documents. After the necessary documentation is completed, we will provide a spreadsheet (Excel file, .xlsx) file with the first name, last name, email address, and location (city, state, country) for about 10,000 users. In the Excel file will be a...
Fixed-Price - Est. Budget: $ 100 Posted
We are looking for a web scraping script that searches through all articles on a number of websites for certain keywords and outputs the article’s entire contents, the frequency of the keywords, and other metadata. We are also looking for a script that compiles the 50 most frequent words in those articles by month. More specific details are provided below. Output • CSV with one row for each article and columns for the following features (see “article summary” tab in attachment for template): o Date o Website o Article title o URL o Location of website headquarters o Article contents o Frequency of keyword 1 in article body o Presence of keyword 1 in article title (true/false or 1/0) o Repeat frequency and presence measures for other keywords • CSV with one row for each of the top 50 most frequent words and columns for the following features (see “top 50 monthly” tab in attachment for template): o Date (month-year) o Keyword o Frequency • Web scraping script(s) in Python...
Hourly - Est. Time: Less than 1 week, 10-30 hrs/week - Posted
Hello, I am looking for an experience data miner and researcher to search LinkedIn and return everyone with the title of CFO / VP Finance / Director of Finance / Controller in the Greater Toronto Area - Canada. The list will likely be extensive - looking for around 5000 names. I can provide the best groups to search and discuss more details with the person chosen. Thanks.
Fixed-Price - Est. Budget: $ 250 Posted
I need a multi-threaded web spider that can work similar to Google crawlers. It will need to crawl the domain to find all external domains and internal links. It will follow internal to find more pages and save all external domains to a separate database. Found domains need to be DNS/WhoIs checked for availability. - I need to be able to easily add domains for crawling through simple form or a text file import - It needs to have a dashboard where I can see live stats on how many crawlers are active (and change the number), how many domains is waiting in, how many is being crawled, how many is finished, number of pages crawled by domain, number of found external domains, number of free/live, server usage/load etc. - I need to be able to change crawler timeouts - I need to be able to pause/continue/stop/remove each 'project' - I need to be able to set how many levels deep crawler should go within the site structure - Adding new domains, exporting results, etc. should not affect web...
Fixed-Price - Est. Budget: $ {{ job.amount.amount | number:0 }} Open to Suggestion Hourly - Est. Time: {{ [job.duration, job.engagement].join(', ') }} - Posted
{{ job.description }}