Data Scraping Jobs

361 were found based on your criteria {{ paging.total|number:0 }} were found based on your criteria

show all
  • Hourly ({{ jobTypeController.getFacetCount("0")|number:0}})
  • Fixed Price ({{ jobTypeController.getFacetCount("1")|number:0}})
Fixed-Price - Intermediate ($$) - Est. Budget: $700 - Posted
I've attached the project overview in Word format (link also here: https://drive.google.com/file/d/0B8ZJVPz-oSKjZ0pNS1paR2Nrc1E/view?usp=sharing), and also included additional SEC FTP resources in it which might be useful. Objective Download 10k of every publicly traded US company dating back to 1994 in PDF format. Overview The SEC makes annual reports (i.e., 10k) available via their website (Microsoft example: www.sec.gov/cgi-bin/browse-edgar?action=getcompany&CIK=0000789019&type=10-k&dateb=&owner=exclude&count=40), and also via FTP for free. When an annual report is posted each year, it is now done so as an HTML file, with a few different “exhibits” (i.e., other documents of relevance). I am mostly concerned with the 10k file itself, but would prefer having both the 10k and exhibits for each year. For example, this is Microsoft’s 2014 annual report from the SEC’s website: i.imgur.com/aBeGa6F.png I would then like each annual report to be in a single PDF file, going back to 1994, in the following format: Symbol Year Company Name.pdf. For example, Microsoft would look like: MSFT 2015 Microsoft.pdf MSFT 2014 Microsoft.pdf MSFT 2013 Microsoft.pdf etc. The links in each PDF file will have to behave as they do on the SEC’s website, as found in this example PDF: https://drive.google.com/file/d/0B8ZJVPz-oSKjaS1TbmNOZ0dlWlU/view?usp=sharing The script will then check for new annual reports (and companies), and if it finds one, download it, convert it to PDF, and name it per the naming convention.
Skills: Data scraping FTP JavaScript Python
Hourly - Intermediate ($$) - Est. Time: Less than 1 week, 10-30 hrs/week - Posted
looking to get a large list of USA real estate emails, name, and phone number - I need to know: -what is your method of building this list and how can I know that they're accurate and real email addresses -how many total contacts you believe you can deliver -how many hours will it take you to build this list
Skills: Data scraping Data mining Web scraping
Hourly - Entry Level ($) - Est. Time: 1 to 3 months, Less than 10 hrs/week - Posted
Looking to populate a database full of user names. Will put together a video showing how to do this task, its pretty straight forward data entry but can be very fun. Lead entry will be one task but also have many more simple data / entry and web based tasks to complete. Looking for the right person to develop a long term relationship who can get into the habit of doing a daily routine of set tasks.
Skills: Data scraping Data Entry Internet research
Fixed-Price - Entry Level ($) - Est. Budget: $20 - Posted
Currently I have a program that downloads data from a series of websites and pastes the data into excel sheets that I then use. I have found another website that I would like the information from added to my program. There are 2 different pages I would like data from on this website. The website has an export button that automatically pastes the data in an excel sheet. What I am looking for is a programmer to add this website to the list of websites that I already have, open it, export the data to an excel sheet, and copy the excel sheet to a different excel sheet I have made. This would need to be done for both web pages I would like the data from.
Skills: Data scraping Microsoft Excel
Fixed-Price - Expert ($$$) - Est. Budget: $110 - Posted
I am looking for a web scraping professional to create a script that crawls through an entire e-commerce site. The majority of this site has no internal links on its pages. I need someone who is an expert at crawling sites, aggregating data, and then storing them either as a CSV file or in MySQL. You will also need to be able to incorporate rotating proxies to lessen the chance of an IP address getting banned. The crawling of an individual page is quite straightforward. I was able to extract the desired information using a short Beautiful Soup script. But since nearly all pages are standalone product landing pages with no internal (or external) links, my limited abilities were not able to crawl the entire site Required Experience: - Comfortable using Selenium, Phantom JS, Beautiful Soup (if you prefer other technologies and feel they will create a better end product, then I would be willing to further discuss the matter) - Smart scraping behavior such as adjusting HTTP headers of the scraper to look like it is coming from a browser. - Ability to store data in a CSV file or in MySQL - Add a rotating proxy service to the script to protect the IP Address from getting banned Attributes that would make you a good fit for this project: - Able to communicate effectively: I need to verbally describe my ideas to the freelancer, so a good command of English is important - Ability to successfully meet deadlines - Good at giving development updates. To Apply: - Send a cover letter that includes an overview of how you intend to accomplish the job - A portfolio of your work. - At least 2 references I genuinely look forward to hearing from you!
Skills: Data scraping Python Selenium Web scraping
Fixed-Price - Intermediate ($$) - Est. Budget: $500 - Posted
We're looking for a person(s) to provide us scrapes of web data using provided links. These links are in the thousands and dynamic. The person must be fluent in English and have basic technical/development skills. Having access to a large network of bots is a requirement to complete this task.
Skills: Data scraping HTML Web scraping
Fixed-Price - Intermediate ($$) - Est. Budget: $20 - Posted
I need Python script that will login to a password protected area (login will be provided) and downloads all the videos (the mp4 links are accessible through the source code of the page), and PDFs and organize them in folders. It would also be nice to combine all the videos as one large file as well after downloading them all.
Skills: Data scraping HTML Python