Data Scraping Jobs

435 were found based on your criteria {{|number:0 }} were found based on your criteria

show all
  • Hourly ({{ jobTypeController.getFacetCount("0")|number:0}})
  • Fixed Price ({{ jobTypeController.getFacetCount("1")|number:0}})
Hourly - Expert ($$$) - Est. Time: Less than 1 month, 10-30 hrs/week - Posted
In summary: I want to be able to configure scrapy for multiple locations via a simple website. I want scrapy to grab a session token, spoof the IP, grab my data and save the CSV to an S3 bucket. I want to be able to: 1) login to my own secure website hosted in AWS 2) display simple 4 column form with column names (see attachment) 3) Setup new scrapes 4) refresh recurring scrapes 3) in detail For setting up New Scrapes: "Get New DataSource" launches new tab or similar (e.g., Chrome extension?) wherein I login into my new datasource and then navigate to the area that I want to scrape, specify the table and somehow specify "Get Data". It should be able to handle easier REST url requests or more difficult ones with an obscured header variables). While I'm open to variation, I'm envisioning something similar to the pinterest chrome extension but with regards to data tables within secure websites. Once, the scrape configuration is saved, then it starts 4) get data "refresh" 4) in detail click "REFRESH" spawns new tab wherein user only logs in. Session token is grabbed by service. All requested data is navigated to and pulled on the back end. Note: some IP spoofing on the login or on the backend service will be required. 5) back end service should exist as AWS Lambda callable code. As such, variables should reside separately and load per request. 6) I anticipate using this with a node.js service ... so, looking for callable compliance (i.e., I know that scrapy is natively python) 7) data should be saved consistently/statically to a dedicated S3 bucket (per logged in user) ... authenticated URL can be made available. Finally, I'm okay with pulling in Scrapy and AWS libraries. I do want to minimize code complexity beyond that am looking for clean, well documented, quick code.
Skills: Data scraping Scrapy Web Crawler
Fixed-Price - Intermediate ($$) - Est. Budget: $150 - Posted
I need to scrape Google Trends data but it gives me a Quota limit when I use this API: I need to bypass the quota limit and I don't know how. I need to scrape it for a list of 150k words. Only bid after reading and knowing you are confident about being able to do the work.
Skills: Data scraping API Development Python
Fixed-Price - Entry Level ($) - Est. Budget: $100 - Posted
We are looking to: • Record a spreadsheet of all 301 business brokers in California, based on directory/ website at (click "Submit" with no filters). We will want to scrape the main listing, as well as the individual broker page (name, business name, email, address and phone). • For brokers with no email address on the profile page, to research and obtain these addresses (may be able to get through the hyperlink on website "contact broker") • Find Upworthy candidates to give large future similar assignments to, ideally up to 40/hrs per week. We are looking for: • Top web researchers, web scrapers and data entry specialists seeking longer-term assignments in CRM data entry. • Reliable, hardworking, and detail-oriented. If you are interested in reliable, profitable and consistent long-term work, we want to work with you! Feel free to ask any questions, and we look forward to working together.
Skills: Data scraping Data Entry Internet research
Fixed-Price - Intermediate ($$) - Est. Budget: $45 - Posted
I need an Excel list created of all the solo and two attorney firms who are members of the Boston Bar (I currently use - By solo and two attorney firms, I mean they is only 1 or 2 attorneys working at the office. I would like them in these categories: Name, Firm Name, address, email, phone and website (if provided). If you can determine if they are a solo or two attorney firm that would be great because that's who I'm targeting. Each will need to be a separate column header on the spreadsheet. This ensures that I can easily filter the data.
Skills: Data scraping Data mining Data Recovery Data Science