Data Mining Jobs

483 were found based on your criteria {{ paging.total|number:0 }} were found based on your criteria

show all
  • Hourly ({{ jobTypeController.getFacetCount("0")|number:0}})
  • Fixed Price ({{ jobTypeController.getFacetCount("1")|number:0}})
Fixed-Price - Intermediate ($$) - Est. Budget: $15 - Posted
IF YOU ARE A WIZ WITH MACROS OR EXCEL DATA COMPARISONS USING AUTOMATIC SEARCH FEATURES... YOU'LL BE PERFECT FOR THIS We have 2 excel task and we need someone who can do the work within a couple hours. Task 1: We have one excel file with 270,000 records in it and another with about 12,000 records in it. We need to find the 12,000 in the file with the 270,000 and highlight them blue. The data it is searching in the 12,000 excel file has records in about 10 columns that needs to be compared to one single column in the big 270k file. Output will be an excel file with 270k records with 12,000 records highlighted blue to denote which came from the smaller list. Task 2: - We have an excel sheet that contains about 22,000 people - We need to recategorise the people in the sheet by a number of "new" list names and create a different csv for each of the categories (about 10 categories) - the 1st row must use specific field names (e.g. FIRST_NAME) so that it can be uploaded to our new system easily
Skills: Data mining Data Entry Data scraping Database Modeling
Hourly - Expert ($$$) - Est. Time: Less than 1 week, Less than 10 hrs/week - Posted
This job is for a 30 minute Skype call to give my development team some tips to improve there scraping. The devs are scraping high volume data from Google Adwords and Search Results. They are using proxies and other methods but are still having issues scaling this. The scraping is a repetitive activity that happens every day. Devs are programming in Python. You will be perfect for this job if you have experience with very high volume data scraping from Google and know how to work around there system. You are bidding on a 30 minute phone call.
Skills: Data mining Data scraping Python Web Crawler
Hourly - Expert ($$$) - Est. Time: 3 to 6 months, 30+ hrs/week - Posted
Data entry clerks work from source documents to enter information into spreadsheets or computer programs. They quickly and accurately type the data according to the data entry instructions. Clerks may need to insert new records before entering data, delete existing records or type over existing data to update data in the computer. A data entry clerk might need to cross-reference data from multiple source documents when entering data for a single record.
Skills: Data mining Data Backup Data Entry
Fixed-Price - Entry Level ($) - Est. Budget: $70 - Posted
I have 674 URLs (42(A) + 632(B)), each with tables in the same format. I need the data to be scraped into a single Excel file, with a tab for each page to be scraped. Note (see in preview links below): I need any hyperlinks to be included, and under the "elected" column, where there is a checkbox (images/check.gif), I need that to be included as a character (e.g. "X"). Here's a sample from the A-List: http://www.lop.parl.gc.ca/About/Parliament/FederalRidingsHistory/hfer.asp?Language=E&Search=Gres&genElection=42&ridProvince=0&submit1=Search And here's a sample from the B-List: http://www.lop.parl.gc.ca/About/Parliament/FederalRidingsHistory/hfer.asp?Language=E&Search=Bres&ridProvince=0&genElection=0&byElection=2012%2F03%2F19&submit1=Search (They *should* both be the same format - but I am not the expert) I will provide, in Excel or CSV format: -A sample output (Pretty straightforward tables as I will be running the post-processing) -A list of the URLs to be scraped -A name for each of the scraped tables Look forward to hearing from you if this will be possible!
Skills: Data mining Microsoft Excel
Fixed-Price - Expert ($$$) - Est. Budget: $50 - Posted
I am looking for someone with strong data mining skills as well as familiarity with soil and ecological government agencies. Specifically I will need you to gather data points according to the the attached spread sheet as well as a list of sources. Functional areas of the data including year (last three years), state (Minnesota, Iowa, Illinois, Indiana), County, and all requested data points. The attached spreadsheet is about 80% of the way done, and will need the following things done: - There are 380 counties across all 4 states, however this spread sheet does not have a complete list of all counties for each of the years (2015, 2014, 2013). Thus I will need you to ensure that there is are 380 data points for all three years for both corn and soy bean acres. Please use the list of sources at the Top of the spread sheet. - Please do not send me half-done edits. The product you send me should be the best work you can do. Do not hesitate to ask me any questions along the way. - I will need this by ASAP, and will throw in a $20 bonus if you can get it done in 24 hours. I would like to schedule a phone call to make absolutely sure that we are on the same page with all expectations.
Skills: Data mining Data scraping Microsoft Excel
Hourly - Entry Level ($) - Est. Time: Less than 1 week, 10-30 hrs/week - Posted
I am in need of an individual who can pull down every single Microsoft Excel template that is located within the Excel program, under "New." 1. Open Microsoft Excel 2. Click "New" 3. Under the Office.com templates section, click under every single folder (starting with analysis) and download every template 4. Place all templates under each folder that it was sourced 5. Zip the file and send it over I am not certain about versions, but I would like every template from Excel 2010 to Excel 2016 if possible.
Skills: Data mining Microsoft Excel
Hourly - Entry Level ($) - Est. Time: More than 6 months, 10-30 hrs/week - Posted
Hey guys, This is a simple task which is outlined below: - I need an automated way for someone to get email address's for me. - Using the company name I have provided, please find the CEO or top decision maker of the person and find their email. - put their email in a spreadsheet with first name, last name, and title This is a recurring bulk job where you will be given 1k emails at a time and expected to turn around in less then 48 hours. I'm not looking for this to be done manual - I have a very large team that does that already. I need someone who can build a tool that can do this in an automated fashion. This is a great job and you will have new work every time you complete the project. I am looking to develop a long-term relationship. I need someone to build the tool, and provide the service. I am extremely focused on results here. I don't expect you to find 100% of the names but do expect verified quality. I'm currently paying $0.11 per each name returned (from my manual team) so you need to be able to beat that rate. Depending on how accurate and fast you work there is a huge opportunity to make money. All payments will happen thru the bonus feature. When I award the project I will use $1 as a placeholder so we can have an open contract. please write "game boy" in the cover letter so I know you read this. --- Skills: linkedin, marketing, email-support, research
Skills: Data mining Email Handling Email Marketing Internet research
Hourly - Expert ($$$) - Est. Time: More than 6 months, 30+ hrs/week - Posted
I'll need 2-3 contacts for each of the first 200 schools listed on this site: www.4icu.org/ph For each contact, I'll only need the following information on a Google Spreadsheet: a. First Name & Last Name b. Current Title (I'll provide you with the actual targets upon hiring) c. Verified Email Address This is a very easy & quick project as all of the info can be found on each of the school's website. Also, I do not need a lot of data, but I will strictly require correct & accurate information: Spelling of First & Last Name, Current Title and most importantly the verified email address for each contact (which is also on the schools' websites) Note: Budget noted is excluding Upwork fees so don't worry, you'll get the whole amount (plus possibly a bonus if you do a really awesome work!)
Skills: Data mining Data Entry Google Docs