Hadoop Jobs

81 were found based on your criteria {{ paging.total|number:0 }} were found based on your criteria

show all
  • Hourly ({{ jobTypeController.getFacetCount("0")|number:0}})
  • Fixed Price ({{ jobTypeController.getFacetCount("1")|number:0}})
Hourly - Entry Level ($) - Est. Time: More than 6 months, 30+ hrs/week - Posted
Python, Java, C++ or any other OOP language) to perform daily assignments •Utilizes Hadoop or other Map Reduce paradigms or Big data, machine learning, statistical modeling and data mining to develop recommendations or provide solutions to requests/issues •May work with external consultants/vendors/research entities •Creates statistical models and/or conducts analysis using appropriate methodologies
Hourly - Entry Level ($) - Est. Time: More than 6 months, 30+ hrs/week - Posted
Required Skills: 3+ years of experience with statistics, data models, data mining, and data analysis. 1+ year of data science and machine learning experience. 1+ year of experience utilizing a big data platform such as SAS, Hadoop, Cassandra, Spark, etc. Master's in Statistics, Math, Computer Science, Economics/Econometrics, Physics, or related.
Hourly - Expert ($$$) - Est. Time: Less than 1 month, 10-30 hrs/week - Posted
Hi, I am looking for a developer with experience using the common crawl dataset. Ideally it would be ran as EC2 instances, so bandwidth isn't an issue, only computation time/cost. Knowledge of additional cost saving measures such as spot instances would be an advantage. I require a script to parse through the metadata (WAT) files and extract data quickly and efficently. The data I am interested in is the number of times a domain references an external image or javascript file. For example: "WARC-Target-URI": "http://internaldomain.com/some-page/" "Scripts": [{​ "path": "SCRIPT@/src", "type": "text/javascript", "url": "http://externaldomain.com/script.js" }​ "Links": [{​ "alt": "Alt Text", "path": "IMG@/src", "url": "http://externaldomain.com/image.jpg" }​ I would like to know how many times externaldomain.com has been referenced throughout the commoncrawl corpus by JS and Images. A more detail spec can be provided after some more discussion. As this is just a very basic overview of what I would like created. References & Examples: http://commoncrawl.org/the-data/get-started/ https://commoncrawl.s3.amazonaws.com/crawl-data/CC-MAIN-2016-26/segments/1466783391519.0/wat/CC-MAIN-20160624154951-00000-ip-10-164-35-72.ec2.internal.warc.wat.gz
Skills: Hadoop Amazon EC2 Data scraping Java
Hourly - Entry Level ($) - Est. Time: More than 6 months, Less than 10 hrs/week - Posted
., is a technology based company that provides IT Services to companies our solutions include BI & Analytics solutions on Qlikview, Microsoft BI and Big Data using Hadoop. We are looking for a website designer to make some fixes and re-designing our existing website.
Skills: Google AdWords Development SEO Backlinking SEO Writing WordPress
Fixed-Price - Intermediate ($$) - Est. Budget: $2,000 - Posted
Qualifications 9+ years of experience developing large scale software/ web applications Experience in Big Data and cloud computing leveraging Hadoop ecosystem ((Hadoop/HDFS and related technologies) Proven prior experience in building large scalable distributed systems At least 2 Big Data implementations Experience in Java and an expert of all Hadoop components Hands on experience with major components like Map Reduce, HBase, Hive, Pig Strong understanding of underlying Hadoop concepts and distributed computing Solid understanding of current status of big-data technology including hands on experience with design, and building big-data infrastructures. ... Qualifications 9+ years of experience developing large scale software/ web applications Experience in Big Data and cloud computing leveraging Hadoop ecosystem ((Hadoop/HDFS and related technologies) Proven prior experience in building large scalable distributed systems At least 2 Big Data implementations Experience in Java and an expert of all Hadoop components Hands on experience with major components like Map Reduce, HBase, Hive, Pig Strong understanding of underlying Hadoop concepts and distributed computing Solid understanding of current status of big-data technology including hands on experience with design, and building big-data infrastructures. ... Qualifications 9+ years of experience developing large scale software/ web applications Experience in Big Data and cloud computing leveraging Hadoop ecosystem ((Hadoop/HDFS and related technologies) Proven prior experience in building large scalable distributed systems At least 2 Big Data implementations Experience in Java and an expert of all Hadoop components Hands on experience with major components like Map Reduce, HBase, Hive, Pig Strong understanding of underlying Hadoop concepts and distributed computing Solid understanding of current status of big-data technology including hands on experience with design, and building big-data infrastructures.
Skills: Hadoop Agile software development Big Data C++
Fixed-Price - Expert ($$$) - Est. Budget: $160 - Posted
We have tie up with so many engineering colleges in pune, we have a plan to conduct training courses in respective college premises as per their requirements, for that we required freelance technical trainers for JAVA, .NET, ANDROID & HADOOP. Roles and Responsibilities: 1. Should have hands on experience in application development and passionate towards sharing knowledge. 2.
Skills: Hadoop .NET Remoting Android Java
Hourly - Intermediate ($$) - Est. Time: 1 to 3 months, 10-30 hrs/week - Posted
Looking for an experienced Hadoop systems Architect/Administrator (Cloudera ONLY). I have two positions (Suitable candidate can assume both the roles) 1) Systems Architect who can advice on the areas to improve with respect to Automation, deployment, performance tuning, capacity management etc, Document with the steps, Diagrams (Please apply if you only have experience large deployments) — This would be hourly job 2) Hadoop Administrator with experience with Troubleshooting of various ecosystem tools, JVM, setting up monitoring like Ganglia, Nagios, Automation experience (Shell and Python) etc here is the detailed Job description 1) Big data System architecture/Administration (Cloudera Hadoop, Elasticsearch, MongoDB) 2) Cloudera Administration, Cloudera Manager API Preferably Cloudera certified 3) In-depth knowledge of Security practice on Cloudera (Kerberos, KMS, Cloudera Navigator, Sentry ) 4) expert in troubleshooting (ecosystem tools, JVM, Hive/Impala Query Tuning ) 5) Solid Scripting in Python, Shell (Proof needed: Github) 6) Someone who has experience with Monitoring setup (Ganglia, Nagios) complimenting existing Cloudera Manager Solid Linux Admin Skills Work ... I have two positions (Suitable candidate can assume both the roles) 1) Systems Architect who can advice on the areas to improve with respect to Automation, deployment, performance tuning, capacity management etc, Document with the steps, Diagrams (Please apply if you only have experience large deployments) — This would be hourly job 2) Hadoop Administrator with experience with Troubleshooting of various ecosystem tools, JVM, setting up monitoring like Ganglia, Nagios, Automation experience (Shell and Python) etc here is the detailed Job description 1) Big data System architecture/Administration (Cloudera Hadoop, Elasticsearch, MongoDB) 2) Cloudera Administration, Cloudera Manager API Preferably Cloudera certified 3) In-depth knowledge of Security practice on Cloudera (Kerberos, KMS, Cloudera Navigator, Sentry ) 4) expert in troubleshooting (ecosystem tools, JVM, Hive/Impala Query Tuning ) 5) Solid Scripting in Python, Shell (Proof needed: Github) 6) Someone who has experience with Monitoring setup (Ganglia, Nagios) complimenting existing Cloudera Manager Solid Linux Admin Skills Work ... I have two positions (Suitable candidate can assume both the roles) 1) Systems Architect who can advice on the areas to improve with respect to Automation, deployment, performance tuning, capacity management etc, Document with the steps, Diagrams (Please apply if you only have experience large deployments) — This would be hourly job 2) Hadoop Administrator with experience with Troubleshooting of various ecosystem tools, JVM, setting up monitoring like Ganglia, Nagios, Automation experience (Shell and Python) etc here is the detailed Job description 1) Big data System architecture/Administration (Cloudera Hadoop, Elasticsearch, MongoDB) 2) Cloudera Administration, Cloudera Manager API Preferably Cloudera certified 3) In-depth knowledge of Security practice on Cloudera (Kerberos, KMS, Cloudera Navigator, Sentry ) 4) expert in troubleshooting (ecosystem tools, JVM, Hive/Impala Query Tuning ) 5) Solid Scripting in Python, Shell (Proof needed: Github) 6) Someone who has experience with Monitoring setup (Ganglia, Nagios) complimenting existing Cloudera Manager Solid Linux Admin Skills Work
Skills: Hadoop Ansible Bash shell scripting Cloudera
Hourly - Intermediate ($$) - Est. Time: More than 6 months, 30+ hrs/week - Posted
----- No agency please as we would be dealing only with quality freelancers----- ----- Preference would be given to applicants from Indian sub-continent because of location constraints, but if you are from Pakistan please dont apply, as due to legal issues we wont be able to hire you even if you clear interviews, so save your job credits---- Hi, We are looking for 2 resources, who are expert in Big Data, with good years of experience in Hadoop, Kafka, Hive, Strom and if you have experience in MapR then it would give you a real edge for selection to join our team.
Skills: Hadoop Apache Hive Apache Kafka Big Data
Fixed-Price - Entry Level ($) - Est. Budget: $80 - Posted
Density-based method is a remarkable class in clustering data streams, which has the ability to discover arbitrary shape clusters and to detect noise. Furthermore, it does not need the number of clusters in advance. We not only summarize the main density-based clustering algorithms on data streams, discuss their uniqueness and limitations, but also explain how they address the challenges in clustering data streams. I need a base paper and problem in the same field which can help in extending the research with a novel problem.
Skills: Hadoop Big Data
Hourly - Intermediate ($$) - Est. Time: Less than 1 month, Less than 10 hrs/week - Posted
We are embarking on a project to improve storage benchmarks. The first step is better understanding how applications access storage. We need to analyze data we receive in the form of trace files (.CSVs with one record per storage I/O). A single trace can easily be a billion or more I/Os. Specifications for a parser, step one, are attached. If the parser goes well we should have more.
Skills: Hadoop Data Analytics Data Visualization