Hire the best MapReduce specialists

Check out MapReduce specialists with the skills you need for your next job.
  • $110 hourly
    Distributed Computing: Apache Spark, Flink, Beam, Hadoop, Dask Cloud Computing: GCP (BigQuery, DataProc, GFS, Dataflow, Pub/Sub), AWS EMR/EC2 Containerization Tools: Docker, Kubernetes Databases: Neo4j, MongoDB, PostgreSQL Languages: Java, Python, C/C++
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Apache Kafka
    Cloud Computing
    Apache Hadoop
    White Paper Writing
    Academic Writing
    Google Cloud Platform
    Dask
    Apache Spark
    Research Paper Writing
    Apache Flink
    Kubernetes
    Python
    Java
  • $45 hourly
    As a highly experienced Data Engineer with over 10+ years of expertise in the field, I have built a strong foundation in designing and implementing scalable, reliable, and efficient data solutions for a wide range of clients. I specialize in developing complex data architectures that leverage the latest technologies, including AWS, Azure, Spark, GCP, SQL, Python, and other big data stacks. My extensive experience includes designing and implementing large-scale data warehouses, data lakes, and ETL pipelines, as well as data processing systems that process and transform data in real-time. I am also well-versed in distributed computing and data modeling, having worked extensively with Hadoop, Spark, and NoSQL databases. As a team leader, I have successfully managed and mentored cross-functional teams of data engineers, data scientists, and data analysts, providing guidance and support to ensure the delivery of high-quality data-driven solutions that meet business objectives. If you are looking for a highly skilled Data Engineer with a proven track record of delivering scalable, reliable, and efficient data solutions, please do not hesitate to contact me. I am confident that I have the skills, experience, and expertise to meet your data needs and exceed your expectations.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Snowflake
    ETL
    PySpark
    MongoDB
    Unix Shell
    Data Migration
    Scala
    Microsoft Azure
    Amazon Web Services
    SQL
    Apache Hadoop
    Cloudera
    Apache Spark
  • $40 hourly
    I am a developer focused on providing highly efficient software solutions. - Full Stack Developer - Data Scientist
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Apache Spark
    Cloudera
    CakePHP
    Apache HBase
    Apache Hadoop
    Laravel
    Python
    PHP
    MongoDB
    JavaScript
  • $25 hourly
     Certification in Big Data/Hadoop Ecosystem  Big Data Environment: Google Cloud Platform, Cloudera, HortonWorks and AWS, SnowFlake, Databricks, DC/OS  Big Data Tools : Apache Hadoop, Apache Spark, Apache Kafka, Apache Nifi, Apache Cassandra, Yarn/Mesos, Oozie, Sqoop, Airflow, Glue, Athena, S3 Buckets, Lambda, Redshift, DynamoDB ,Delta Lake, Docker, GIT, Bash Scripts Jenkins, Postgres, MongoDB, Elastic Search, Kibana, Ignite, TiDB  Certification SQL Server, Database Development and Crystal Report.  SQL Server Tools: SQL Management Studio, BIDS, SSIS, SSAS and SSRS  BI/Dashboarding Tools: Power BI, Tableau, Kibana  Big Data Development Programing Languages: Scala and python. ======================================================================= ************************************* Big Data Engineer**********************************************  Hands on experience with Google cloud platform, Big Query, Google Data Studio and Flow  Developing ETL pipeline for SQL server as well using SSIS.  For Reporting and Analysis using SSIS, SSRS and SSAS cubes.  Having amazing experience with Big data framework and open source technologies (Apache Nifi, Kafka, Spark and Cassandra, HDFS, Hive Docker/Cassandra/ Postgres SQL, Git, Bash Scripts Jenkins, MongoDB, Elastic Search, Ignite, TiDB.  Managing data warehouse Big Data cluster services and developments of Data Flows.  Writing big data/Spark ETL applications for different sources (SQL, Oracle, CSV, XML,JSON) to support different department for analytics.  Extensive work with Hive, Hadoop, Spark, Docker, Apache Nifi  Supporting different department for big data analytics.  Build multiple end to end Fraud monitoring alert based systems.  Preferable language is Scala and python as well. ************Big Data Engineer– Fraud Management at VEON *************  Devolved ETL Pipeline from Kafka to Cassandra using Spark in Scala Language.  Using Big Data Tools with Horton Works and AWS (Apache Nifi, Kafka, Spark and Cassandra, Elastic Search)  Dashboard Developments - Tableau and Kibana.  Writing SQL server complex queries, procedures and Functions.  Developing ETL pipeline for SQL server as well using SSIS.  For Reporting and Analysis using SSIS, SSRS and SSAS cubes.  Developing and designing Auto Email Reports.  Offline Data Analytics for Fraud Detection and Setting up controls for prevention.  SQL Database Development.  System Support of Fraud Management.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Google Cloud Platform
    SQL Programming
    Data Warehousing
    Database
    AWS Glue
    PySpark
    MongoDB
    Python Script
    Docker
    Apache Hadoop
    Apache Spark
    Databricks Platform
    Apache Kafka
    Apache Hive
  • $55 hourly
    I focus on data engineering, software engineering, ETL/ELT, SQL reporting, high-volume data flows, and development of robust APIs using Java and Scala. I prioritize three key elements: reliability, efficiency, and simplicity. I hold a Bachelor's degree in Information Systems from Pontifícia Universidade Católica do Rio Grande do Sul as well as graduate degrees in Software Engineering from Infnet/FGV and Data Science (Big Data) from IGTI. In addition to my academic qualifications I have acquired a set of certifications: - Databricks Certified Data Engineer Professional - AWS Certified Solutions Architect – Associate - Databricks Certified Associate Developer for Apache Spark 3.0 - AWS Certified Cloud Practitioner - Databricks Certified Data Engineer Associate - Academy Accreditation - Databricks Lakehouse Fundamentals - Microsoft Certified: Azure Data Engineer Associate - Microsoft Certified: DP-200 Implementing an Azure Data Solution - Microsoft Certified: DP-201 Designing an Azure Data Solution - Microsoft Certified: Azure Data Fundamentals - Microsoft Certified: Azure Fundamentals - Cloudera CCA Spark and Hadoop Developer - Oracle Certified Professional, Java SE 6 Programmer My professional journey has been marked by a deep involvement in the world of Big Data solutions. I've fine-tuned my skills with Apache Spark, Apache Flink, Hadoop, and a range of associated technologies such as HBase, Cassandra, MongoDB, Ignite, MapReduce, Apache Pig, Apache Crunch and RHadoop. Initially, I worked extensively with on-premise environments but over the past five years my focus has shifted predominantly to cloud based platforms. I've dedicated over two years to mastering Azure and I’m currently immersed in AWS. I have a great experience with Linux environments as well as strong knowledge in programming languages like Scala (8+ years) and Java (15+ years). In my earlier career phases, I had experience working with Java web applications and Java EE applications, primarily leveraging the WebLogic application server and databases like SQL Server, MySQL, and Oracle.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Scala
    Apache Solr
    Apache Kafka
    Apache Spark
    Bash Programming
    Elasticsearch
    Java
    Progress Chef
    Apache Flink
    Apache HBase
    Apache Hadoop
    MongoDB
    Docker
  • $95 hourly
    => Let's Connect Hello, I'm Dima, a seasoned CyberSecurity Specialist and Turnkey Infrastructure Expert specializing in BigData solutions and data analysis, utilizing a DevOps approach. => Expertise Overview With a robust passion for constructing SOC, SOAR, and SIEM solutions, my primary focus lies in developing data ingestion, enrichment, and analysis pipelines, ensuring they are highly available and fault-tolerant. My expertise extends to building central logging and real-time processing platforms from the ground up, optimizing them for performance, security, and reliability across multiple environments, whether in the cloud or on-premise. => Value Proposition My commitment is to deliver solutions that not only centralize security and threat intelligence but also facilitate enhanced control over data, ultimately contributing to infrastructure cost savings. => Technological Summary CyberSecurity:------- > Wazuh, Suricata, pfSense BigData:--------------- > Kafka, ElasticSearch, OpenSearch Data Processing:----- > FluentD, Vector.dev, Apache NiFi Infra as Code:--------- > Terraform, cdktf, cdk8s Virtualization:--------- > Proxmox, VMware Containerization:----- > Kubernetes Clouds:---------------- > AWS, Hetzner, DigitalOcean, Linode Automation:----------- > Jenkins, GitHub Actions Monitoring:----------- > Zabbix, Grafana, Kibana, Prometheus, Thanos Mail:--------------------> MailCow SMTP/IMAP, Postfix VPN:------------------- > OpenVPN Server Programming:-------- > Bash, Python, TypeScript Operating Systems:- > CentOS, RHEL, Rocky Linux, Ubuntu, Debian => Personal Attributes • Leadership: Leading by example with a team-first approach • End-to-End Execution: Proficient from POC to Enterprise-level implementation • Resilience: Demonstrating high thoroughness and endurance • Adaptability: A quick, can-do architect and experienced troubleshooter • Optimization: Adept in process and performance optimization • Documentation: Skilled technical documentation writer • Vision: A visionary in technological implementation and solution provision
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Elasticsearch
    Linux System Administration
    Apache Kafka
    Apache Hadoop
    Email Security
    Machine Learning
    ELK Stack
    Cloudera
    Zabbix
    MySQL
    Big Data
    Apache NiFi
    PfSense
    Red Hat Administration
    Proxmox VE
    Amazon Web Services
  • $22 hourly
    Data Engineering in Python ♣ ETL/Data Pipeline: Apache Airflow, Pandas, PySpark, Apache Kafka, SQLAlchemy, DBT Data Science in Python ♣ Machine Learning: Hugging Face, Pandas, Numpy, Scikit-learn, NLTK, Keras, Tensorflow, Optuna, PySpark, Matplotlib Tools Expertise ♣ Databases: PostgreSQL ♣ Data Analytics: Metabase, Cluvio, Chartio, DBT, Jinja, Segment, RudderStack, Stitch, Amplitude, Qlik, Apache Superset, BigQuery ♣ Cloud platforms: GCP, AWS, ♣ Version Control: GitHub, GitLab ♣ DevOps: Linux, VMs, Jupyter Notebook, Colab Web Develoment ♣ Django Web Scraping in Python ♣ BeautifulSoup Technical Skill-Set: ♣ Data Architecture Design ♣ Database Optimization & Tuning ♣ Data Pipeline Development ♣ Schema Design & Management ♣ Data Governance & Security ♣ Performance Optimization ♣ Data Quality Management ♣ Version Control & CI/CD ♣ Documentation & Technical Writing ♣ Problem-solving & Debugging ♣ Data munging/pre-processing ♣ Regression & classification modeling ♣ Descriptive & inferential statistics ♣ Building data products
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Model Optimization
    Keras
    pandas
    Artificial Intelligence
    Natural Language Processing
    Analytics
    Machine Learning
    Python
    Business Intelligence
    Deep Learning
    Data Transformation
    BigQuery
    Data Processing
    ETL Pipeline
    Data Engineering
  • $25 hourly
    I’m a developer with experience in NLP、LLM 、 CV and recommendation system and bigdata. 1. I’m experienced in RAG、langchian、deepfast、llm、 object dection、video generation 2.I’m experienced in hadoop/hdfs/yarn/hbase/redis/kafka/hive, 2. I’m experienced in tensorflow/kubeflow/tf serving/trtion serving 3.I’m experienced in k8s/docker/html/jqurey/spring boot/mybatis 4.I’m experienced in aws componet, such as emr/ec2/s3/code deploy
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    jQuery
    LangChain
    LLM Prompt Engineering
    PySpark
    AWS Application
    Big Data
    Artificial Intelligence
    Apache Hadoop
    Spring Boot
    Kubernetes
    Apache Flink
    Java
    TensorFlow
    Apache Spark
  • $25 hourly
    I bring extensive hands-on experience in the realm of data science, showcasing proficiency in various Hadoop components such as MapReduce, Hive, Pig, alongside a deep understanding of AWS cloud services. Over the course of my career, I have successfully executed numerous projects utilizing machine learning techniques for in-depth data analysis. Specifically, I leverage Apache Spark to efficiently process vast datasets for analytical purposes. My expertise extends to the full spectrum of Spark's capabilities, including Spark Streaming, Spark MLlib, and Spark GraphX, which have proven instrumental in enhancing the speed and scalability of data processing in various projects. I have implemented Spark MLlib to develop machine learning models tailored to meet specific client requirements, focusing on prediction and classification tasks. In my current role, I am deeply involved in working with Hadoop components, and I continue to harness the advanced features of Spark, such as Spark Streaming, MLlib, and GraphX, for real-time data processing requirements. Moreover, I actively incorporate DevOps practices into my workflow to ensure seamless collaboration between development and operations teams. This includes the integration of continuous integration/continuous deployment (CI/CD) pipelines, automated testing, and infrastructure as code (IaC) principles. Embracing a DevOps mindset enhances the overall efficiency and reliability of the software development lifecycle. I take pride in my ability to align machine learning methodologies with data processing workflows to meet client demands effectively. This involves leveraging Spark MLlib for predictive modeling and classification tasks, ensuring a holistic approach to addressing client requirements and business objectives. Throughout my journey in data science, I have remained dedicated to staying at the forefront of technology, constantly adapting to new tools and methodologies. I am enthusiastic about bringing this multifaceted expertise, encompassing data science and DevOps practices, to tackle new challenges and make meaningful contributions to future projects.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Data Scraping
    Google Analytics
    AWS Lambda
    Apache Kafka
    Amazon DynamoDB
    Apache Hadoop
    BigQuery
    Big Data
    Amazon ECS
    SQL
    Sentiment Analysis
    Machine Learning
    NLTK
    Apache Spark MLlib
    Apache Spark
  • $20 hourly
    Innovative, passionate, and quick software developer and architect with deep knowledge of programming concepts and internals. 24+ years of hands-on industry experience in programming, designing, managing, & leading software projects & companies. Change agent, problem solver with a passion for technology; skilled in grasping the big picture, conceptualizing, developing & implementing solutions by partnering closely with business leaders. Worked for different type of industries, which includes Health-care Informatics, Lab Informatics, MOOC, Online Education, ERPs, Electronic Design Automation (EDA), Semi Conductor, Heavy mechanical manufacturing, Travel, Pharmaceutical & e-commerce. Skilled in: Python | PHP | Apex | Objective C | Java | C/C++ | Flex | ActionScript | JavaScript | Perl | VB6 iPhone SDK | Android SDK | BlackBerry SDK | Flex SDK | Sancha Django | Django REST | Odoo | Web2Py | Zope | Plone | Open edX | CodeIgniter | .NET | Java EE SQL Server | MySQL | PostgreSQL | Sybase | DB2 | SQLite | MS Access Odoo | OpenERP | Salesforce CRM | MS SharePoint git | SVN | CVS | VSS | Unfuddle Scrum | UP | UML | CASE Tools | Poseidon | Rambaugh OMT ASP.NET | ASP | Telerik Visual C++| MFC | Win32 | COM | DCOM | OLE | Multithreading ANT | JDK | Digester| Struts | Servlets | JSP | EJB | WebSphere | Eclipse Web 2.0 | AJAX | CSS | XML | XSD OpenSource | github Knowledge Management | MediaWiki | PHP Health-care Informatics | ICD9 | ICD10 | HIPPA | HL7 SEMI Standards | EFEM | SEMI E87 | SEMI E90 | SEMI E40 | Robotics | SECS/GEM | SCADA Customs ERP | Landing Cost | HS Codes | WTO Standards Lab Informatics | Bika | OLiMS | LiMS | LIS Windows | MacOS | Linux | Ubuntu | Unix | EC2 Windows CE | Micro Controller Programming | Dynamic C
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Ehealth
    Mapbox
    Apache Hadoop
    LIMS
    Android
    iOS
    Odoo
    web2py
    Django
    Laravel
    Python
    PHP
    WordPress
    React
    JavaScript
  • $150 hourly
    I'm a Software Engineer and Devops Enginner with 10+ years experience in both fields. Ready to take on any challenge, quick learner and an excellent problem solver. I am always eager to learn new tools and technologies and I try to spend some time weekly on this topic. Devops skills: - AWS Cloud, Azure Cloud, Google Cloud - Kubernetes, Helm, Docker ( Docker-compose, Docker registry) - ARM-Templates, CloudFormation, Terraform, Salt, Ansible - Prometheus, Grafana, Nagios - Linux ( Ubuntu, Centos), Windows, Mac OS X - Spark, Hadoop, Kafka, Yarn, Mesos, Hbase, Cassandra, Hive, Hdfs, Zookeeper, ElasticSearch - Postgresql, Mysql, Azure Sql, Mongo - Apache, Nginx, Tomcat - Git, Gitlab - Jira, Confluence - Jenkins Programming skills: - Scala - Java - C# - Python - Basic and debugging skills for C and C++ and JavaScript - bash, powershell With most of my projects, I am responsible for everything from ground up. This includes the code of the application, but also the automated pipelines, automated deployments, version control, and Infrastructure-as-Code. I have vast experience with designing and developing system architectures in Azure Cloud and Amazon Web Services.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Azure DevOps
    DevOps
    Linux
    Amazon EC2
    GitHub
    Kubernetes
    CI/CD
    Bash
    Java
    Python
    Terraform
    Microsoft Azure
    Amazon Web Services
  • $35 hourly
    5+ years of experience in Big Dat Technologies like Spark, Hadoop, Hive, Sqoop, ADF, Databricks. 5+ years of experience in ELK Stack( Elasticsearch, Logstash and Kibana). Microsoft Azure Certified Data Engineer. Elasticsearch and Kibana Certified. MongoDB Certified Developer.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Microsoft Azure
    Databricks Platform
    Apache Spark
    PySpark
    MongoDB
    Logstash
    Elasticsearch
    Grok Framework
    ELK Stack
    Apache Hadoop
    Hive
    Bash
    SQL
    Kibana
  • $25 hourly
    I am a Senior AI Engineer and Full-Stack Developer with over 10 years of experience in developing intelligent, data-driven solutions that drive business growth. My expertise spans Generative AI, Large Language Models (LLMs), and full-stack development, helping businesses integrate advanced automation and AI-driven intelligence into their applications. - 𝗘𝘅𝗽𝗲𝗿𝘁𝗶𝘀𝗲: ✔𝗔𝗜 & 𝗠𝗮𝗰𝗵𝗶𝗻𝗲 𝗟𝗲𝗮𝗿𝗻𝗶𝗻𝗴: LLMs, RAG, LangChain, OpenAI, LLaMA, Hugging Face, Transformers, Diffusion Models, TensorFlow, PyTorch, Scikit-learn, OpenCV ✔𝗕𝗮𝗰𝗸𝗲𝗻𝗱 𝗗𝗲𝘃𝗲𝗹𝗼𝗽𝗺𝗲𝗻𝘁: Python, Node.js, Java, Spring Boot, Django, Flask, Microservices ✔𝗙𝗿𝗼𝗻𝘁𝗲𝗻𝗱 𝗧𝗲𝗰𝗵𝗻𝗼𝗹𝗼𝗴𝗶𝗲𝘀: React, Angular, Vue.js, Bootstrap ✔𝗗𝗮𝘁𝗮𝗯𝗮𝘀𝗲𝘀: MongoDB, MySQL, PostgreSQL, Vector Databases (Pinecone, FAISS) ✔𝗖𝗹𝗼𝘂𝗱 & 𝗗𝗲𝘃𝗢𝗽𝘀: AWS, GCP, Kubernetes, Docker Swarm, CI/CD Pipelines, MLOps (Model Deployment, Jenkins, GitLab CI, CircleCI) ✔𝗕𝗶𝗴 𝗗𝗮𝘁𝗮 𝗣𝗿𝗼𝗰𝗲𝘀𝘀𝗶𝗻𝗴: Spark, Hadoop I specialize in building scalable, AI-enhanced applications that streamline workflows, automate tasks, and improve decision-making. Whether you need a custom AI model, an intelligent web application, or a cloud-based deployment. 💡 𝗜𝗻𝗱𝘂𝘀𝘁𝗿𝘆 𝗘𝘅𝗽𝗲𝗿𝗶𝗲𝗻𝗰𝗲: EdTech, FinTech, E-Commerce, Healthcare, Media, Travel, Real Estate, Manufacturing, and Energy & Utilities. 𝗟𝗲𝘁'𝘀 𝗰𝗼𝗹𝗹𝗮𝗯𝗼𝗿𝗮𝘁𝗲 𝗮𝗻𝗱 𝘁𝘂𝗿𝗻 𝘆𝗼𝘂𝗿 𝗔𝗜 𝘃𝗶𝘀𝗶𝗼𝗻 𝗶𝗻𝘁𝗼 𝗿𝗲𝗮𝗹𝗶𝘁𝘆!
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    TypeScript
    React
    Java
    Python
    OpenCV
    Python Scikit-Learn
    PyTorch
    TensorFlow
    Vector Database
    Hugging Face
    Diffusion Model
    LLaMA
    Retrieval Augmented Generation
    LLM Prompt Engineering
    OpenAI API
  • $30 hourly
    Good Day Over 9+ years of extensive hands-on experience in Big Data technologies from core Hadoop Eco System to GCP-AWS cloud-based platforms. Expertise in Cloud (GCP, AWS) and In-Premise (Hadoop) systems and its various components Experience working in different Google Cloud Platform Technologies like Big Query, Dataflow, Dataproc, Pub sub, Composer and AWS like EMR, Redshift, Lambda, Step Functions, EKS over open source Ecosystem Hadoop, HDFS, MapReduce, Kafka, Spark, Hive. - Design and Development of Ingestion Framework over Google Cloud, AWS and Hadoop cluster. - Good Knowledge on Hadoop Cluster architecture and monitoring. - Extensive Experience on importing and exporting data using Kafka. - Strong hands-on experience in ETL processing using Spark, Scala/Python and Kafka. - Integration various data science models into Data Engineering platform over cloud and in-premise - End-to-End Bigdata platform setup over In-premise to cloud - Migration of tradition data systems into cost-friendly, reliable, scalable data systems - Developing and scheduling ETL workflows in Hadoop using Oozie, Airflow, Google Cloud Composer - Setup, manage and optimize Distributed data warehouses like Hive, Big Query, Redshift - Managing different queues over pub/sub, Kafka - Handled various frequency based ingestion with Realtime, Near Realtime, Scheduled batch flow - Handling integration with RDBMS like MSSQL, MySQL and NOSQLs like MongoDB, Elasticsearch - Complete Data driven system build experience from Data Ingestion, Transformation, Store and Analytics over BI platforms like PowerBI, DataStudio - Managing various loggings using ELK (Elasticsearch, Logstash and Kibana) - Setting up various webserver configurations - Setting and managing Devops pipelines over Kubernetes, Docker, Azure Devops, Github devops
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    AI Content Creation
    Database
    Python Script
    Machine Learning
    Flask
    API
    Data Migration
    Apache Airflow
    ETL
    Apache Kafka
    Google Cloud Platform
    Python
    SQL
    Amazon Web Services
    JavaScript
  • $70 hourly
    🎓 𝗖𝗲𝗿𝘁𝗶𝗳𝗶𝗲𝗱 𝗗𝗮𝘁𝗮 𝗣𝗿𝗼𝗳𝗲𝘀𝘀𝗶𝗼𝗻𝗮𝗹 with 𝟲+ 𝘆𝗲𝗮𝗿𝘀 of experience and hands-on expertise in Designing and Implementing Data Solutions. 🔥 4+ Startup Tech Partnerships ⭐️ 100% Job Success Score 🏆 In the top 3% of all Upwork freelancers with Top Rated Plus 🏆 ✅ Excellent communication skills and fluent English If you’re reading my profile, you’ve got a challenge you need to solve and you are looking for someone with a broad skill set, minimal oversight and ownership mentality, then I’m your go-to expert. 📞 Connect with me today and let's discuss how we can turn your ideas into reality with creative and strategic partnership.📞 ⚡️Invite me to your job on Upwork to schedule a complimentary consultation call to discuss in detail the value and strength I can bring to your business, and how we can create a tailored solution for your exact needs. 𝙄 𝙝𝙖𝙫𝙚 𝙚𝙭𝙥𝙚𝙧𝙞𝙚𝙣𝙘𝙚 𝙞𝙣 𝙩𝙝𝙚 𝙛𝙤𝙡𝙡𝙤𝙬𝙞𝙣𝙜 𝙖𝙧𝙚𝙖𝙨, 𝙩𝙤𝙤𝙡𝙨 𝙖𝙣𝙙 𝙩𝙚𝙘𝙝𝙣𝙤𝙡𝙤𝙜𝙞𝙚𝙨: ► BIG DATA & DATA ENGINEERING Apache Spark, Hadoop, MapReduce, YARN, Pig, Hive, Kudu, HBase, Impala, Delta Lake, Oozie, NiFi, Kafka, Airflow, Kylin, Druid, Flink, Presto, Drill, Phoenix, Ambari, Ranger, Cloudera Manager, Zookeeper, Spark-Streaming, Streamsets, Snowflake ► CLOUD AWS -- EC2, S3, RDS, EMR, Redshift, Lambda, VPC, DynamoDB, Athena, Kinesis, Glue GCP -- BigQuery, Dataflow, Pub/Sub, Dataproc, Cloud Data Fusion Azure -- Data Factory, Synapse. HDInsight ► ANALYTICS, BI & DATA VISUALIZATION Tableau, Power BI, SSAS, SSMS, Superset, Grafana, Looker ► DATABASE SQL, NoSQL, Oracle, SQL Server, MySQL, PostgreSQL, MongoDB, PL/SQL, HBase, Cassandra ► OTHER SKILLS & TOOLS Docker, Kubernetes, Ansible, Pentaho, Python, Scala, Java, C, C++, C# 𝙒𝙝𝙚𝙣 𝙮𝙤𝙪 𝙝𝙞𝙧𝙚 𝙢𝙚, 𝙮𝙤𝙪 𝙘𝙖𝙣 𝙚𝙭𝙥𝙚𝙘𝙩: 🔸 Outstanding results and service 🔸 High-quality output on time, every time 🔸 Strong communication 🔸 Regular & ongoing updates Your complete satisfaction is what I aim for, so the job is not complete until you are satisfied! Whether you are a 𝗦𝘁𝗮𝗿𝘁𝘂𝗽, 𝗘𝘀𝘁𝗮𝗯𝗹𝗶𝘀𝗵𝗲𝗱 𝗕𝘂𝘀𝗶𝗻𝗲𝘀𝘀 𝗼𝗿 𝗹𝗼𝗼𝗸𝗶𝗻𝗴 𝗳𝗼𝗿 your next 𝗠𝗩𝗣, you will get 𝗛𝗶𝗴𝗵-𝗤𝘂𝗮𝗹𝗶𝘁𝘆 𝗦𝗲𝗿𝘃𝗶𝗰𝗲𝘀 at an 𝗔𝗳𝗳𝗼𝗿𝗱𝗮𝗯𝗹𝗲 𝗖𝗼𝘀𝘁, 𝗚𝘂𝗮𝗿𝗮𝗻𝘁𝗲𝗲𝗱. I hope you become one of my many happy clients. Reach out by inviting me to your project. I look forward to it! All the best, Anas ⭐️⭐️⭐️⭐️⭐️ 🗣❝ Muhammad is really great with AWS services and knows how to optimize each so that it runs at peak performance while also minimizing costs. Highly recommended! ❞ ⭐️⭐️⭐️⭐️⭐️ 🗣❝ You would be silly not to hire Anas, he is fantastic at data visualizations and data transformation. ❞ 🗣❝ Incredibly talented data architect, the results thus far have exceeded our expectations and we will continue to use Anas for our data projects. ❞ ⭐️⭐️⭐️⭐️⭐️ 🗣❝ The skills and expertise of Anas exceeded my expectations. The job was delivered ahead of schedule. He was enthusiastic and professional and went the extra mile to make sure the job was completed to our liking with the tech that we were already using. I enjoyed working with him and will be reaching out for any additional help in the future. I would definitely recommend Anas as an expert resource. ❞ ⭐️⭐️⭐️⭐️⭐️ 🗣❝ Muhammad was a great resource and did more than expected! I loved his communication skills and always kept me up to date. I would definitely rehire again. ❞ ⭐️⭐️⭐️⭐️⭐️ 🗣❝ Anas is simply the best person I have ever come across. Apart from being an exceptional tech genius, he is a man of utmost stature. We blasted off with our startup, high on dreams and code. We were mere steps from the MVP. Then, pandemic crash. Team bailed, funding dried up. Me and my partner were stranded and dread gnawed at us. A hefty chunk of cash, Anas and his team's livelihood, hung in the balance, It felt like a betrayal. We scheduled a meeting with Anas to let him know we were quitting and request to repay him gradually over a year, he heard us out. Then, something magical happened. A smile. "Forget it," he said, not a flicker of doubt in his voice. "The project matters. Let's make it happen!" We were floored. This guy, owed a small fortune, just waved it away? Not only that, he offered to keep building, even pulled his team in to replace our vanished crew. As he spoke, his passion was a spark that reignited us. He believed. In us. In our dream. In what he had developed so far. That's the day Anas became our partner. Not just a contractor, but a brother in arms. Our success story owes its spark not to our own leap of faith, but from the guy who had every reason to walk away. Thanks, Anas, for believing when we couldn't.❞
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Solution Architecture Consultation
    AWS Lambda
    ETL Pipeline
    Data Management
    Data Warehousing
    AWS Glue
    Apache Spark
    Amazon Redshift
    ETL
    Python
    SQL
    Marketing Analytics
    Big Data
    Data Visualization
    Artificial Intelligence
  • $40 hourly
    I'm Linux DevOps and Cloud architect since 2002. Most of my professional career is with design, setup and DevOps of medium and high loaded web farms, NoSQL databases which are time-critical and require 24/7/365 uptime. During the last several years, I'm concentrated on architect & administration of Hadoop ecosystem, Big-Data systems (Cassandra, ElasticSearch, Riak ...) and distributed storage Ceph. I have big experience with a variety of web servers and load balancers (Apache, Nginx, HAProxy, Tomcat, Jetty etc .. ) as well as with cloud services such as AWS, Azure and GCP.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Big Data
    Apache HBase
    Linux System Administration
    Apache Cassandra
    Golang
    Nomad
    CI/CD Platform
    Apache Hadoop
    Consul
    Kubernetes
    Elasticsearch
    Google Cloud Platform
    Python
    Amazon Web Services
    Linux
  • $35 hourly
    Over 5 years of working experience in data engineering, ETL, AWS, ML and python. AWS data analytics and machine learning certified.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    OpenAI Embeddings
    Docker
    Terraform
    Amazon ECS
    AWS Lambda
    Amazon Redshift
    Amazon S3
    Amazon Web Services
    Analytics
    PostgreSQL
    PySpark
    SQL
    pandas
    AWS Glue
    Python
  • $175 hourly
    Mr. Joshua B. Seagroves is a seasoned professional having served as an Enterprise Architect/Senior Data Engineer for multiple Fortune 100 Companies. With a successful track record as a startup founder and CTO, Mr. Seagroves brings a wealth of experience to his role, specializing in the strategic design, development, and implementation of advanced technology systems. Throughout his career, Mr. Seagroves has demonstrated expertise in architecting and delivering cutting-edge solutions, particularly in the realm of data engineering and sciences. He has successfully spearheaded the implementation of multiple such systems and applications for a diverse range of clients. As part of his current responsibilities, Mr. Seagroves actively contributes to the prototyping and research efforts in the field of data engineering/data science, specifically in the development of operational systems for critical mission systems. Leveraging his extensive background in architecture and software modeling methodologies, he has consistently led and collaborated with multidisciplinary teams, successfully integrating various distributed computing technologies, including Hadoop, NiFi, HBase, Accumulo, and MongoDB. Mr. Seagroves' exceptional professional achievements and extensive experience make him a highly sought-after expert in his field. His comprehensive knowledge and hands-on expertise in advanced technology systems and big data make him a valuable asset to any organization.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    YARN
    Apache Hadoop
    Big Data
    Apache Zookeeper
    TensorFlow
    Apache Spark
    Apache NiFi
    Apache Kafka
    Artificial Neural Network
    Artificial Intelligence
  • $40 hourly
    🔍🚀 Welcome to a world of data-driven excellence! 🌐📊 Greetings, fellow professionals! I am thrilled to introduce myself as a dedicated Data Consultant / Engineer, leveraging years of honed expertise across a diverse spectrum of data stacks 🌍. My journey has been enriched by a wealth of experience, empowering me with a comprehensive skill set that spans Warehousing📦, ETL⚙, Analytics📈, and Cloud Services☁. Having earned the esteemed title of GCP Certified Professional Data Engineer 🛠, I am your partner in navigating the complex data landscape. My mission is to unearth actionable insights from raw data, shaping it into a strategic asset that fuels growth and innovation. With a deep-rooted passion for transforming data into valuable solutions, I am committed to crafting intelligent strategies that empower businesses to flourish. Let's embark on a collaborative journey to unlock the full potential of your data. Whether it's architecting robust data pipelines ⛓, optimizing storage solutions 🗃, or designing analytics frameworks 📊, I am dedicated to delivering excellence that transcends expectations. Reach out to me, and together, let's sculpt a future where data powers success. Thanks!
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    PySpark
    Machine Learning
    Natural Language Processing
    Informatica
    Data Science
    Data Warehousing
    Snowflake
    Data Analysis
    Big Data
    BigQuery
    ETL
    Apache Airflow
    Apache Hadoop
    Apache Spark
    Databricks Platform
    Python
    Apache Hive
  • $50 hourly
    A Backend Software Engineering with more than 6 years of experience. Have worked with large-scale backend/distributed systems and big data systems. A DevOps engineer with 4 years of experience - both on-premises and AWS, experienced with K8s, Terraform, Ansible, CI/CD. Currently working as Principal Engineer/ Solution Architect role.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Architectural Design
    GraphQL
    Serverless Computing
    Amazon Web Services
    DevOps
    API Development
    Elasticsearch
    Apache Kafka
    Scala
    Apache Spark
    Docker
    Apache Hadoop
    Kubernetes
  • $25 hourly
    I’m a Senior Data Engineer with 6+ years of experience designing and optimizing Big Data solutions, ETL pipelines, and cloud architectures. I specialize in Google Cloud (GCP), Apache Spark (PySpark & Scala), and Big Data platforms such as Cloudera, Hortonworks, and AWS EMR. 🔹 Expertise & Skills: ✔ Data Engineering & ETL (NiFi, Apache Airflow, Sqoop, Kafka) ✔ Big Data Processing (Spark, Hadoop, Hive, MongoDB, ElasticSearch) ✔ Cloud Technologies (Google Cloud, AWS, Docker, Kubernetes) ✔ Workflow Automation & CI/CD (GitLab, Ansible, Terraform) ✔ Large-scale Data Migration & Performance Optimization 💼 Recent Projects & Achievements: ✅ Designed and optimized Spark-based ETL pipelines , processing petabytes of data ✅ Migrated legacy BI systems to modern Big Data architectures ✅ Developed a real-time customer analytics platform using GCP & BigQuery ✅ Implemented scalable Hadoop/Spark clusters 🎓 Certifications & Education: 📌 Google Cloud Professional Data Engineer Certified 📌 Master’s in Computer Engineering – Esprit School of Engineering 💡 If you're looking for a data-driven expert to optimize your Big Data & Cloud Infrastructure, let’s connect!
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Cloud Engineering
    Cloudera
    Apache Hadoop
    Data Warehousing
    Apache NiFi
    Linux
    Apache Spark
    Data Lake
    Data Analysis
    SQL
    Big Data
    Business Intelligence
    Scala
    Apache Hive
    Python
  • $40 hourly
    I have completed 58+ jobs, earned $200K (on Upwork), and have a stellar 5/5 feedback. ⭐️⭐️⭐️⭐️⭐️ $145K+ - Golang/Python Microservices "Having worked with several overseas teams in the past, I have to say that I struck gold finding Fahad and his team. They really became an extension to my team, are very disciplined about understanding what needs to happen, contributing at the highest levels. Fahad is an honest guy, smart, and always helpful in meeting deadlines and pleasing customers. I look forward to continuing to work with Fahad and his team as they are a true asset and highly productive development group." ⭐️⭐️⭐️⭐️⭐️ $28K+ - Python/Flask "Fahad knows his stuff and has integrated very well into our team. This is due to his communication skills and motivation. I highly recommend him" And many more! My experience represents countless hours spent mastering skills and solving complex problems, ensuring you don't have to navigate these challenges yourself. Hire me if you: ✅ Want a SWE with strong technical skills ✅ Need a Python, Go, or a Rust developer ✅ Seek to leverage AI for predictive analytics, enhancing data-driven decision-making ✅ Require AI-based optimization of existing software for efficiency and scalability ✅ Wish to integrate AI and machine learning models to automate tasks and processes ✅ Need expert guidance in selecting and implementing the right AI technologies for your project ✅ Desire a detail-oriented person who asks questions and figures out things on his own ✅ Even have a requirement in your mind but are not able to craft it into a technical format ✅ Want advice on what tools or tech you want to implement in your next big project ✅ Are stuck in a data modeling problem and need a solution architect ✅ Want to optimize a data pipeline Don't hire me if you: ❌ Have a huge project that needs to be done overnight ❌ Have academic work to be done About me: ⭐️ A data engineer with proven experience in designing and implementing big data solutions ⭐️ Skilled in integrating AI technologies to solve complex problems, improve efficiency, and innovate within projects ⭐️ A Go developer specialized in creating microservices ⭐️ A certified Data Engineer on AWS technologies ⭐️ Will optimize your code in every single commit without even mentioning or charging extra hours ⭐️ Diverse experience with start-ups and enterprises taught me how to work under pressure yet work professionally
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Web Scraping
    Microservice
    ETL Pipeline
    Big Data
    AI Bot
    OpenAI API
    Artificial Intelligence
    Generative AI
    Large Language Model
    Golang
    Apache Spark
    Python
  • $30 hourly
    With close to 10 years of industry experience in the specialised field of software design and development, I possess proven capabilities to develop high quality software applications. My aim is to obtain a challenging position that will utilise my skills and experiences and which will also provide me with the opportunity for growth and advancement. Languages- Java, Python, Javascript. Skills- Core: Data Structures and Algorithms. Data Analysis: Hadoop MapReduce Backend: Java, Spring, Spring Boot, Microservices, Struts, Design Principles, Design Patterns, SQL, Webservices, SOA(REST and SOAP), JMS, Servlets, Swing, JSP, MAVEN, Sub versioning(svn, git), Jenkins. Frontend: HTML5, CSS3, Javascript, Jquery, Bootstrap, React.js.  IDE/Tools - Atom, Notepad++, Brackets, Eclipse, NetBeans, Excel, RapidSQL, Squirrel, Pycharm.  Databases - Oracle, DB2, MySQL, PostgreSQL. Achievements-  Won INFOSYS' Quarterly Manufacturing Unit Level award for my outstanding performance in Quarter 4, 2010.  Won Royal Bank of Scotland's monthly awards for outstanding performance during the period Aug'14 and July'15. It is a certificate of recognition of commitment, hard work and continued contribution to the business.  Won Royal Bank of Scotland's Star Team of The Month award for supporting colleagues and making a positive contribution to the business. Projects- 1. User Interface Development 2. Enterprise Application Development 3. Website Development 4. Desktop Software Development 5. Peer-to-peer application development 6. Webservices English Exam(s)- Pearson Test of English(PTE) Academic - Overall Score - 76 with 90/90 in English Writing. IELTS General - Overall Score - Band 7 with 8.5 band in English Listening.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Big Data
    API
    Database
    Spring Framework
    CSS
    Apache Tomcat
    Spring Boot
    Microservice
    Apache Hadoop
    Java
    Python
    JavaScript
  • $40 hourly
    Personal skills • Quality-oriented. I am interested in researching best practices and creating solutions that are resilient beyond the current context. In the long term, 5-stars systems pay off the extra effort. • Communicative. I don't like waiting for 3 days to receive a reply, so I don't do that to others. You can expect same day feedback from me, at the very least a "I'll get back to you later on this one". • Critical thinker: If I think you’re wrong in you're requirements, I’ll tell you and suggest alternative solutions :) Other than that, I consider myself a friendly and approachable person, who loves to help my colleagues and clients whenever I can! Technical skills —— Data Engineering • Expertise in Python: I ranked in the top 15% out of 1.3 million people on the LinkedIn Python assessment (see portfolio). • SQL • ETL/ELT with Python, Databricks (Pyspark), DBT, Dagster, Airbyte and a lot of AWS services. • Python Google Style Guide. • Agile, Extreme Programming (XP) & Clean Code (and Google Python Style Guide) —— Cloud/DevOps • AWS: Batch, Step Functions, Glue, Athena, Boto3, Lambda, S3, EC2, IAM, KMS, SQS, etc. • Bash. Docker + ECS. CI/CD - Github actions. Terraform, SAM, CodePipeline.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    ETL
    Data Engineering
    Microsoft Azure
    Azure DevOps
    Microsoft Power BI
    Terraform
    pandas
    Google Cloud Platform
    Amazon Web Services
    dbt
    PySpark
    Databricks Platform
    Python
    Docker
  • $35 hourly
    I have 18+ years of experience in software development in Telecom, Banking, and Healthcare domains. Primary skillsets include Big Data eco-systems (Apache Spark, Hive, Map Reduce, Cassandra), Scala, Core Java, Python, C++. I am well versed in designing and implementing Big data solutions, ETL and Data Pipelines, Serverless and event-driven architectures on Google Cloud Platform (GCP), and Cloudera Hadoop 5.5. I like to work with organizations to develop sustainable, scalable, and modern data-oriented software systems. - Keen eye on scalability, sustainability of the solution - Can come up with maintainable & good object-oriented designs quickly - Highly experienced in seamlessly working with remote teams effectively - Aptitude for recognizing business requirements and solving the root cause of the problem - Can quickly learn new technologies Sound experience in following technology stacks: Big Data: Apache Spark, Spark Streaming, HDFS, Hadoop MR, Hive, Apache Kafka, Cassandra, Google Cloud Platform (Dataproc, Cloud storage, Cloud Function, Data Store, Pub/Sub), Cloudera Hadoop 5.x Languages: Scala, Python, Java, C++, C, Scala with Akka and Play frameworks Build Tools: Sbt, Maven Databases: Postgres, Oracle, MongoDB/CosmosDB, Cassandra, Hive GCP Services: GCS, DataProc, Cloud functions, Pub/Sub, Data-store, BigQuery AWS Services: S3, VM, VM Auto-scaling Group, EMR, S3 Java APIs, Redshift Azure Services: Blob, VM, VM scale-set, Blob Java APIs, Synapse Other Tools/Technologies: Dockerization, Terraform Worked with different types of Input & Storage formats: CSV, XML, JSON file, Mongodb, Parquet, ORC
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    C++
    Java
    Apache Spark
    Scala
    Apache Hadoop
    Python
    Apache Cassandra
    Oracle PLSQL
    Apache Hive
    Cloudera
    Google Cloud Platform
  • $110 hourly
    Top-rated developer working (mostly) with big data, artificial intelligence, machine learning, analytics & back-end architecture. I am specialized in Bigdata (Hadoop, Apache Spark, Sqoop, Flume, Hive, Pig, Scala, Apache Kudu, kafka, python, shell scripting core Java, Machine Learning). As a Big Data architect I work as part of a team responsible for building, designing application for online analytics, Outgoing, motivated team player eager to contribute dynamic customer service, administrative, supervisory, team building, and organizational skills towards supporting the objectives of an organization that rewards reliability, dedication, and solid work ethics with opportunities for professional growth. skillSet: Hadoop,spark, scala, python, bash, Tableau,jenkins, Ansible,Hbase, Sqoop, Flume, Ne04j, Machine Learning, java, Nifi, Awz, Azure, GCP, DataBricks, DataMeer, kafka, Confluent, Schema Registry, SQl, DB2, CDC Why should you hire me ? ✅ 1400+ Upwork Hours Completed+ productive hours logged with 100% customer satisfaction » Passion for Data Engineering and Machine Learning » Experience with functional scala: shapeless, cats, itto-csv, neotypes » Familiar with Hadoop ecosystem; Apache Spark, Hive, YARN, Apache Drill, Sqoop, Flume, Zookeeper, HDFS, MapReduce, Machine Learning, airflow » Worked with JWT authentication, reactive JDBC-like connectors for PostgreSQL, MySQL & MariaDB, reactive MongoDB » Micro-services expert. Worked mostly with Lagom; Akka persistence, event-sourcing » Defining a scalable architecture on top of AWS, Google Cloud, Digital Ocean, Alibaba Cloud » ElasticSearch stack pro; ElasticSearch, Logstash, Beats, Kibana » Efficient project manager Let's discuss your idea and build the next big thing!
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Google Cloud Platform
    Apache HBase
    Snowflake
    Machine Learning
    Apache Spark MLlib
    Databricks Platform
    ETL Pipeline
    AWS Glue
    Apache Hive
    Scala
    SQL
    Docker
    Apache NiFi
    Apache Kafka
    Apache Spark
    Apache Hadoop
  • $120 hourly
    With over 12 years of experience, of which about 8yrs I have worked with different Bigdata technologies(Hadoop, Spark) and the remaining time I mostly worked on writing python scrappers, scripts, API services and also built iOS applications using Objective-C - Experience in building data pipelines to process Petabyte scale data and optimise them for cost and performance - Experience in fine tuning the Spark jobs to the most optimal level and thereby cutting down infrastructure costs by 50-80% - Experience with building Data lakes for major e-commerce and fintech companies - Worked at different startups throughout my career and highly adaptable to different working methodologies like Agile and Kanban
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Apache Spark
    Big Data
    Apache Hadoop
    PySpark
    Scala
    Python
  • Want to browse more freelancers?
    Sign up

How it works

1. Post a job

Tell us what you need. Provide as many details as possible, but don’t worry about getting it perfect.

2. Talent comes to you

Get qualified proposals within 24 hours, and meet the candidates you’re excited about. Hire as soon as you’re ready.

3. Collaborate easily

Use Upwork to chat or video call, share files, and track project progress right from the app.

4. Payment simplified

Receive invoices and make payments through Upwork. Only pay for work you authorize.

Trusted by

How do I hire a MapReduce Specialist on Upwork?

You can hire a MapReduce Specialist on Upwork in four simple steps:

  • Create a job post tailored to your MapReduce Specialist project scope. We’ll walk you through the process step by step.
  • Browse top MapReduce Specialist talent on Upwork and invite them to your project.
  • Once the proposals start flowing in, create a shortlist of top MapReduce Specialist profiles and interview.
  • Hire the right MapReduce Specialist for your project from Upwork, the world’s largest work marketplace.

At Upwork, we believe talent staffing should be easy.

How much does it cost to hire a MapReduce Specialist?

Rates charged by MapReduce Specialists on Upwork can vary with a number of factors including experience, location, and market conditions. See hourly rates for in-demand skills on Upwork.

Why hire a MapReduce Specialist on Upwork?

As the world’s work marketplace, we connect highly-skilled freelance MapReduce Specialists and businesses and help them build trusted, long-term relationships so they can achieve more together. Let us help you build the dream MapReduce Specialist team you need to succeed.

Can I hire a MapReduce Specialist within 24 hours on Upwork?

Depending on availability and the quality of your job post, it’s entirely possible to sign up for Upwork and receive MapReduce Specialist proposals within 24 hours of posting a job description.

Schedule a call