Hire the best MapReduce specialists

Check out MapReduce specialists with the skills you need for your next job.
  • $40 hourly
    I am a developer focused on providing highly efficient software solutions. - Full Stack Developer - Data Scientist
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Apache Spark
    Cloudera
    CakePHP
    Apache HBase
    Apache Hadoop
    Laravel
    Python
    PHP
    MongoDB
    JavaScript
  • $110 hourly
    Distributed Computing: Apache Spark, Flink, Beam, Hadoop, Dask Cloud Computing: GCP (BigQuery, DataProc, GFS, Dataflow, Pub/Sub), AWS EMR/EC2 Containerization Tools: Docker, Kubernetes Databases: MongoDB, Postgres-XL, PostgreSQL Languages: Java, Python, C/C++
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Apache Kafka
    Cloud Computing
    Apache Hadoop
    White Paper Writing
    Academic Writing
    Google Cloud Platform
    Dask
    Apache Spark
    Research Paper Writing
    Apache Flink
    Kubernetes
    Python
    Java
  • $25 hourly
     Certification in Big Data/Hadoop Ecosystem  Big Data Environment: Google Cloud Platform, Cloudera, HortonWorks and AWS, SnowFlake, Databricks, DC/OS  Big Data Tools : Apache Hadoop, Apache Spark, Apache Kafka, Apache Nifi, Apache Cassandra, Yarn/Mesos, Oozie, Sqoop, Airflow, Glue, Athena, S3 Buckets, Lambda, Redshift, DynamoDB ,Delta Lake, Docker, GIT, Bash Scripts Jenkins, Postgres, MongoDB, Elastic Search, Kibana, Ignite, TiDB  Certification SQL Server, Database Development and Crystal Report.  SQL Server Tools: SQL Management Studio, BIDS, SSIS, SSAS and SSRS  BI/Dashboarding Tools: Power BI, Tableau, Kibana  Big Data Development Programing Languages: Scala and python. ======================================================================= ************************************* Big Data Engineer**********************************************  Hands on experience with Google cloud platform, Big Query, Google Data Studio and Flow  Developing ETL pipeline for SQL server as well using SSIS.  For Reporting and Analysis using SSIS, SSRS and SSAS cubes.  Having amazing experience with Big data framework and open source technologies (Apache Nifi, Kafka, Spark and Cassandra, HDFS, Hive Docker/Cassandra/ Postgres SQL, Git, Bash Scripts Jenkins, MongoDB, Elastic Search, Ignite, TiDB.  Managing data warehouse Big Data cluster services and developments of Data Flows.  Writing big data/Spark ETL applications for different sources (SQL, Oracle, CSV, XML,JSON) to support different department for analytics.  Extensive work with Hive, Hadoop, Spark, Docker, Apache Nifi  Supporting different department for big data analytics.  Build multiple end to end Fraud monitoring alert based systems.  Preferable language is Scala and python as well. ************Big Data Engineer– Fraud Management at VEON *************  Devolved ETL Pipeline from Kafka to Cassandra using Spark in Scala Language.  Using Big Data Tools with Horton Works and AWS (Apache Nifi, Kafka, Spark and Cassandra, Elastic Search)  Dashboard Developments - Tableau and Kibana.  Writing SQL server complex queries, procedures and Functions.  Developing ETL pipeline for SQL server as well using SSIS.  For Reporting and Analysis using SSIS, SSRS and SSAS cubes.  Developing and designing Auto Email Reports.  Offline Data Analytics for Fraud Detection and Setting up controls for prevention.  SQL Database Development.  System Support of Fraud Management.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Google Cloud Platform
    SQL Programming
    Data Warehousing
    Database
    AWS Glue
    PySpark
    MongoDB
    Python Script
    Docker
    Apache Hadoop
    Apache Spark
    Databricks Platform
    Apache Kafka
    Apache Hive
  • $65 hourly
    A Full Stack Developer, experienced with Java, Javascript, Hadoop, C/C++, Solidity and Jasper Reports, Experienced with Solidity smart contracts and integrating DApps with different Blockchain networks. Also experienced with React and ExpressJS. Experienced with the Java language for Spring MVC and Big Data using Hadoop and Spark. Experienced with report writing using Jasper Studio.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Chatbot Development
    Dialogflow API
    Python
    ChatGPT
    API Development
    Hibernate
    Apache Hadoop
    Node.js
    React Native
    Solidity
    Java
    JavaScript
    React
  • $100 hourly
    I have over 4 years of experience in Data Engineering (especially using Spark and pySpark to gain value from massive amounts of data). I worked with analysts and data scientists by conducting workshops on working in Hadoop/Spark and resolving their issues with big data ecosystem. I also have experience on Hadoop maintenace and building ETL, especially between Hadoop and Kafka. You can find my profile on stackoverflow (link in Portfolio section) - I help mostly in spark and pyspark tagged questions.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    MongoDB
    Data Warehousing
    Data Scraping
    ETL
    Data Visualization
    PySpark
    Python
    Data Migration
    Apache Airflow
    Apache Spark
    Apache Kafka
    Apache Hadoop
  • $50 hourly
    Development experience in information management solutions, ETL processes, database design and storage systems; Responsible, able to work and solve problems independently. Software Developer, Integration process Architect Envion Software Creating a Hadoop cluster system to process heterogeneous data (ETL, Hadoop cluster, RDF/SparQL, NoSQL DB, IBM DashDB) ETL processes for big amount of database DataWarehouses creation and support Database Developer and Data Scientist A software development company Programming Analytics Stream processing Associate Professor Saint-Petersburg State University Member of the Database and Information Management Research Group
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Java
    DataTables
    Data Management
    Apache Spark
    Apache Hadoop
    Pentaho
    BigQuery
    Apache Airflow
    ETL Pipeline
    Python
    SQL
    Scala
    ETL
  • $30 hourly
    My motto is to build and innovate together! My clientele is consistently impressed by my creativity, dedication, and durability. Why I believe myself to be the best candidate and a great asset to your project. 🅐 Local Development Team Leader 🅑 Certified Full Stack Developer | React | Angular | Next | Node | MERN | MEAN | SAAS Developer 🅒 Proactive Communicator with keen eye to detail and focusing on Client Satisfaction ➢ Frontend Stack: React/Redux, JavaScript, Typescript, GraphQL, Gatsby, HTML/CSS, jQuery, Bootstrap, ECMAScript 6, HTML/CSS, Next.js, REST APIs, Three.js, React Hooks, MobX, Webpack. ➢ Backend Stack: NodeJS, Express, GraphQL, TypeScript, JavaScript, NestJS, Firebase. ➢Database: MySQL, Firebase, SQLite, PostgreSQL, MongoDB. ➢ Cloud Services: Firebase, Heroku, AWS ( EC2, S3, Amplify, Lambda, DynamoDB) ➢ Payment Providers - PayPal, Hyperpay, Stripe. ➢ Version Control - Git, Gitflow, SVN, smart GIT, Bitbucket, Gitlab, Github. I've accomplished over 50 projects. I value long-term partnerships with clients who can rely on me at any moment. Let’s connect and get you the best Web Soft product you can’t get anywhere else. Hoping to hear from you soon
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Web Development
    Amazon Web Services
    Redux
    TypeScript
    Back-End Development
    JavaScript
    CSS
    SaaS Development
    Next.js
    SQL
    DevOps
    Full-Stack Development
    MongoDB
    Node.js
    React
  • $70 hourly
    🎓 𝗖𝗲𝗿𝘁𝗶𝗳𝗶𝗲𝗱 𝗗𝗮𝘁𝗮 𝗣𝗿𝗼𝗳𝗲𝘀𝘀𝗶𝗼𝗻𝗮𝗹 with 𝟲+ 𝘆𝗲𝗮𝗿𝘀 of experience and hands-on expertise in Designing and Implementing Data Solutions. 🔥 4+ Startup Tech Partnerships ⭐️ 100% Job Success Score 🏆 In the top 3% of all Upwork freelancers with Top Rated Plus 🏆 ✅ Excellent communication skills and fluent English If you’re reading my profile, you’ve got a challenge you need to solve and you are looking for someone with a broad skill set, minimal oversight and ownership mentality, then I’m your go-to expert. 📞 Connect with me today and let's discuss how we can turn your ideas into reality with creative and strategic partnership.📞 ⚡️Invite me to your job on Upwork to schedule a complimentary consultation call to discuss in detail the value and strength I can bring to your business, and how we can create a tailored solution for your exact needs. 𝙄 𝙝𝙖𝙫𝙚 𝙚𝙭𝙥𝙚𝙧𝙞𝙚𝙣𝙘𝙚 𝙞𝙣 𝙩𝙝𝙚 𝙛𝙤𝙡𝙡𝙤𝙬𝙞𝙣𝙜 𝙖𝙧𝙚𝙖𝙨, 𝙩𝙤𝙤𝙡𝙨 𝙖𝙣𝙙 𝙩𝙚𝙘𝙝𝙣𝙤𝙡𝙤𝙜𝙞𝙚𝙨: ► BIG DATA & DATA ENGINEERING Apache Spark, Hadoop, MapReduce, YARN, Pig, Hive, Kudu, HBase, Impala, Delta Lake, Oozie, NiFi, Kafka, Airflow, Kylin, Druid, Flink, Presto, Drill, Phoenix, Ambari, Ranger, Cloudera Manager, Zookeeper, Spark-Streaming, Streamsets, Snowflake ► CLOUD AWS -- EC2, S3, RDS, EMR, Redshift, Lambda, VPC, DynamoDB, Athena, Kinesis, Glue GCP -- BigQuery, Dataflow, Pub/Sub, Dataproc, Cloud Data Fusion Azure -- Data Factory, Synapse. HDInsight ► ANALYTICS, BI & DATA VISUALIZATION Tableau, Power BI, SSAS, SSMS, Superset, Grafana, Looker ► DATABASE SQL, NoSQL, Oracle, SQL Server, MySQL, PostgreSQL, MongoDB, PL/SQL, HBase, Cassandra ► OTHER SKILLS & TOOLS Docker, Kubernetes, Ansible, Pentaho, Python, Scala, Java, C, C++, C# 𝙒𝙝𝙚𝙣 𝙮𝙤𝙪 𝙝𝙞𝙧𝙚 𝙢𝙚, 𝙮𝙤𝙪 𝙘𝙖𝙣 𝙚𝙭𝙥𝙚𝙘𝙩: 🔸 Outstanding results and service 🔸 High-quality output on time, every time 🔸 Strong communication 🔸 Regular & ongoing updates Your complete satisfaction is what I aim for, so the job is not complete until you are satisfied! Whether you are a 𝗦𝘁𝗮𝗿𝘁𝘂𝗽, 𝗘𝘀𝘁𝗮𝗯𝗹𝗶𝘀𝗵𝗲𝗱 𝗕𝘂𝘀𝗶𝗻𝗲𝘀𝘀 𝗼𝗿 𝗹𝗼𝗼𝗸𝗶𝗻𝗴 𝗳𝗼𝗿 your next 𝗠𝗩𝗣, you will get 𝗛𝗶𝗴𝗵-𝗤𝘂𝗮𝗹𝗶𝘁𝘆 𝗦𝗲𝗿𝘃𝗶𝗰𝗲𝘀 at an 𝗔𝗳𝗳𝗼𝗿𝗱𝗮𝗯𝗹𝗲 𝗖𝗼𝘀𝘁, 𝗚𝘂𝗮𝗿𝗮𝗻𝘁𝗲𝗲𝗱. I hope you become one of my many happy clients. Reach out by inviting me to your project. I look forward to it! All the best, Anas ⭐️⭐️⭐️⭐️⭐️ 🗣❝ Muhammad is really great with AWS services and knows how to optimize each so that it runs at peak performance while also minimizing costs. Highly recommended! ❞ ⭐️⭐️⭐️⭐️⭐️ 🗣❝ You would be silly not to hire Anas, he is fantastic at data visualizations and data transformation. ❞ 🗣❝ Incredibly talented data architect, the results thus far have exceeded our expectations and we will continue to use Anas for our data projects. ❞ ⭐️⭐️⭐️⭐️⭐️ 🗣❝ The skills and expertise of Anas exceeded my expectations. The job was delivered ahead of schedule. He was enthusiastic and professional and went the extra mile to make sure the job was completed to our liking with the tech that we were already using. I enjoyed working with him and will be reaching out for any additional help in the future. I would definitely recommend Anas as an expert resource. ❞ ⭐️⭐️⭐️⭐️⭐️ 🗣❝ Muhammad was a great resource and did more than expected! I loved his communication skills and always kept me up to date. I would definitely rehire again. ❞ ⭐️⭐️⭐️⭐️⭐️ 🗣❝ Anas is simply the best person I have ever come across. Apart from being an exceptional tech genius, he is a man of utmost stature. We blasted off with our startup, high on dreams and code. We were mere steps from the MVP. Then, pandemic crash. Team bailed, funding dried up. Me and my partner were stranded and dread gnawed at us. A hefty chunk of cash, Anas and his team's livelihood, hung in the balance, It felt like a betrayal. We scheduled a meeting with Anas to let him know we were quitting and request to repay him gradually over a year, he heard us out. Then, something magical happened. A smile. "Forget it," he said, not a flicker of doubt in his voice. "The project matters. Let's make it happen!" We were floored. This guy, owed a small fortune, just waved it away? Not only that, he offered to keep building, even pulled his team in to replace our vanished crew. As he spoke, his passion was a spark that reignited us. He believed. In us. In our dream. In what he had developed so far. That's the day Anas became our partner. Not just a contractor, but a brother in arms. Our success story owes its spark not to our own leap of faith, but from the guy who had every reason to walk away. Thanks, Anas, for believing when we couldn't.❞
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Solution Architecture Consultation
    AWS Lambda
    ETL Pipeline
    Data Management
    Data Warehousing
    AWS Glue
    Apache Spark
    Amazon Redshift
    ETL
    Python
    SQL
    Marketing Analytics
    Big Data
    Data Visualization
    Artificial Intelligence
  • $40 hourly
    With 6+ years of Software Engineering, AI development, ML Engineering, Web Development, Data Scientist, Scriptwriting, and Automation. Programming Languages: Java, Python, Clojure, JavaScript, Kotlin, HTML, CSS, SQL • AI & ML expertise: NLP, GAN, Recommendation Systems, Object Recognition • AI & ML tools: Pandas, TensorFlow, PyTorch, Numpy • Python: Django, Flask, Rest framework • JavaScript: TypeScript, React, Redux, Node.js • Kotlin: MVVM Design, Android Studio, Dependency Injection • Java: Spring, JavaFX • Other tools: Spark, Git, Linux, Bash I ‣ Don't give up until the client is satisfied ‣ Completed everything before the deadline. ‣ Respond quickly. ‣ Am Easy to communicate with. I am at your service, so feel free to contact me any time, I will be happy to help you out. Thanks & Regards Joy Longawis
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    DevOps
    Data Science
    UX & UI
    App Development
    Web Development
    Web Scraping
    API
    AI Development
    Artificial Intelligence
    JavaScript
    Neural Network
    Machine Learning
    Natural Language Processing
    Java
    Python
  • $45 hourly
    As a highly experienced Data Engineer with over 10+ years of expertise in the field, I have built a strong foundation in designing and implementing scalable, reliable, and efficient data solutions for a wide range of clients. I specialize in developing complex data architectures that leverage the latest technologies, including AWS, Azure, Spark, GCP, SQL, Python, and other big data stacks. My extensive experience includes designing and implementing large-scale data warehouses, data lakes, and ETL pipelines, as well as data processing systems that process and transform data in real-time. I am also well-versed in distributed computing and data modeling, having worked extensively with Hadoop, Spark, and NoSQL databases. As a team leader, I have successfully managed and mentored cross-functional teams of data engineers, data scientists, and data analysts, providing guidance and support to ensure the delivery of high-quality data-driven solutions that meet business objectives. If you are looking for a highly skilled Data Engineer with a proven track record of delivering scalable, reliable, and efficient data solutions, please do not hesitate to contact me. I am confident that I have the skills, experience, and expertise to meet your data needs and exceed your expectations.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Snowflake
    ETL
    PySpark
    MongoDB
    Unix Shell
    Data Migration
    Scala
    Microsoft Azure
    Amazon Web Services
    SQL
    Apache Hadoop
    Cloudera
    Apache Spark
  • $40 hourly
    I'm Linux DevOps and Cloud architect since 2002. Most of my professional career is with design, setup and DevOps of medium and high loaded web farms, NoSQL databases which are time-critical and require 24/7/365 uptime. During the last several years, I'm concentrated on architect & administration of Hadoop ecosystem, Big-Data systems (Cassandra, ElasticSearch, Riak ...) and distributed storage Ceph. I have big experience with a variety of web servers and load balancers (Apache, Nginx, HAProxy, Tomcat, Jetty etc .. ) as well as with cloud services such as AWS, Azure and GCP.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Big Data
    Apache HBase
    Linux System Administration
    Apache Cassandra
    Golang
    Nomad
    CI/CD Platform
    Apache Hadoop
    Consul
    Kubernetes
    Elasticsearch
    Google Cloud Platform
    Python
    Amazon Web Services
    Linux
  • $30 hourly
    Seasoned data engineer with over 11 years of experience in building sophisticated and reliable ETL applications using Big Data and cloud stacks (Azure and AWS). TOP RATED PLUS . Collaborated with over 20 clients, accumulating more than 2000 hours on Upwork. 🏆 Expert in creating robust, scalable and cost-effective solutions using Big Data technologies for past 9 years. 🏆 The main areas of expertise are: 📍 Big data - Apache Spark, Spark Streaming, Hadoop, Kafka, Kafka Streams, HDFS, Hive, Solr, Airflow, Sqoop, NiFi, Flink 📍 AWS Cloud Services - AWS S3, AWS EC2, AWS Glue, AWS RedShift, AWS SQS, AWS RDS, AWS EMR 📍 Azure Cloud Services - Azure Data Factory, Azure Databricks, Azure HDInsights, Azure SQL 📍 Google Cloud Services - GCP DataProc 📍 Search Engine - Apache Solr 📍 NoSQL - HBase, Cassandra, MongoDB 📍 Platform - Data Warehousing, Data lake 📍 Visualization - Power BI 📍 Distributions - Cloudera 📍 DevOps - Jenkins 📍 Accelerators - Data Quality, Data Curation, Data Catalog
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    SQL
    AWS Glue
    PySpark
    Apache Cassandra
    ETL Pipeline
    Apache Hive
    Apache NiFi
    Apache Kafka
    Big Data
    Apache Hadoop
    Scala
    Apache Spark
  • $40 hourly
    🔍🚀 Welcome to a world of data-driven excellence! 🌐📊 Greetings, fellow professionals! I am thrilled to introduce myself as a dedicated Data Consultant / Engineer, leveraging years of honed expertise across a diverse spectrum of data stacks 🌍. My journey has been enriched by a wealth of experience, empowering me with a comprehensive skill set that spans Warehousing📦, ETL⚙, Analytics📈, and Cloud Services☁. Having earned the esteemed title of GCP Certified Professional Data Engineer 🛠, I am your partner in navigating the complex data landscape. My mission is to unearth actionable insights from raw data, shaping it into a strategic asset that fuels growth and innovation. With a deep-rooted passion for transforming data into valuable solutions, I am committed to crafting intelligent strategies that empower businesses to flourish. Let's embark on a collaborative journey to unlock the full potential of your data. Whether it's architecting robust data pipelines ⛓, optimizing storage solutions 🗃, or designing analytics frameworks 📊, I am dedicated to delivering excellence that transcends expectations. Reach out to me, and together, let's sculpt a future where data powers success. Thanks!
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    PySpark
    Machine Learning
    Natural Language Processing
    Informatica
    Data Science
    Data Warehousing
    Snowflake
    Data Analysis
    Big Data
    BigQuery
    ETL
    Apache Airflow
    Apache Hadoop
    Apache Spark
    Databricks Platform
    Python
    Apache Hive
  • $80 hourly
    A Backend Software Engineering with more than 6 years of experience. Have worked with large-scale backend/distributed systems and big data systems. A DevOps engineer with 4 years of experience - both on-premises and AWS, experienced with K8s, Terraform, Ansible, CI/CD. Currently working as Principal Engineer/ Solution Architect role.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Architectural Design
    GraphQL
    Serverless Computing
    Amazon Web Services
    DevOps
    API Development
    Elasticsearch
    Apache Kafka
    Scala
    Apache Spark
    Docker
    Apache Hadoop
    Kubernetes
  • $50 hourly
    Experienced IT professional with a demonstrated history of working in the Cloudera industry. Skilled in IT Service Management, AWS Cloud, Linux, Hadoop, Python, Databases. Strong information technology professional with more than 5+ years of experience. If you have any queries, please feel free to contact me at any time. Thank you for taking the time to visit my profile and I look forward to hearing from you! Thanking With Regards, Mohammed Ismail
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Microsoft Power BI Data Visualization
    Kerberos
    Microsoft SQL Server Administration
    Hive Technology
    Microsoft SQL Server
    Administrative Support
    Apache Kafka
    Cloudera
    Apache Hadoop
    Apache Hive
    Apache Spark
  • $20 hourly
    • Having 6+ years of experience as a Hadoop/Pyspark Developer. • Having extensive knowledge on Hadoop technology experience in Storage, writing Queries, processing, and analysis of data. • Experience on migrating on Premises ETL process to Hadoop Layer. • Experience in optimizing Hive SQL queries and Spark Jobs. • Implemented various frameworks like Data Quality Analysis and Data Validation with the help of technologies like Bigdata, Spark, Python. • Primary technical skills in Pyspark, HDFS, YARN, Hive, Sqoop, Impala, Oozie. • Good exposure on advanced topics Like Analytical Functions, Indexes, Partition Tables. • Experience with creation of Technical document for Functional Requirement, Impact Analysis, Technical Design documents, Data Flow Diagram. • Quick learner and up to date with industry trends, Excellent written and oral communications, analytical and problem-solving skills and good team player, Ability to work independently and well-organized.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    PySpark
    Apache Impala
    Sqoop
    Python Script
    Apache Hadoop
    Apache Spark
    SQL
    Python
    Apache Hive
    Apache Airflow
  • $25 hourly
    I am a data engineer expert with over than 5 years experience in data ingestion, integration and manipulation. Till date, I have done many projects in data engineering and big data. I worked on business analytics and telco analytics, i used multy data platforms and framework such as Cloudera data platform, Nifi, R studio, Spark, Hadoop, Kafka ... If this is what you want, then get in touch with me
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Cloud Engineering
    Cloudera
    Apache Hadoop
    Data Warehousing
    Apache NiFi
    Linux
    Apache Spark
    Data Lake
    Data Analysis
    SQL
    Big Data
    Business Intelligence
    Scala
    Apache Hive
    Python
  • $50 hourly
    Raghu is an IT expert, specializes in Amazon web service consulting , developing serverless application on AWS and migrating business environments, applications and data centers to the AWS cloud. He is a certified Solution Architect and his toughest work is migrating telecom infra to cloud. He can help in solving problem using opensource tools and specializes in freeRADIUS, Squid, strongswan, openVPN and few more. Contact Raghu, if you need highly recommended AWS certified Architect to evaluate your needs, make the move to the cloud, and vastly improve your technology environment. Raghu had used below AWS services to build his customer infra, Serverless AWS API Gateway AWS Lambda Amazon DynamoDB Analytics Amazon Kinesis Amazon Redshift AWS Glue Amazon QuickSight Application Integration SQS SNS Compute Amazon EC2 Amazon EC2 Auto Scaling Amazon Elastic Container Service AWS Elastic Beanstalk AWS Lambda Machine Learning Amazon SageMaker Amazon Comprehend Amazon Lex Amazon Polly Amazon Rekognition Migration and Transfer AWS Database Migration Service AWS Transfer for SFTP Networking and Content Delivery Amazon VPC Amazon CloudFront Amazon Route 53 Elastic Load Balancing Security Identity and Compliance AWS Key Management Service AWS Shield AWS Single Sign-On AWS WAF Storage Amazon Simple Storage Service (S3) Amazon Elastic Block Store (EBS) Amazon Elastic File System (EFS) Amazon S3 Glacier AWS Backup AWS Storage Gateway Database Amazon Aurora Amazon RDS Amazon Redshift AWS Database Migration Service Management & Governance Amazon CloudWatch AWS Auto Scaling AWS CloudFormation AWS CloudTrail AWS Command Line Interface AWS Management Console AWS Organizations Customer Engagement Amazon Simple Email Service (SES)
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Amazon S3
    Amazon Athena
    Backup & Migration
    AWS Glue
    Internet of Things Solutions Design
    Amazon Cognito
    Amazon API Gateway
    Serverless Computing
    AWS Server Migration
    AWS CloudFront
    API Integration
    AWS Lambda
    Amazon Web Services
    Amazon DynamoDB
    Amazon Aurora
  • $95 hourly
    => Let's Connect Hello, I'm Dima, a seasoned CyberSecurity Specialist and Turnkey Infrastructure Expert specializing in BigData solutions and data analysis, utilizing a DevOps approach. => Expertise Overview With a robust passion for constructing SOC, SOAR, and SIEM solutions, my primary focus lies in developing data ingestion, enrichment, and analysis pipelines, ensuring they are highly available and fault-tolerant. My expertise extends to building central logging and real-time processing platforms from the ground up, optimizing them for performance, security, and reliability across multiple environments, whether in the cloud or on-premise. => Value Proposition My commitment is to deliver solutions that not only centralize security and threat intelligence but also facilitate enhanced control over data, ultimately contributing to infrastructure cost savings. => Technological Summary CyberSecurity:------- > Wazuh, Suricata, pfSense BigData:--------------- > Kafka, ElasticSearch, OpenSearch Data Processing:----- > FluentD, Vector.dev, Apache NiFi Infra as Code:--------- > Terraform, cdktf, cdk8s Virtualization:--------- > Proxmox, VMware Containerization:----- > Kubernetes Clouds:---------------- > AWS, Hetzner, DigitalOcean, Linode Automation:----------- > Jenkins, GitHub Actions Monitoring:----------- > Zabbix, Grafana, Kibana, Prometheus, Thanos Mail:--------------------> MailCow SMTP/IMAP, Postfix VPN:------------------- > OpenVPN Server Programming:-------- > Bash, Python, TypeScript Operating Systems:- > CentOS, RHEL, Rocky Linux, Ubuntu, Debian => Personal Attributes • Leadership: Leading by example with a team-first approach • End-to-End Execution: Proficient from POC to Enterprise-level implementation • Resilience: Demonstrating high thoroughness and endurance • Adaptability: A quick, can-do architect and experienced troubleshooter • Optimization: Adept in process and performance optimization • Documentation: Skilled technical documentation writer • Vision: A visionary in technological implementation and solution provision
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Elasticsearch
    Linux System Administration
    Apache Kafka
    Apache Hadoop
    Email Security
    Machine Learning
    ELK Stack
    Cloudera
    Zabbix
    MySQL
    Big Data
    Apache NiFi
    PfSense
    Red Hat Administration
    Proxmox VE
    Amazon Web Services
  • $25 hourly
    Good Day Over 9+ years of extensive hands-on experience in Big Data technologies from core Hadoop Eco System to GCP-AWS cloud-based platforms. Expertise in Cloud (GCP, AWS) and In-Premise (Hadoop) systems and its various components Experience working in different Google Cloud Platform Technologies like Big Query, Dataflow, Dataproc, Pub sub, Composer and AWS like EMR, Redshift, Lambda, Step Functions, EKS over open source Ecosystem Hadoop, HDFS, MapReduce, Kafka, Spark, Hive. - Design and Development of Ingestion Framework over Google Cloud, AWS and Hadoop cluster. - Good Knowledge on Hadoop Cluster architecture and monitoring. - Extensive Experience on importing and exporting data using Kafka. - Strong hands-on experience in ETL processing using Spark, Scala/Python and Kafka. - Integration various data science models into Data Engineering platform over cloud and in-premise - End-to-End Bigdata platform setup over In-premise to cloud - Migration of tradition data systems into cost-friendly, reliable, scalable data systems - Developing and scheduling ETL workflows in Hadoop using Oozie, Airflow, Google Cloud Composer - Setup, manage and optimize Distributed data warehouses like Hive, Big Query, Redshift - Managing different queues over pub/sub, Kafka - Handled various frequency based ingestion with Realtime, Near Realtime, Scheduled batch flow - Handling integration with RDBMS like MSSQL, MySQL and NOSQLs like MongoDB, Elasticsearch - Complete Data driven system build experience from Data Ingestion, Transformation, Store and Analytics over BI platforms like PowerBI, DataStudio - Managing various loggings using ELK (Elasticsearch, Logstash and Kibana) - Setting up various webserver configurations - Setting and managing Devops pipelines over Kubernetes, Docker, Azure Devops, Github devops
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    AI Content Creation
    Database
    Python Script
    Machine Learning
    Flask
    API
    Data Migration
    Apache Airflow
    ETL
    Apache Kafka
    Google Cloud Platform
    Python
    SQL
    Amazon Web Services
    JavaScript
  • $55 hourly
    I am a data engineer with 8 years of experience designing and implementing big data solutions. My expertise lies in Apache Spark, the Hadoop ecosystem, NoSQL databases, and writing ETLs using Python, Scala, and Java. Also, an avid Go developer who specializes in creating microservices using this programming language. I have experience in migrating large enterprises from legacy software to Go-based microservices My educational background is in Computer Science, and I began my career as a software developer before transitioning to big data. In my current role as a Data Engineer, I am responsible for building data pipelines, processing and analyzing large-scale data sets, and creating data models to support data-driven decision-making. I have a strong understanding of distributed computing and extensive experience in optimizing data processing workflows for performance and scalability. I have worked on a variety of projects, including real-time data processing, data warehousing, and building streaming data pipelines. My technical skills include expertise in programming languages such as Python, Scala, Java, and technologies such as Apache Spark, Hadoop, and NoSQL databases like MongoDB and Cassandra.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Software Architecture & Design
    Amazon Web Services
    Golang
    Microservice
    Play Framework
    API Integration
    Scala
    ETL Pipeline
    Serverless Computing
    Big Data
    Apache Spark
    Apache Hadoop
    Python
  • $30 hourly
    With close to 10 years of industry experience in the specialised field of software design and development, I possess proven capabilities to develop high quality software applications. My aim is to obtain a challenging position that will utilise my skills and experiences and which will also provide me with the opportunity for growth and advancement. Languages- Java, Python, Javascript. Skills- Core: Data Structures and Algorithms. Data Analysis: Hadoop MapReduce Backend: Java, Spring, Spring Boot, Microservices, Struts, Design Principles, Design Patterns, SQL, Webservices, SOA(REST and SOAP), JMS, Servlets, Swing, JSP, MAVEN, Sub versioning(svn, git), Jenkins. Frontend: HTML5, CSS3, Javascript, Jquery, Bootstrap, React.js.  IDE/Tools - Atom, Notepad++, Brackets, Eclipse, NetBeans, Excel, RapidSQL, Squirrel, Pycharm.  Databases - Oracle, DB2, MySQL, PostgreSQL. Achievements-  Won INFOSYS' Quarterly Manufacturing Unit Level award for my outstanding performance in Quarter 4, 2010.  Won Royal Bank of Scotland's monthly awards for outstanding performance during the period Aug'14 and July'15. It is a certificate of recognition of commitment, hard work and continued contribution to the business.  Won Royal Bank of Scotland's Star Team of The Month award for supporting colleagues and making a positive contribution to the business. Projects- 1. User Interface Development 2. Enterprise Application Development 3. Website Development 4. Desktop Software Development 5. Peer-to-peer application development 6. Webservices English Exam(s)- Pearson Test of English(PTE) Academic - Overall Score - 76 with 90/90 in English Writing. IELTS General - Overall Score - Band 7 with 8.5 band in English Listening.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Big Data
    API
    Database
    Spring Framework
    CSS
    Apache Tomcat
    Spring Boot
    Microservice
    Apache Hadoop
    Java
    Python
    JavaScript
  • $55 hourly
    I focus on data engineering, software engineering, ETL/ELT, SQL reporting, high-volume data flows, and development of robust APIs using Java and Scala. I prioritize three key elements: reliability, efficiency, and simplicity. I hold a Bachelor's degree in Information Systems from Pontifícia Universidade Católica do Rio Grande do Sul as well as graduate degrees in Software Engineering from Infnet/FGV and Data Science (Big Data) from IGTI. In addition to my academic qualifications I have acquired a set of certifications: - Databricks Certified Data Engineer Professional - AWS Certified Solutions Architect – Associate - Databricks Certified Associate Developer for Apache Spark 3.0 - AWS Certified Cloud Practitioner - Databricks Certified Data Engineer Associate - Academy Accreditation - Databricks Lakehouse Fundamentals - Microsoft Certified: Azure Data Engineer Associate - Microsoft Certified: DP-200 Implementing an Azure Data Solution - Microsoft Certified: DP-201 Designing an Azure Data Solution - Microsoft Certified: Azure Data Fundamentals - Microsoft Certified: Azure Fundamentals - Cloudera CCA Spark and Hadoop Developer - Oracle Certified Professional, Java SE 6 Programmer My professional journey has been marked by a deep involvement in the world of Big Data solutions. I've fine-tuned my skills with Apache Spark, Apache Flink, Hadoop, and a range of associated technologies such as HBase, Cassandra, MongoDB, Ignite, MapReduce, Apache Pig, Apache Crunch and RHadoop. Initially, I worked extensively with on-premise environments but over the past five years my focus has shifted predominantly to cloud based platforms. I've dedicated over two years to mastering Azure and I’m currently immersed in AWS. I have a great experience with Linux environments as well as strong knowledge in programming languages like Scala (8+ years) and Java (15+ years). In my earlier career phases, I had experience working with Java web applications and Java EE applications, primarily leveraging the WebLogic application server and databases like SQL Server, MySQL, and Oracle.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Scala
    Apache Solr
    Apache Kafka
    Apache Spark
    Bash Programming
    Elasticsearch
    Java
    Progress Chef
    Apache Flink
    Apache HBase
    Apache Hadoop
    MongoDB
    Docker
  • $30 hourly
    I hold a Masters' degree in Data Science and a MicroMasters in the same from UCSD. I worked as a Machine Learning Engineer for about three years in the Data Science Industry before starting my professional freelance career. I have worked with various high-profile multinational firms on various Data Science projects and have provided successful consultancies for different Big Data solutions. Some of the clients that I have worked with include: ✅ Mercy Hospitals - 4th largest Hospital Chain in the USA ✅ KPMG - One of the Big Four accounting organizations ✅ ComfortDelgro, Singapore (one of the largest taxi service providers in Asia) ✅ Telenor Pakistan - Second-largest cellular & digital services provider in Pakistan ✅ EZLink - The predominant public transit card in Singapore; and more Please see my Projects section for more insights into what the projects were. In addition to the practical work, I have also provided successful consultancies for various Big Data Solutions to a large Health Care Enterprise in the US with very positive Gartner reviews. I am very skilled in the following areas and tools in the domain of Data Science (which contains Machine Learning, Big data analytics, and Data Engineering as core fields): ⭐ Python (Basic to Advanced) ⭐ DS & ML Algorithms (Basic to Advanced), a few of them include: ---------✨ Regression (e.g. Linear Regression, Logistic Regression, Least-Squares, and more) ---------✨ Instance-based (e.g. K-Nearest Neighbor (KNN), Support Vector Machines (SVM)) ---------✨ Regularization Algorithms (e.g. Ridge Regression, LASSO, and more) ---------✨ Decision Tree Algorithms (e.g. CART, C5.0, Chi-squared, and more) ---------✨ Bayesian Algorithms (e.g. Naive Bayes, Multinomial Naive Bayes, and more) ---------✨ Clustering Algorithms (e.g K-Means, k-Medians, EM, and more) ---------✨ Association Rule Learning Algorithms (e.g. Apriori Algorithm) ---------✨ Artificial Neural Network Algorithms (e.g. Single Layer MLP) ---------✨ Deep Learning Algorithms (e.g. CNNs, RNNs, LSTMS, Auto-Encoders, and more) ---------✨ Ensemble Algorithms (e.g. Boosting, Bagging, AdaBoost, Random Forest, etc.) ---------✨ Dimensionality Reduction Algorithms (e.g. PCA, LDA, MDA, and more) ---------✨ And more... ⭐ Specialty subfields of Machine Learning ---------✨ Computer Vision (CV) - (e.g. Object Recognition, Pose-Estimation, etc.) ------------------🌠 Image Classification ------------------🌠 Object Detection ------------------🌠 Object Tracking ------------------🌠 Semantic Segmentation ------------------🌠 Instance Segmentation ------------------🌠 Image Reconstruction or Super Resolution *(See my research paper)* ------------------🌠 and more... ---------✨ Natural Language Processing (NLP) ------------------🌠 Tokenization ------------------🌠 Part of Speech (POS) Tagging ------------------🌠 Named Entity Recognition (NER) ------------------🌠 Sentiment Analysis ------------------🌠 Categorization and Classification ------------------🌠 Chatbots ------------------🌠 Keyword Spotting ------------------🌠 and many more... ---------✨ Recommender Systems ---------✨ Speech Processing *(See my Masters' Thesis)* ---------✨ And more... ⭐ Advanced Visualizations using ---------✨ Tableau ---------✨ Python - matplotlib, seaborn, ggplot, plotly, and more ---------✨ And more... ⭐ Pyspark for Distributed Analytics and Distributed Machine Learning ⭐ Data Lake on Apache Hadoop ⭐ Enterprise Data Warehouse on Apache Hive ⭐ ETL using ---------✨ Cron (Bash) ---------✨ Talend ---------✨ SLJM ⭐ CRUD operations using SQL ⭐ Data Modelling using SQL ⭐ Git I am also an experienced researcher and I provide corporate training in Artificial Intelligence (AI) from time to time. I love to work on managed services provided by Cloud Platforms such as GCP, Azure, and AWS. Big Data on the cloud is the next big thing. I also provide consultancy on Data Architectures/Data Analytics Platforms for companies to migrate towards the Cloud or build a Distributed Analytics Platform on-premise.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Apache Hadoop
    Google Cloud Platform
    Data Visualization
    Cloud Computing
    Big Data
    SQL
    Data Science
    Machine Learning
    Apache Spark
    Computer Vision
    Natural Language Processing
  • $18 hourly
    🏆 M.S. in Big Data Analytics & Artificial Intelligence 🏆 M.S. in Business Consulting 🏆 50+ projects completed I'm Rohan - Your Tech Bestie. With my technical skills and business acumen, I'll understand your needs, solve the problems, and scale your business. I follow a professional, friendly, and collaborative approach to achieve the outcomes with the highest quality. Here is what my valuable clients have to say about me: ✅ "Rohan is a very good freelancer! He knows his job, has very quick communication and nice quality. Recommended!" ✅ "Rohan was very responsive and easy to work with." ✅ "Great to work with Rohan. He is completed all the tasks on time and is a good in tech skills." ✅ "Rohan works quickly and communicates his limits efficiently." ✅ "Rohan was easy to work with, he communicated well and did an outstanding job. He was efficient and great at trouble shooting to get the job done. I would definitely recommend him highly and I will also rehire him for future jobs." ✅ "Fast turn around time and good quality of work and good english writing skills. Easy to communicate with. Recommended." 👉 Send me a message about your vision, and I'll reply with my best advice or refer you in the right direction.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Neural Network
    Model Optimization
    Model Tuning
    pandas
    Artificial Intelligence
    Natural Language Processing
    Analytics
    Research Papers
    Quantitative Analysis
    Keras
    Machine Learning
    Data Visualization
    Python
    Business Intelligence
  • $45 hourly
    I am a professional software developer with a degree in Computer Systems Engineering. My experience includes building everything from web applications to Artificial Intelligence-based systems with the latest technologies, mostly Python and Node JS. My history and review reflects my performance working with the top Enterprise Clients on Upwork. My communication in English is fluent, so it won't be a barrier at all. Following are the areas that I am proficient at: **** Web apps ✅ Figma/PSD to HTML ✅ Angular ✅ React JS ✅. Vue JS ✅. Django ✅ Flask ✅. FAST API ✅. HTML, CSS, Javascript, jQuery ✅. Express JS ✅. Node JS ✅ Databases (MySQL, SQLite, Postgresql, MongoDB, Firebase) ✅ Web Scraping using Selenium, Beautifulsoup, and Scrapy ✅ REST APIs/ GraphQL ✅ Unit testing using Pytest, JEST ✅. Elasticsearch **** Data Engineering ✅. Ingest data from data sources. ✅. Databases (Redshift, Bigquery, PostgreSQL, MySQL, MongoDB, etc). ✅. Build and maintain Data warehouse. ✅. Schedule and automate ETL pipelines using Airflow. ✅. Optimize queries. ✅. BI tools/dashboards **** Artificial Intelligence ✅. Open AI ✅. GPT3 and GPT4 prompting and integrations ✅. Machine Learning ✅. NLP using spaCy and NLTK ✅. Deep Learning ✅. Neural Networks ✅. Time-series Analysis ✅. Recommendation Engine **** 3rd Party APIs: ✅. Telegram ✅. Paypal, Stripe, and many other payment platforms. ✅. Instagram, Twitter and FB ✅. Twilio ✅. Stock exchange APIs **** Deployment ✅. AWS ✅. Lambda Functions ✅. Serverless Architecture ✅. Digitalocean ✅. Heroku ✅. GCP ✅. Docker As a top-rated seller, I always try to meet the client's satisfaction with my top-quality work. Feel free to reach out anytime.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Data Scraping
    RESTful API
    Web Development
    Data Management
    Django
    Node.js
    GraphQL
    Database Maintenance
    Automated Deployment Pipeline
    React
    ETL Pipeline
    Flask
    Data Science
    Machine Learning
    Python
  • $80 hourly
    Hi, I am Data Architect/Snr. Data Engineer with 10 years experience with RDBMS/NoSQL databases and processing large amounts of data. My experience related to Enterprise level/High profile projects in the past, but now I'm helping alot startups and small-mid sized companies. My core competences are : Data Modelling, Data Architecture on Cloud platforms, Database development, ETL and Business Intelligence, Database Administration Modelling of OLTP and Datawarehouse systems. It could be design of new schema, normalization/denormalization of existing model, Enterprise datawarehouse design based on Kimball/Inmon, Data Lake and Data Vault architectures, Modernization of existing Datamodel(s). DBA Activities - DB migrations, Backup & Recovery, Upgrades, Instance configurations, DB Monitoring, Horizontal scaling, Streaming/BDR replications. Sharding with postgreSQL extensions. Data Integration and ETL : Traditional batch ETL - Informatica, Talend, AWS Datapipeline, Matillion ETL Serverless ETL - AWS Lambda, Glue, Batch, AWS DMS, Google Cloud Functions Streaming ETL - Apache NiFi, Kafka, Kinesis streams SaaS ETL - Stitch, Alooma, Fivetran Direct loading with DBMS tools & scripting Building BI layer with Crystal reports , Tableau/QlickSense or other modern BI SaaS tool. Cloud containerization and deployment : Docker, Mesos/Kubernetes Java development : EE/SE , Spring, Hibernate, RESTful APIs, Maven Clouds : - Cloud migrations (AWS, Azure, GCP) - Cloud infrastructures (VPCs, EC2, Loadbalancing, Autoscaling, Security in AWS/GCP/Azure) - Processing in EMR Hadoop/HDInsight/Azure DF/Google PubSub - Athena, DynamoDB/Cosmos DB, Amazon Aurora - Development & Administration of RDS/Azure SQL/GCP databases - Building Analytics Solutions in Amazon Redshift/Azure PDW/Google Bigquery/Snowflake with End-to-End BI implementations Thank you for getting to the end of this boring details and looking forward working on exciting projects together :) Best Regards, Yegor.
    vsuc_fltilesrefresh_TrophyIcon MapReduce
    Oracle Database Administration
    Amazon EC2
    Amazon RDS
    Amazon Web Services
    Amazon Redshift
    Tableau
    Oracle Performance Tuning
    PostgreSQL Programming
    Oracle PLSQL
    ETL
  • Want to browse more freelancers?
    Sign up

How it works

1. Post a job (it’s free)

Tell us what you need. Provide as many details as possible, but don’t worry about getting it perfect.

2. Talent comes to you

Get qualified proposals within 24 hours, and meet the candidates you’re excited about. Hire as soon as you’re ready.

3. Collaborate easily

Use Upwork to chat or video call, share files, and track project progress right from the app.

4. Payment simplified

Receive invoices and make payments through Upwork. Only pay for work you authorize.

Trusted by

How do I hire a MapReduce Specialist on Upwork?

You can hire a MapReduce Specialist on Upwork in four simple steps:

  • Create a job post tailored to your MapReduce Specialist project scope. We’ll walk you through the process step by step.
  • Browse top MapReduce Specialist talent on Upwork and invite them to your project.
  • Once the proposals start flowing in, create a shortlist of top MapReduce Specialist profiles and interview.
  • Hire the right MapReduce Specialist for your project from Upwork, the world’s largest work marketplace.

At Upwork, we believe talent staffing should be easy.

How much does it cost to hire a MapReduce Specialist?

Rates charged by MapReduce Specialists on Upwork can vary with a number of factors including experience, location, and market conditions. See hourly rates for in-demand skills on Upwork.

Why hire a MapReduce Specialist on Upwork?

As the world’s work marketplace, we connect highly-skilled freelance MapReduce Specialists and businesses and help them build trusted, long-term relationships so they can achieve more together. Let us help you build the dream MapReduce Specialist team you need to succeed.

Can I hire a MapReduce Specialist within 24 hours on Upwork?

Depending on availability and the quality of your job post, it’s entirely possible to sign up for Upwork and receive MapReduce Specialist proposals within 24 hours of posting a job description.

Schedule a call