Hire the best Apache Spark Engineers in Mumbai, IN

Check out Apache Spark Engineers in Mumbai, IN with the skills you need for your next job.
Clients rate Apache Spark Engineers
Rating is 4.7 out of 5.
4.7/5
based on 283 client reviews
  • $35 hourly
    Academician, Freelance Machine learning and coding instructor, researcher ,visiting lecturer researcher at university of Mumbai. I have more than 15 years of teaching experience in College ,University ,institutes and corporates
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    GitHub
    HTML
    Python Script
    Oracle PLSQL
    PySpark
    Content Writing
    C#
    JavaScript
    SQL Programming
    SQL
    R
    Data Science
    Python
    Scala
    Machine Learning
  • $30 hourly
    I'm a dynamic data expert with proven ability to deliver short or long-term projects in data engineering, data warehousing, machine learning, and business intelligence realm. My passion is to partner with my clients to deliver top-notch, scalable data solutions to provide immediate and lasting value. I specialize in the following data solutions: ✔️ Data strategy advisory & technology selection/recommendation ✔️ Building data warehouses using modern cloud platforms and technologies ✔️ Creating and automating data pipelines, real-time streaming & ETL processes ✔️ Building highly intuitive, interactive dashboards. ✔️ Data Cleaning, Processing, and Machine Learning models ✔️ Data Migration (Heterogenous and Homogenous) Some of the technologies I most frequently work with are: ☁️ Cloud: AWS, GCP & Azure 👨‍💻 Databases: SQL Server, Redshift, BigQuery, Snowflake, RDS, PostgreSQL, MySQL, S3, DynamoDB, MongoDB, Cloud Data Store, Redshift ⚙️ Data Integration/ETL: Talend, Stitch, Informatica, SSIS, AWS Glue & EMR, Alteryx, GCP DataFlow & DataProc 📊 BI/Visualization: Tableau, PowerBI, Spotfire, Excel, Google Data Studio, AWS QuickSight 🤖 Machine learning - Natural Language Processing, Keras, Jupyter Notebook, Python, TensorFlow, Pandas, Numpy, Pytorch, JS. ==What my clients say about me== ------------------------------------------------------ "Darshil is a python savvy, technically sound resource. He is very good at Pyspark and is good at understanding complex ETL pipelines. It was an amazing experience working with him and would love to work with him again." ⭐⭐⭐⭐⭐ ------------------------------------------------------ "Darshil delivered on all of the project requirements and went above and beyond my expectations. He makes himself available to answer my questions, as basic as some of them are. I've learned a great deal from just one project, and I look forward to working with him some more." ⭐⭐⭐⭐⭐ ------------------------------------------------------ I am highly attentive to detail, organized, efficient, and responsive. Let's get to work! 💪
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    API
    Data Extraction
    Amazon Web Services
    Google Cloud Platform
    Data Management
    Data Visualization
    Amazon Redshift
    BigQuery
    SQL
    Machine Learning
    Natural Language Processing
    ETL Pipeline
    Looker Studio
    Python
  • $40 hourly
    Hello, I'm Saurabh 👋🏻 I have 2+ years of experience in the data field. I have helped businesses set up efficient data pipelines and improve reporting by 50%. I have participated in 6 data hackathons and won 4 of them. I have a proven track record of success in helping businesses to: ✅Improve their data quality and accessibility. ✅Build and maintain scalable data pipelines. ✅Develop and deploy machine learning models. ✅Create and share insights with stakeholders. I am also an expert in data visualization and storytelling, and I can help you communicate your data findings clearly and concisely. Here are some specific examples of the types of projects that I can help you with: ✅Building and maintaining data pipelines to ingest and process data from various sources. ✅Designing and implementing data warehouses to store and manage data at scale. ✅Creating and sharing data insights with stakeholders through dashboards, reports, and presentations. I'm proficient in the following tools and technologies: -- Languages: Python, Kafka, PySpark, SQL -- Databases: PostgreSQL, BigQuery, SQLite, MySQL -- Transformation: DBT (data build tool) -- ETL Tool: Airbyte -- Orchestration: Airflow -- BI Tool: Looker, Tableau, Preset -- Cloud: Google Cloud Platform -- DevOps: Bash, Docker, Docker Compose, GitHub actions -- Misc: Agile, Clean Code If you are looking for an Analytics Engineer, Contact me today to learn more about how I can help you make better decisions with data.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    ETL Pipeline
    dbt
    A/B Testing
    Data Modeling
    CI/CD
    Docker
    Google Cloud Platform
    Apache Airflow
    Apache Kafka
    BigQuery
    Snowflake
    Tableau
    SQL
    Python
  • $35 hourly
    I am developer predominantly working on data engineering domain. Proficient in creating data models and data pipeline needed for business analytics needs
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    PostgreSQL
    Git
    ETL Pipeline
    Python
  • $30 hourly
     Highly Skilled IT Professional Experience in Azure / GCP Cloud with over 3+ years of experience working as Azure/GCP Data Engineer.  Overall, 14+ years in IT Experience. In Software Design, Development, Analysis, Testing, Data Warehouse and Business Intelligence tools.  Working within an Agile delivery methodology production implementation in iterative sprints Cloud Data Platform (Azure):  Implemented standard Databricks Notebook used to load Full load and Delta load (Type 1) to process a large volume of data with all the business rules and transformation.  Used Scala and Spark SQL language to create standard Databricks Notebook  Implemented Azure Key Vault service to store all the vital credentials like Service Principal key, Database credentials, Storage connection string, and others  Knowledge of ADLA (U-SQL) replacement with Azure Databricks for data processing  Designed and implemented highly performant data ingestion ADF pipelines from multiple sources using Azure Databricks  Created some UDF for logging, History load from multiple date folders from data lake storage GCP : BigQuery – Created external table on GCS Parque files, Views with Dedupe logic, Wrote the dynamic batch script to find the current/invalid parquet files in GCS 🖎 DataProc – Written in PySpark / Spark SQL program to transform data and used dataproc cluster to run the job. Implemented Delta lake 🖎 Composer, Apache Airflow – Used for workflow orchestrations. 🖎 Git – Used to maintain as a code repository 🖎 GCS – Storage used to keep processed data. Parquet/CSV files used to store data. 🖎 Programming – PySpark, Python, and Spark SQL used for script DW& Data Modeling:  Experience in OLTP/OLAP System Study, developing Database Schemas like Star Schema and Snowflake Schema used in relational, dimensional, and multidimensional modeling.  Experience in analysis, design, and construction of Data warehouses.  Expert Experience in Normalization, De-normalization.  Implemented Slowly Changing Dimensions - Type I & II in Dimension tables as per the requirements ETL:  Proficient in using SQL Server Integration Services to build Data Integration and Workflow Solutions, Extract, Transform and Load (ETL) solutions for Data warehousing applications.  Skilled in Business Intelligence tools like SQL Server 2008R2 Integration Services (SSIS).  Experience creating SSIS packages to automate the Import and Export of data to and from SQL Server 2008 using SSIS tools like Import and Export Wizard, Package Installation, and BIDS. T-SQL:  Extensive experience in using T-SQL (DML, DDL) in SQL Server 2012 / 2000 platforms.  Experienced in creating Tables, Stored Procedures, Views, Indexes, Cursors, Triggers, User Profiles, User Defined Functions, Relational Database Models, and Data Integrity in observing Business Rules.  Implemented Change Tracking (CT) and Change Data Capture (CDC) functionality  Extensive knowledge in tuning T-SQL, Query Optimization to improve the Stored Procedures/Functions performance and availability.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    BigQuery
    Apache Airflow
    Query Tuning
    Microsoft Azure SQL Database
    Google Cloud Platform
    Big Data
    Python
    Microsoft Azure
    PySpark
    API Integration
    Databricks Platform
    Data Warehousing
    SQL
    Data Migration
  • $30 hourly
    I am highly experienced in Big data technologies and Data analysis using Spark, Python and SQL. For this regular communication is really important to me. I already manage a complete project on above mentioned tools.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Apache Hadoop
    Big Data
    Data Analysis
    Tableau
    Seaborn
    Matplotlib
    NumPy
    PySpark
    pandas
    Python
    SQL
  • $22 hourly
    Summary Graduate in Business Analytics and AWS Certified Solution Architect - Associate with proficiency in machine learning, cloud computing and data modeling. Have led a couple of cross-functional teams to design reporting and predictive modeling from disparate data source in agile methodology saving several thousand in cost per month.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Kubernetes
    Data Warehousing & ETL Software
    Cloud Computing
    Microsoft Azure
    Google Cloud Platform
    Apache NiFi
    MySQL
    Microsoft Excel
    Apache Hive
    MongoDB
    PostgreSQL
    PySpark
    Python
    Amazon Web Services
  • $6 hourly
    Career Objective Experienced System Engineer with a demonstrated history of working in the IT and services industry. Seeking a position that offers a professional challenge in development of Big data solutions by utilizing analytical, technical and problem-solving skills so as to significantly contribute to an organization and enhance my knowledge Professional Summary 1. 4+ years of experience in Big data Analytics and Development 2. Working proficiently on various Big data tools such as Hive, Sqoop, Spark with good understanding of HDFS 3. Extensive working experience in an AGILE development Methodology. 4. Having good hands-on experience in SQL and ETL process. 5. Certified Quantexa Data Engineer. 6. Basic knowledge of streaming using Kafka and cloud technologies - AWS
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    SQL Programming
    Apache Hive
    SQL
    Hive
    Scala
    Apache Hadoop
  • $16 hourly
    Passionate and experienced data engineer with a knack for turning complex datasets into actionable insights. With a strong foundation in data architecture and ETL processes, I specialize in designing and implementing scalable data solutions that drive business success. Proficient in a variety of technologies including Apache Spark, Hadoop, SQL, and cloud platforms like Azure and AWS, I thrive in transforming raw data into valuable assets. Whether it's optimizing data pipelines, building robust data warehouses, or developing innovative analytics solutions, I am dedicated to leveraging the power of data to fuel innovation and decision-making. Let's unlock the potential of your data together * Knows SQL, Python, Pyspark, Databricks, Hadoop, ADF, ADL, Snowflake, PowerBI * Full project management and assurance of quality of work throughout * Regular communication and updates are important to me, so lets keep in touch.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Databricks Platform
    PySpark
    Tableau
    Kubernetes
    Docker
    Jenkins
    GitHub
    Apache Pig
    Hive
    Snowflake
    Microsoft Azure
    Apache Hadoop
    Python
    SQL
  • $18 hourly
    A passionate coder and researcher with extensive experience in developing AI assisted systems. My work involves developing efficient ML solutions to various problems and efficiently deploying them on real work applications.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Research & Development
    Vue.js
    PyTorch
    OpenCV
    Django
    PostgreSQL
    Linux
    R
    Swift
    Java
    Python
    C++
    C
    Machine Learning
  • $20 hourly
    Objective: * To work in a challenging atmosphere by exhibiting my skill with utmost sincerity & dedicated smart work for the growth of organization along with myself. * A young energetic man with high level of dedication and strong sense of responsibility, time management. Also a quick learner and a team player and willing to work in an innovative, competitive environment which offers growth to the organization as well as to prove myself and be part of the company's success story. I can utilize my skills, expertise and creativity to add value to myself and contribute to the growth of the organization. With more than several years experience in Sales & Operation, Marketing, Customer service. Ability to understand business need quickly Excellent ability to solve problem creatively independently or effectively within the team culture
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Central Reservation Systems
    Sabre
    Amadeus CRS
    Customer Service
  • $40 hourly
    Hi, Glad you found me! I am a Data and Software Engineer interested in solving complex problems. I have 11+ years of diverse Data Engineering experience and currently, I lead a team of data engineers and data scientists to develop robust Data Ecosystems that could process high-speed streaming data and generate valuable insights from it, the current tech stack is Python, Spark, Kafka, AWS, Airflow, and Terraform. I have designed high DAU low latency data architecture, with production-grade data quality. My regular chores include code review, data modeling, analyzing streaming data, spark data pipeline/framework development, CI/CD management, and scrum planning. I have implemented Data Lakes, Warehouses, and Lakehouses over AWS, Azure, and GCP. It would be great if we could have further discussion.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Microsoft Azure
    ETL
    Sentiment Analysis
    Machine Learning
    Natural Language Processing
    Databricks Platform
    Database Design
    Apache Airflow
    Python
    Scala
    Apache Cassandra
    Elasticsearch
    Apache Hadoop
  • $3 hourly
    I am a developer with experience of designing and developing highly scalable systems, deploying them to cloud and writing data management workflows
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Amazon Web Services
    DevOps
    Engineering & Architecture
    AWS Lambda
    PostgreSQL
    Amazon Redshift
    Django
    NodeJS Framework
    JavaScript
    Golang
    Data Engineering
    Java
    Python
  • $3 hourly
    Passionate Data Science Enthusiast | Aspiring Data Analyst, Data Engineer | Open to Opportunities Hello there! I'm Amol Deshmukh, and I'm on a journey to explore the fascinating world of data science. As I have recently completed Data Science Bootcamp from ODIN School, I'm excited to embark on a career in data science and contribute to innovative solutions in this dynamic field. I see data science as a powerful tool to drive informed decisionmaking and make a positive impact on businesses and society. I've honed my skills in data analysis, Python programming, SQL, and Apache Spark through coursework, projects, and self-study. I thrive in collaborative environments and believe that diverse perspectives lead to innovative solutions. I'm eager to work alongside experienced professionals in data science and learn from their expertise while contributing my passion and fresh ideas. The field of data science is ever-evolving, and I'm committed to
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Looker
    Business Process Management
    Project Management
    Microsoft Power BI
    Python
    SQL
  • $40 hourly
    Backend Engineer with experience in B2B SaaS startups. * Expert in API Development with Python. * Up to date on modern software development best practices like CI/CD, Trunk-Based Development and more. * Proficient in SQL data modeling and query optimization.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Docker
    Flask
    SQLAlchemy
    Back-End Development
    PostgreSQL
    DevOps
    API Development
    Data Engineering
    Google Cloud Platform
    Amazon Web Services
    SQL
    Python
  • $30 hourly
    Summary Experience: 2.5 years of hands-on experience in Big Data technologies including Hadoop, PySpark, AWS, Hive, Impala, Vertica, Spark SQL, Python and SQL. Proficient in Python programming language and data manipulation libraries such as Pandas and PySpark. Skills: Expertise in Big Data technologies, demonstrating a deep understanding of Hadoop ecosystem components and their practical applications. Proficient in PySpark for efficient data processing and analysis, leveraging distributed computing capabilities. Experienced in utilizing Hive for data warehousing and executing SQL-like queries on large datasets. Skilled in Spark SQL for interactive querying and analysis of structured data, enabling comprehensive insights. Strong command of SQL for effective management and manipulation of relational databases.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Microsoft Power BI Data Visualization
    Apache Impala
    Vertica
    Amazon Web Services
    Big Data
    Apache Kafka
    Apache Hive
    SQL
    PySpark
    Apache Hadoop
  • $30 hourly
    I am a data engineer with extensive project experience in my current setup. I have professional knowledge of various cloud integration & migration systems. I can be of immense help with providing realistic solutions to your business needs. - Strong Experience with Airflow, Python, SQL and Azure Platform - I will always be available for a quick call
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Microsoft Azure
    Databricks Platform
    SQL
    Python
    Apache Airflow
    Big Data
    Data Engineering
  • $30 hourly
    I am Cloud Data Engineer experienced in architecting and building solutions across GCP and AWS cloud. My experience in building pipelines and data modelling can really help. I have extensively worked on Python programming and various SQL & No SQL databases. Also I have built APIs using Python Flask.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    REST API
    TIBCO Spotfire
    Informatica
    PostgreSQL
    MongoDB
    Hive
    Big Data
    PySpark
    NoSQL Database
    SQL
    Flask
    Python
    Google Cloud Platform
    Data Engineering
  • Want to browse more freelancers?
    Sign up

How hiring on Upwork works

1. Post a job (it’s free)

Tell us what you need. Provide as many details as possible, but don’t worry about getting it perfect.

2. Talent comes to you

Get qualified proposals within 24 hours, and meet the candidates you’re excited about. Hire as soon as you’re ready.

3. Collaborate easily

Use Upwork to chat or video call, share files, and track project progress right from the app.

4. Payment simplified

Receive invoices and make payments through Upwork. Only pay for work you authorize.

Trusted by

How do I hire a Apache Spark Engineer near Mumbai, on Upwork?

You can hire a Apache Spark Engineer near Mumbai, on Upwork in four simple steps:

  • Create a job post tailored to your Apache Spark Engineer project scope. We’ll walk you through the process step by step.
  • Browse top Apache Spark Engineer talent on Upwork and invite them to your project.
  • Once the proposals start flowing in, create a shortlist of top Apache Spark Engineer profiles and interview.
  • Hire the right Apache Spark Engineer for your project from Upwork, the world’s largest work marketplace.

At Upwork, we believe talent staffing should be easy.

How much does it cost to hire a Apache Spark Engineer?

Rates charged by Apache Spark Engineers on Upwork can vary with a number of factors including experience, location, and market conditions. See hourly rates for in-demand skills on Upwork.

Why hire a Apache Spark Engineer near Mumbai, on Upwork?

As the world’s work marketplace, we connect highly-skilled freelance Apache Spark Engineers and businesses and help them build trusted, long-term relationships so they can achieve more together. Let us help you build the dream Apache Spark Engineer team you need to succeed.

Can I hire a Apache Spark Engineer near Mumbai, within 24 hours on Upwork?

Depending on availability and the quality of your job post, it’s entirely possible to sign up for Upwork and receive Apache Spark Engineer proposals within 24 hours of posting a job description.