Hire the best Apache Spark Engineers in Gurgaon, IN

Check out Apache Spark Engineers in Gurgaon, IN with the skills you need for your next job.
Clients rate Apache Spark Engineers
Rating is 4.7 out of 5.
4.7/5
based on 283 client reviews
  • $25 hourly
    Hello, I'm Aditya Johar, a seasoned Data Scientist and Full Stack Developer. With over 9 years of hands-on experience, I bring a wealth of expertise to the table. Here are the top 5 qualities that make me a reliable, highly experienced, and talented expert for your project collaborations: [1] My journey through the world of data science and full-stack development has exposed me to a plethora of tools and technologies. I am well-versed in Python, along with its data science libraries like Pandas, NumPy, and Scikit-Learn. For deep learning and AI, I work with frameworks such as TensorFlow and PyTorch. In the full-stack domain, I'm proficient in Node.js, Express, and MongoDB. [2] I thrive on tackling complex challenges. I have a track record of turning data into actionable insights using data visualization tools like Matplotlib and Seaborn, and efficiently managing databases through SQL and NoSQL systems. When it comes to full-stack development, I excel in both front-end technologies (React, HTML, CSS) and back-end frameworks (Django, Flask). [3] My versatility allows me to cover the entire project pipeline. I can dive into data analysis, modeling, and end-to-end application development with proficiency in React and Redux. This comprehensive approach streamlines project execution. [4] Effective communication is at the core of my work. I excel at translating complex technical concepts into plain language, making it accessible to non-technical stakeholders. [5] I've accumulated a portfolio of successful projects, showcasing my ability to deliver high-quality solutions on time and within budget. You'll find case studies, project highlights, and testimonials from satisfied clients. ------------------------------------------TOP USE CASES COVERED--------------------------------- ✅ NATURAL LANGUAGE PROCESSING (NLP): Sentiment Analysis, Text Summarization, Chatbots and Virtual Assistants, Language Translation ✅COMPUTER VISION: Image and Video Classification, Object Detection, Facial Recognition, Medical Image Analysis ✅RECOMMENDATION SYSTEMS: Product Recommendations (e.g., e-commerce), Content Recommendations (e.g., streaming services), Personalized Marketing ✅PREDICTIVE ANALYTICS: Sales and Demand Forecasting, Customer Churn Prediction, Stock Price Prediction, Equipment Maintenance Prediction ✅E-COMMERCE OPTIMIZATION: Dynamic Pricing, Inventory Management, Customer Lifetime Value Prediction ✅TIME SERIES ANALYSIS: Financial Market Analysis, Energy Consumption Forecasting, Weather Forecasting ✅SPEECH RECOGNITION: Virtual Call Center Agents, Voice Assistants (e.g., Siri, Alexa) ✅AI IN FINANCE: Credit Scoring, Algorithmic Trading, Fraud Prevention ✅AI IN HR: Candidate Screening, Employee Performance Analysis, Workforce Planning ✅CONVERSATIONAL AI: Customer Support Chatbots, Virtual Shopping Assistants, Voice Interfaces ✅AI IN EDUCATION: Personalized Learning Paths, Educational Chatbots, Plagiarism Detection ✅AI IN MARKETING: Customer Segmentation, Content Personalization, A/B Testing ✅SUPPLY CHAIN OPTIMIZATION: Demand Forecasting, Inventory Optimization, Route Planning And Many More use cases that we can discuss while we connect. "Ready to turn these possibilities into realities? I'm just a click away! Feel free to contact me, or if you're eager to get started, simply click the 'Invite to Job' or 'Hire Now' button in the top right corner of your screen. Let's kick off your project and make it a success!"
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Django
    Apache Airflow
    Apache Hadoop
    Terraform
    PySpark
    Apache Kafka
    Flask
    BigQuery
    BERT
    Python Scikit-Learn
    pandas
    Python
    TensorFlow
    Data Science
  • $30 hourly
    Highly Skilled IT Professional Experience in GCP Cloud with over 3+ years of experience working as Cloud Data Engineer.  Overall, 3+ years in IT Experience. In ML Pipeline Development, Analysis, Testing, Data Warehouse and Business Intelligence tools.  Working within an Agile delivery methodology production implementation in iterative sprints My Skills: Cloud Based * Google Cloud Platform * Bigquery * Airflow * Dataproc, VetexAI * Composer * Google Cloud Storage * Google Cloud Function * Compute Engine Programming Languages: * Python * SQL * Shell Scripting.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Vertex AI
    Cloud Migration
    Data Engineering
    Google Sheets
    SAS
    Data Scraping
    Data Warehousing
    Python
    SQL
    Big Data
    Data Migration
    Apache Airflow
    Google Cloud Platform
    BigQuery
  • $6 hourly
    Achievement-driven & innovative professional with over 4 years of extensive experience in the field of Software Development  Currently associated as Senior Software Engineer with Park+  Understanding of Software Development Lifecycle (SDLC) right from requirement analysis, documentation (functional specifications, technical design), coding, and testing (preparation of test cases along with implementation) to the maintenance of proposed applications  Expertise in end-to-end implementation of various projects including designing, development, coding, integration and implementation of software applications  Capable of independently executing various initiatives & ensuring timely & and smooth completion of the same; skilled at communicating complex technical requirements to non-technical stakeholders in an effective manner  Accomplished at thriving in fast-paced environments, readily adapting to evolving business and technology challenges; exposure to designing, developing, testing, and debugging software; reviewing code and design  Effective communicator with excellent relationship-building & and interpersonal skills; strong analytical, problem-solving & and organizational capabilities with a flexible & detail-oriented attitude
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Google
    Amazon
    Terraform
    NoSQL Database
    FastAPI
    Django Stack
    Product Development
    Software Development
    Azure DevOps
    Ansible
    pandas
    SQL
    Flask
    Python
  • $25 hourly
    A highly motivated person with strong technical, problem-solving, and excellent time management skills who likely to create an impact on the organization/work, he is part of and always love to socialize and experience new things in life. My hunger for the new challenges make me unique.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    MySQL
    Scala
    Microsoft Azure
    Data Analytics
    Snowflake
    SQL Programming
    Data Engineering
    Data Warehousing & ETL Software
    ETL Pipeline
    PySpark
    SQL
    Databricks Platform
    Python
  • $50 hourly
    SUMMARY 9+ years experienced enterprising software engineering professional, gained hands-on experience in designing and implementing software solutions. Adept at brainstorming & creating various tools and applications by producing efficient code, conducting systems analysis to improve performance, and automating tasks through appropriate tools.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Media Analytics
    Amazon Redshift
    Bitbucket
    Media & Entertainment
    Amazon EC2
    Amazon S3
    Hive
    Big Data
    Apache Airflow
    Apache Kafka
    Data Analytics
    Android SDK
    Android
    Android Studio
  • $35 hourly
    I am a Data Engineer, with expertose in managing end to end data flow, starting from ingestion till reporting.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Artificial Intelligence
    SQL
    PySpark
    SQL Programming
    Apache NiFi
    Apache Spark MLlib
    AWS Glue
  • $20 hourly
    Microsoft certified Data Engineer, Experienced web developer with outstanding programming skills that allow to create high-quality and bug-free products. I have strong experience in the following areas: + Azure Data Factory + Azure Data Lake + Azure Databricks + Azure Servicebus +Azure Logic Apps + ASP.NET, + C#,  + MSSQL,  + MVC,  + JavaScript,  + JQuery,  + Ajax + Entity Framework,  + LINQ + Node.Js
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Data Warehousing & ETL Software
    Microsoft Azure
    Data Lake
    Big Data
    Apache Hive
    Angular 4
    R
    ASP.NET MVC
    pandas
    Python
    C#
    SQL
  • $25 hourly
    I am a Full Stack (Python with React)/(PHP with React) Developer experienced in the LAMP/WAMP/PYTHON/DJANGO platform. I have a Master's in Software Engineering. Listed below are my experience and skills. - 10+ years of experience in web application development using PHP, Python, MySQL, and JavaScript; - 4+ years of relevant experience in Python with Django and API development; - 3+ years of relevant experience in Web Scraping, Data Extraction, Web Crawling, Data Mining, Data Engineer, Python, PySpark,Scrapy; - Expert knowledge of Codeigniter, Laravel, Yii frameworks, also have experience with Symfony, CakePHP frameworks; - In-depth knowledge of Search Technology (Elastic search and Solr Search); - Good experience of AWS Services; - In-depth knowledge of REST and SOAP with 3rd party APIs like Amazon MWS, eBay API, and other public APIs like Facebook API and Google API; - In-depth knowledge of systems architecture and software design methodologies; - Good experience in GUI development using Jquery, ReactJS, Twitter Bootstrap; - Extensive experience in relational database management systems, mainly with MySQL and PostgreSQL; - Experience with NoSQL database MongoDB; - Experience in Linux systems administration (LAMP, Nginx); - I practice Agile Methodologies in the development process; - I use Git for version control, CI/CD Pipeline, Jenkins, Codedeploy; I am a self-driven, very fast learner.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    FastAPI
    Elasticsearch
    CodeIgniter
    Django
    Apache Hadoop
    PHP
    PySpark
    MySQL
    Apache Solr
    Laravel
    API
    Apache Airflow
    Python
    Data Scraping
  • $3 hourly
    Been working as a data engineer in companies for nearly 3 years now. Have experience with Python, SQL, pyspark, ETL. AWS experience with services such as Glue, EMR, Redshift, S3, RDS and have some snowflake experience too.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Selenium WebDriver
    Linux
    Amazon EC2
    Amazon Athena
    AWS Lambda
    Amazon S3
    Amazon Redshift
    AWS Glue
    PySpark
    SQL
    Python
  • $100 hourly
    Experience in creating scalable data architecture and pipelines on cloud. Expert in Tech Stack: Python, SQL, SPARK, Snowflake, AWS, AZURE, Data Modeling, Data Warehousing, Airflow, DBT, Data Processing, Data Analytics
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Amazon Athena
    Amazon Redshift
    AWS CloudFormation
    Amazon DynamoDB
    AWS Lambda
    AWS Glue
    Data Warehousing
    Apache Airflow
    Data Engineering
    Data Analytics
    dbt
    Snowflake
    Git
    PySpark
    Docker
    Python
    SQL
  • $20 hourly
    SUMMARY * Data Engineer with two plus years of experience on Azure Cloud Platform, Data Lake architecture and Data Warehousing. Worked closely with major clients in FMCG and BFSI domain. * Expertise in developing CI/CD pipelines using Azure Devops. * Major skills in Pyspark, Python, T-SQL, Snowflake, DBT and Azure cloud .
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    CI/CD
    Microsoft Azure SQL Database
    Distributed Computing
    Data Warehousing
    Azure DevOps
    Cloud Computing
    Microsoft Azure
    Data Lake
    Databricks Platform
    PySpark
    ETL Pipeline
    Python
    Microsoft Excel
  • $10 hourly
    PROJECTS Building Real Time data pipeline Handling customer experience data Used Kafka, Snowflake, Python, Scylla, Spark, Hive, Spark Optimization, SCD techniques to develop this system Upstream flows events through kafka topics and we wrote our processor code (python) to handle the processing logic for different events Post processing, domain events are produced to our domain kafka topic From domain kafka topic it goes to snowflake sink table and finally gets merged into Snowflake final DM table Further looker explore/dashboards are created, keeping Snowflake Dm table as source Currently working on snowflake to hive migration for cost cutting Developed snowflake to hive backfilling pipeline to transfer historical data Applied CAP theorem for optimal tuning of consistency with availability Exception Handling Framework Automate Exception Handling mechanism for Real time data pipeline Reduced 90% of human effort while encountering any kind of
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    MySQL
    SQL
    NoSQL Database
    Apache Cassandra
    Apache HBase
    Apache Airflow
    Snowflake
    Apache Kafka
    Scala
    Python
    PySpark
    Hive
    Real Time Stream Processing
  • $15 hourly
    A Big Data Developer with ~3 years of experience, leveraging years of honed expertise across a diverse spectrum of data stacks. My journey has been enriched by a wealth of experience, empowering me with a comprehensive skill set that spans Warehousing, ETL, Analytics, and Cloud Services. You have the data? Great !! I can help you analyze it using Python. It involves performing exploratory data analysis, hypothesis testing, and data visualization. You have Big Data? Even Better !! I can help you clean, transform, store and analyze it using big data technologies and productionize it using cloud services like AWS and GCP. You want to track business KPIs and metrics? Consider it done!! I can even help you develop reports using Tableau and PowerBI; this will always keep you ahead in your business. Specialized in the following data solutions: ✔️ Data Cleaning, Processing, and Machine Learning models ✔️ Building data warehouses using modern cloud platforms and technologies ✔️ Creating and automating data pipelines, real-time streaming & ETL processes ✔️ Building highly intuitive, interactive dashboards. ✔️ Data Migration (Heterogenous and Homogenous) Below are tools and technologies I have worked with - - Cloud: GCP (Google Cloud Platform), AWS (Amazon Web Services) - Databases: SQL Server, Snowflake, PostgreSQL, MySQL, S3 - Language & Libraries: Python, Pandas, Numpy, Matplotlib - Data Engineering: Spark, AWS Glue, PySpark, BiqQuery, Snowflake, ETL, Datawarehouse, Databricks, Data Lake - Orchestration Tools - Apache Airflow, and CronJobs, etc. - Reporting Tech: PowerBI, Tableau, Excel Let's collaborate and transform your idea into reality !!
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    TensorFlow
    Matplotlib
    Seaborn
    Python Scikit-Learn
    pandas
    Analytics
    Data Lake
    Deep Learning
    Data Analytics
    Machine Learning
    Amazon Web Services
    Snowflake
    SQL
    Python
  • $30 hourly
    Data engineer with 10+ years of experience in different industries like insurance, payment and finance
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    dbt
    Snowflake
    Apache Airflow
    Data Warehousing
    Scala
    Hive
    Apache Hadoop
    SQL
    Python
  • $4 hourly
    As an enthusiast of data and related fields, I strongly feel that a solution is not to be found rather the problem needs to be understood. Skilled in SQL, Python, Tableau & Machine Learning.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Machine Learning Model
    Information Analysis
    Analytical Presentation
    Data Analysis
    Python
    SQL
  • $20 hourly
    A data scientist impacting organizational performance through statistical modelling and insights. Innovative, self-motivated, hardworking individual with ten years of experience
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Databricks Platform
    Big Data
    Python
    Dashboard
    Microsoft Power BI
    Data Visualization
    R
  • $15 hourly
    Hello! I'm Devdatta Singh Yadav, a dedicated Software Engineer with a passion for crafting innovative solutions. Skills: Proficient in programming languages like Java, Ruby, C++, and JavaScript. Skilled in database management with PostgreSQL, RDBMS, and Active Record()ORM. Experienced in Android App Development using Android Studio and retrofit. Adept in web development with Ruby on Rails, React Js, Sidekiq, and Rspec. Familiar with tools like VS Code and Metabase. Experience: As an Associate Software Engineer at Cogoport Pvt. Ltd., I implemented real-time tracking solutions for enhanced truck visibility and customer satisfaction. I added new roles to manage truck availability and modified rate calculation systems to offer flexible pricing options. I also contributed to warehouse operations features and provided prompt technical support. During my internship at Khelgully Gaming Pvt. Ltd., I developed eSports tournaments modules, integrated user activity tracking, and enhanced customer communication. I also integrated payment gateways and implemented deep-link sharing for increased user engagement.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    React Native
    React
    Cache Management
    RSpec
    Software Development
    Web Application
    Android
    Kotlin
    Spring Framework
    PostgreSQL
    MySQL
    Ruby
    Java
    Ruby on Rails
  • $15 hourly
    Experienced data engineer skilled in harnessing big data technologies such as Hadoop, Spark, Snowflake, Databricks, Airflow, and Python to proficiently orchestrate data processing pipelines and analytical workflows. Demonstrated advanced expertise in infrastructure deployment, utilizing Docker and Kubernetes. Possesses a solid foundation in general computing principles and development methodologies. Additionally, excels in DevOps practices including CI/CD workflows, Jenkins automation, Git version control, and navigating cloud platforms for seamless deployment and scalability.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Amazon S3
    AWS Lambda
    Data Warehousing & ETL Software
    Apache Kafka
    Apache Hive
    Apache Hadoop
    SQL
    Big Data
    BigQuery
    Jenkins
    Apache Airflow
    Databricks Platform
    Snowflake
    Python
  • $5 hourly
    As a data scientist, I am responsible for extracting valuable insights from complex datasets to support data-driven decision-making within the organization. My primary focus is on conducting thorough analyses, identifying patterns and trends, and presenting the findings in a clear and concise manner to stakeholders. On a daily basis, I work with various data sources, including databases, spreadsheets, and external data feeds, to gather and clean the necessary information. I utilize a range of analytical tools and techniques, such as SQL, Python, and statistical modeling, to uncover meaningful insights that can help drive business objectives. One of my key responsibilities is to create interactive dashboards and visualizations that effectively communicate the analysis results. These dashboards serve as a central hub for stakeholders to monitor key performance indicators, track progress, and make informed decisions. I work closely with cross-functional teams to understand their specific data needs and tailor the dashboards accordingly. In addition to my analytical skills, I place a strong emphasis on effective communication and collaboration. I regularly present my findings to executives, managers, and subject matter experts, ensuring that the insights are easily understandable and actionable. I also work closely with data engineers and scientists to streamline data pipelines and improve the overall data ecosystem within the organization. As a data scientist, I am passionate about leveraging data to solve complex business problems and drive positive change. I continuously strive to enhance my skills, stay up-to-date with the latest industry trends, and contribute to the growth and success of the organization.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Apache Airflow
    Advanced Analytics
    Python
    BigQuery
    Big Data
    Cloud Engineering
    Cloud Computing
    PySpark
    SciPy
    NumPy
    Analytics Dashboard
    Deep Learning
    Machine Learning
    Data Analysis
  • Want to browse more freelancers?
    Sign up

How hiring on Upwork works

1. Post a job (it’s free)

Tell us what you need. Provide as many details as possible, but don’t worry about getting it perfect.

2. Talent comes to you

Get qualified proposals within 24 hours, and meet the candidates you’re excited about. Hire as soon as you’re ready.

3. Collaborate easily

Use Upwork to chat or video call, share files, and track project progress right from the app.

4. Payment simplified

Receive invoices and make payments through Upwork. Only pay for work you authorize.

Trusted by

How do I hire a Apache Spark Engineer near Gurgaon, on Upwork?

You can hire a Apache Spark Engineer near Gurgaon, on Upwork in four simple steps:

  • Create a job post tailored to your Apache Spark Engineer project scope. We’ll walk you through the process step by step.
  • Browse top Apache Spark Engineer talent on Upwork and invite them to your project.
  • Once the proposals start flowing in, create a shortlist of top Apache Spark Engineer profiles and interview.
  • Hire the right Apache Spark Engineer for your project from Upwork, the world’s largest work marketplace.

At Upwork, we believe talent staffing should be easy.

How much does it cost to hire a Apache Spark Engineer?

Rates charged by Apache Spark Engineers on Upwork can vary with a number of factors including experience, location, and market conditions. See hourly rates for in-demand skills on Upwork.

Why hire a Apache Spark Engineer near Gurgaon, on Upwork?

As the world’s work marketplace, we connect highly-skilled freelance Apache Spark Engineers and businesses and help them build trusted, long-term relationships so they can achieve more together. Let us help you build the dream Apache Spark Engineer team you need to succeed.

Can I hire a Apache Spark Engineer near Gurgaon, within 24 hours on Upwork?

Depending on availability and the quality of your job post, it’s entirely possible to sign up for Upwork and receive Apache Spark Engineer proposals within 24 hours of posting a job description.