Hire the best Pyspark Developers in Chennai, IN

Check out Pyspark Developers in Chennai, IN with the skills you need for your next job.
  • $35 hourly
    Microsoft Azure Ecosystem: 1. Expertise in Azure Functions for real-time data processing, automation, and seamless integration with various Azure services. 2. Developed Azure Data Factory (ADF) pipelines to orchestrate ETL workflows, enabling efficient data ingestion, transformation, and movement across cloud environments. 3. Built scalable data processing solutions using Azure Databricks, optimizing large-scale analytics and AI-driven insights. 4. Implemented Logic Apps to automate complex business workflows, integrating with third-party APIs, SharePoint, and Microsoft Teams. 5. Managed and optimized Azure SQL Database, Azure Data Lake, and Dataverse, ensuring secure, efficient, and scalable data storage and processing. Power Tools: 6. Developed Power Automate workflows to streamline business processes such as email automation, document management, task scheduling, and notifications. 7. Integrated Power Automate with SharePoint, Microsoft Teams, and Dataverse to enable seamless data movement, approvals, and workflow automation. 8. Designed custom Power Automate workflows for manual process triggers, automated scheduling, and enhanced operational efficiency. 9. Proficient in Power BI for data visualization, report automation, and interactive dashboards, leveraging Power Query and Power Pivot for advanced data transformations. 10. Experienced in Power Apps for building custom applications, enabling users to interact with Azure and Power BI data sources dynamically.
    Featured Skill Pyspark
    Azure App Service
    Azure Service Fabric
    Microsoft Azure SQL Database
    Microsoft Azure
    Microsoft Office
    Microsoft Dynamics 365
    Microsoft Power Automate Administration
    Microsoft Power Automate
    PySpark
    Microsoft PowerApps
    Microsoft Excel PowerPivot
    Microsoft SharePoint
    Microsoft SharePoint Development
    Power Query
    Data Analysis
  • $35 hourly
    Seasoned data engineer with over 11 years of experience in building sophisticated and reliable ETL applications using Big Data and cloud stacks (Azure and AWS). TOP RATED PLUS . Collaborated with over 20 clients, accumulating more than 2000 hours on Upwork. 🏆 Expert in creating robust, scalable and cost-effective solutions using Big Data technologies for past 9 years. 🏆 The main areas of expertise are: 📍 Big data - Apache Spark, Spark Streaming, Hadoop, Kafka, Kafka Streams, Trino, HDFS, Hive, Solr, Airflow, Sqoop, NiFi, Flink 📍 AWS Cloud Services - AWS S3, AWS EC2, AWS Glue, AWS RedShift, AWS SQS, AWS RDS, AWS EMR 📍 Azure Cloud Services - Azure Data Factory, Azure Databricks, Azure HDInsights, Azure SQL 📍 Google Cloud Services - GCP DataProc 📍 Search Engine - Apache Solr 📍 NoSQL - HBase, Cassandra, MongoDB 📍 Platform - Data Warehousing, Data lake 📍 Visualization - Power BI 📍 Distributions - Cloudera 📍 DevOps - Jenkins 📍 Accelerators - Data Quality, Data Curation, Data Catalog
    Featured Skill Pyspark
    SQL
    AWS Glue
    PySpark
    Apache Cassandra
    ETL Pipeline
    Apache Hive
    Apache NiFi
    Apache Kafka
    Big Data
    Apache Hadoop
    Scala
    Apache Spark
  • $40 hourly
    I have over 8 years of experience in building graph based solutions both in OLTP and OLAP and also in predictive analytics. I have collaborated with IT Project Managers to understand their requirements, and independently delivered over 30 dashboards (each comprises more than 100 KPIs and viz) which is to analyze the productivity of teams.
    Featured Skill Pyspark
    Spring Boot
    Docker
    Grafana
    Regex Writing
    PySpark
    Apache Spark
    Apache Lucene
    Python Scikit-Learn
    Oracle
    Amazon Neptune
    Neo4j
    TypeScript
    Node.js
    Java
    Python
  • $70 hourly
    I Specialize in building Data Warehouses, Data lakes, Data Integration, Data Modeling, Data lifecycle Management, Data Governance, and architecture solutions with an extensive experience of more than 18 years. I am an expert in building ETL/ELT data pipelines on distributed infrastructure applications, both on-prem and cloud. As an expert in Dimensional Modeling, I have built several Data marts over on-prem and cloud Datawarehouses.
    Featured Skill Pyspark
    Snowflake
    Orchestration
    Containerization
    Microservice
    Data Management
    Data Ingestion
    Data Modeling
    Data Lake
    Informatica Data Quality
    Informatica
    PySpark
    Big Data
    Data Engineering
    Python
  • $150 hourly
    Hello! I'm Vishnu Vyas, a seasoned data scientist and full-stack developer with a deep passion for leveraging the power of Python, machine learning, and NLP to drive insights and actionable results. With my proficiency in Python/Pyspark, AWS cloud services, and my intimate knowledge of Healthcare technologies such as FHIR and Medicare Advantage, I bring a holistic approach to solving complex problems in the healthcare sector. In addition to my data science skills, I'm also experienced in web development, with a toolkit that includes React, Node.js, and other frontend technologies. My goal is to create comprehensive solutions that are scalable, efficient, and user-friendly.
    Featured Skill Pyspark
    Node.js
    React
    Machine Learning
    Data Science
    PySpark
    Python
  • $35 hourly
    With 7+ years of Total experience, worked as Technical Lead, Data analyst, Data Engineer, Business process lead, Hadoop Developer, Oracle Developer, Application & Technical Support.
    Featured Skill Pyspark
    PySpark
    Apache Hadoop
    ETL
    ETL Pipeline
    Data Extraction
  • $40 hourly
    Data Analyst | Power BI, Azure, PySpark & Data Automation I am a data analyst with a strong background in data visualization, automation, and pipeline management, helping businesses transform raw data into actionable insights. With expertise in Power BI, Azure Databricks, PySpark, and Azure Data Factory, I specialize in: ✔ Building and enhancing Power BI dashboards for real-time insights and seamless decision-making. ✔ Automating data workflows using PySpark in Azure Databricks, improving efficiency for large-scale datasets. ✔ Executing and monitoring Azure Data Factory pipelines to ensure smooth data integration and transformation. ✔ Conducting exploratory data analysis (EDA) with Pandas, resolving anomalies, and standardizing datasets. ✔ Designing impactful data visualizations with Seaborn & Tableau, making complex data easy to understand. ✔ Leading cross-functional projects, optimizing workflows, and delivering data-driven solutions. With a results-driven approach, I ensure clean, efficient, and insightful data that empowers businesses to make informed decisions. 🚀 Let’s connect and turn your data into strategic value!
    Featured Skill Pyspark
    Microsoft Power BI
    Data Visualization
    Databricks Platform
    Alteryx, Inc.
    Microsoft Excel
    Tableau
    PySpark
    Python
    Data Analysis
    ETL Pipeline
    ETL
    Data Extraction
  • $20 hourly
    Data Scientist: I am a Data Scientist proficient in leveraging statistical analysis, machine learning techniques, and domain expertise to derive actionable insights from complex datasets. With a strong foundation in mathematics, statistics, and programming languages like Python, I excel in manipulating large datasets and building predictive models. Collaborating closely with stakeholders, I identify business problems, analyze data, and develop algorithms to drive strategic initiatives and optimize processes. My expertise in data visualization and communication enables me to convey findings effectively to non-technical audiences, facilitating informed decision-making within organizations. Data Engineer: As a Data Engineer, I specialize in designing, building, and maintaining robust data infrastructure to support storage, processing, and analysis requirements. Proficient in programming languages such as SQL, Java, and Scala, I architect scalable data pipelines and integrate disparate data sources to ensure data quality and reliability. With expertise in big data technologies like Hadoop, Spark, and Kafka, I collaborate with cross-functional teams to understand data needs and implement solutions that meet organizational objectives. I also prioritize data governance and security practices to safeguard sensitive information, enabling businesses to derive actionable insights and make informed, data-driven decisions.
    Featured Skill Pyspark
    MongoDB
    Apache Spark
    PySpark
    Apache Hadoop
    Data Engineering
    Data Science Consultation
    Data Science
    Statistical Analysis
    JupyterLab
    PyTorch
    pandas
    Python Scikit-Learn
    SciPy
    NumPy
    Python
  • $25 hourly
    As a Data Engineer, I specialize in designing and implementing ETL/ELT pipelines across various platforms to create robust Data Warehouses and Data Lakes. Expertise in PySpark, AWS Glue, Databricks, AWS Lambda, Amazon S3, SnapLogic, ETL processes, PL/SQL, SQL, and Amazon Redshift.
    Featured Skill Pyspark
    Databricks Platform
    SQL Server Reporting Services
    Microsoft SQL Server
    Amazon Redshift
    Apache Airflow
    SnapLogic
    AWS Lambda
    AWS Glue
    SQL
    PySpark
    ETL Pipeline
    ETL
    Data Extraction
  • $5 hourly
    I’m a Data Engineer with 3+ years of experience in data transformation, visualization, and automation. I specialize in building scalable data pipelines, integrating APIs, and developing dashboards to empower businesses with data-driven decision-making. 💡 What I Can Do for You: ✅ Data Engineering & ETL – Extract, transform, and load (ETL) data from multiple sources into structured formats for analysis. ✅ Python Development – Automate data processes, develop APIs, and work with JSON for seamless data exchanges. ✅ Data Visualization – Build interactive Power BI dashboards to uncover insights and track KPIs. ✅ Airtable Customization – Design workflow automation and database structures to streamline business operations. ✅ AI Model Evaluation – Train and assess models for accurate forecasting and data predictions. ✅ API Integration – Connect internal and external systems to automate workflows and ensure smooth data flow. 🛠 Tech Stack & Tools: 🔹 Python (Pandas, NumPy, JSON, APIs) 🔹 SQL (Data Queries & Optimization) 🔹 Power BI (Dashboards & Data Modeling) 🔹 Airtable (Database Structuring & Automation) 🔹 Jupyter Notebooks (Analysis & Documentation) 🔹 Data Warehousing & Transformation 🌟 Why Work with Me? ✔️ Detail-Oriented – I ensure clean, structured, and meaningful data. ✔️ Business-Focused – I align technical solutions with business needs. ✔️ Automation-Driven – I optimize workflows and eliminate manual bottlenecks.
    Featured Skill Pyspark
    Big Data
    Databricks MLflow
    Databricks Platform
    PySpark
    Data Cleaning
    Python
    SQL
    Tableau
    Zapier
    Machine Learning Model
    Data Analysis
    Analytical Presentation
    n8n
    Airtable
    Microsoft Power BI
  • $20 hourly
    🚀 Expert Data Engineer | Databricks | PySpark | ETL | AWS | Big Data | SQL | Python | Power BI 🔹 8+ years of experience designing and building scalable data pipelines and ETL workflows in Databricks, transforming raw data into structured insights. 🔹 Proven expertise in data movement, transformation, and reporting—from ingesting data from APIs, databases, and cloud storage (AWS S3, Salesforce, JSON, etc.) to building Delta Lake architectures for analytics. 🔹 Hands-on experience optimizing incremental and full-load ETL processes for performance, ensuring secure, GDPR-compliant data handling. 🔹 Strong in SQL, Python (PySpark), and Databricks, automating complex data workflows and enabling seamless data integration across platforms. 🔹 Experience in building AI-powered solutions, including NLP-based chatbots for SQL generation and data-driven insights using machine learning models. 🔹 Data Visualization & Reporting – Designed interactive Power BI dashboards for executives, tracking KPIs, sales performance, and product analytics. 💡 Services I Offer: ✅ Databricks ETL Pipeline Development (Delta Lake, PySpark, SQL) ✅ Data Engineering & Cloud Integration (AWS, Azure, APIs, JSON, S3) ✅ Optimized SQL Query Writing & Performance Tuning ✅ Automated Data Pipelines & Workflow Orchestration ✅ Data Modeling & Warehouse Design ✅ Power BI Dashboards & Reports (for business insights) 📩 Looking for an experienced Data specialist to streamline your data workflows and visualise actionable insights? Let’s connect and experience never-seen-before professionalism and craftsmanship in every aspect of your project!
    Featured Skill Pyspark
    Microsoft Excel
    Database Report
    Data Visualization
    AWS Lambda
    AWS Glue
    SQL
    Microsoft Power BI
    Python
    PySpark
    Databricks Platform
    ETL
  • $10 hourly
    I am highly motivated and dedicated to gain more experience and work hard towards completing my task perfectly within the deadlines. Interested in Data Analytics, Data Engineering, Data Science, Finance, Management, and other roles that can add value to both parties.
    Featured Skill Pyspark
    Tutoring
    Management Skills
    Typing
    Data Entry
    Data Engineering
    PySpark
    Financial Analysis
    Operational Planning
    Business Process Management
    Databricks Platform
    Python
    SQL
    Microsoft Power BI
    ETL
    Data Analysis
  • $14 hourly
    I am a data engineer specialising in ETL. I have experience in python, spark and ETL tools like Informatica. I have more than six years of experience in the field.
    Featured Skill Pyspark
    Data Extraction
    PySpark
    Python
    Database
    ETL Pipeline
    ETL
  • $30 hourly
    Full Stack Data Engineer | End-to-End Data Solutions | Reliable. Scalable. Insightful. Hi! I’m a versatile Full Stack Data Engineer with a strong foundation in both backend development and data infrastructure. I help businesses make smarter decisions by building reliable data pipelines, managing cloud-based data systems, and delivering clean, actionable insights. What I bring to the table: ✔️ Data pipeline design & automation ✔️ Database development & optimization ✔️ Cloud & on-premises data integration ✔️ Scalable backend services ✔️ Interactive dashboards & data apps ✔️ Strong focus on performance, quality, and results Whether you're starting from scratch or need to level up your existing data systems, I can jump in and drive your project forward—end to end. Let’s connect and turn your data into a powerful asset
    Featured Skill Pyspark
    Data Extraction
    Generative AI
    MongoDB
    Talend Open Studio
    ETL Pipeline
    PySpark
    Big Data
    Data Migration
    Linux
    Unix Shell
    Oracle
    SQL
    Snowflake
    Python
  • $15 hourly
    Overall 6 years of experience in IT industry with 4 years of relevant experience in Big Data Engineer, handling and transforming heterogeneous data into Key information using Hadoop ecosystem. - Expertise with the tools in Hadoop Ecosystem – HDFS, Hive , Sqoop, Spark, Kafka, Nifi. - Experience working with Elastic Search, Kibana and good knowledge on Oozie, Hbase, Phonix. - Good understanding of distributed systems, HDFS architecture, internal working details of MapReduce, Yarn and Spark processing frameworks. - More than two year of hands on experience using Spark framework with Scala. - Expertise in Inbound and Outbound (importing/exporting) data form/to traditional RDBMS using ApacheSQOOP. - Extensively worked on HiveQL, join operations, writing custom UDF’s and having good experience in optimizing Hive Queries. - Experience in data processing like collecting, aggregating, moving from various sources using Apache Nifi and Kafka. - Worked with various formats of files like delimited text files , JSON files, XML Files - Having basic knowledge on Amazon Web Services.
    Featured Skill Pyspark
    Data Lake
    AWS CloudFormation
    AWS Glue
    Elasticsearch
    Kibana
    Sqoop
    Apache NiFi
    PySpark
    Scala
    SQL
    Apache Hadoop
    Apache Kafka
    Apache Hive
    Apache Spark
  • $40 hourly
    With over 16 years of experience in Information Technology and having played a Technical Specialist/Architect role for various assignments with cutting edge technologies involving Application Design, Development, Functional Studies and component development on projects for clients such as Nokia-Siemens (Velocix), Verizon Business, Walgreens, and Qatar Airways, Citi. Currently working as Product Architect for Media product company (Velocix) with exposure to SAAS based product on public cloud (AWS).
    Featured Skill Pyspark
    Apache Solr
    Kubernetes
    Rancher
    Spring Boot
    Java
    Apache Kafka
    Spring AMQP
    PySpark
  • $50 hourly
    Intermediate Data Engineer I'm a mid level Data Engineer who finds data is the new gold landscape intriguing and exciting. I have been a python developer and is into ETL. I use native python coding, pyspark for any distributed processing and well versed with some of the services offered by AWS. I have mid level knowledge with snowflake Data warehouse and some good hands on experiene with postgresql.
    Featured Skill Pyspark
    Machine Learning Model
    ETL
    PySpark
    Data Engineering
    PostgreSQL
    Snowflake
    Python
  • $15 hourly
    6+ years of experience in Data Science and Machine learning 6+ years of experience as a Python and R Developer - Data Analysis, data cleansing, Exploratory Data Analysis(EDA), Dimensionality reduction, Feature scaling, Modeling, Model evaluation, and Hyper-Parameter tuning. - data analysis and visualizations using Tableau, Python, and R. - NLP(Natural Language Processing) with packages like NLTK, Spacy, Bag of Words, Scikit Learn, etc. - mining PDF using PDF Plumber, PYPDF2, Spacy PDF reader, OCR, and Tesseract. - PDF extraction through LLM and Generative AI like ChatGPT - Time series analysis, Neural networks, Deep learning. - Experience in Content writing on Data Science and AI. - Strong written and oral communication skills, Permanent Freelancer. Always available 24x7.
    Featured Skill Pyspark
    PySpark
    Scala
    R
    Web Design
    SEO Content
    LLM Prompt Engineering
    Generative AI
    Machine Learning
    Data Science
    Python
  • $30 hourly
    I am a Microsoft certified Data Engineer and Data Analyst, with Experience in Python , LLM, AI,Machine learning ,deeplearning, AI agents SQL,Power Bi, Microsoft Fabric and Tableau. i can work with data end to end, whether it can be building data pipelines or data flows or data warehousing, or building AI models all rounder in any thing data
    Featured Skill Pyspark
    Data Lake
    Data Engineering
    PySpark
    Python
    SQL
    LLM Prompt
    Deep Learning
    Machine Learning
    Tableau
    Microsoft Power BI
    Business Intelligence
    Data Analysis
  • $10 hourly
    Your Pocket Data Developer . Anytime whenever you have Emergency Sr. Data Engineer with wide knowledge of MSSQL, SSIS, SSRS and Power BI . Key Skill. Data Integration, Data transformation and warehousing Over the last 9+ years, I am senior consultant for all MS SQL services. I have experience in the following areas: MSSQL Server, SSRS, SSIS, power BI. My core competency lies in complete solution design for Database development project. I am seeking opportunities to build these services from the ground up for you or your business. Also I have hand-on experience with azure in different areas (IAAS, PAAS) Key Skills SSIS ETL development and improvement API integration and delta load Database development Database Performance tuning Database 24/7 Monitoring Support. Database Consulting Power BI Management Technologies MS SQL server SSIS SSRS Power BI
    Featured Skill Pyspark
    PySpark
    AWS Glue
    Amazon S3
    Amazon Redshift
    Databricks Platform
    Unix
    Microsoft Power BI
    SQL Programming
    SQL Server Reporting Services
    SQL Server Integration Services
    Database
    GitHub
    Python
  • $15 hourly
    As a Senior Data Engineer my skills in AWS, SnowFlake, and SQL to design, develop, and optimize data solutions for the automotive industry. I have over 5 years of experience in data engineering, working with IT OPENDOORS PVT LTD as an Azure Data Engineer and a Power BI Developer . I have contributed to multiple projects involving data integration, transformation, visualization, and analysis, supporting business intelligence, analytics, and reporting needs. I am passionate about leveraging data to drive innovation, efficiency, and value for the automotive sector, and I am always eager to learn new technologies and best practices. I am a team player, a problem solver, and a customer-focused professional, who strives to deliver high-quality data solutions that meet or exceed expectations.
    Featured Skill Pyspark
    PySpark
    Data Visualization
    Machine Learning
    pandas
    SQL
    Talend Open Studio
    Snowflake
    Databricks Platform
    AWS Glue
    Microsoft Power BI
    Data Engineering
    Microsoft Azure
  • $15 hourly
    I'm a DataOps and Data Platform Engineer experienced in building Data Platform environment, Databricks workspace on Azure and AWS platform. Experienced in ETL Pipeline using Azure Data Factory, Azure Databricks and Azure Synapse Analytics. Experienced in Devops(Azure and AWS) using Github, Azure devops, Github actions.
    Featured Skill Pyspark
    Architectural Design
    Linux System Administration
    PySpark
    Python
    CI/CD
    Microsoft Azure
    Docker
    Kubernetes
    Azure DevOps
    Terraform
    Databricks Platform
    DevOps
  • $20 hourly
    Big Data developer specializing in Performance tuning and writing efficient code particularly in apache spark
    Featured Skill Pyspark
    Data Extraction
    ETL
    Cloudera
    Scripting Language
    Unix Shell
    Apache Spark
    PySpark
  • $3 hourly
    I'm Venkatesh Anantharaman, an Experienced data analyst with a dynamic background in engineering and a knack for uncovering insights that drive impactful decisions. I thrive on solving complex problems with data-driven strategies. What I Bring to the Table? 1. Data Visualization: Crafting visuals that distill complex datasets into clear, actionable insights. 2. Data Analytics: Diving deep into numbers to uncover trends, patterns, and opportunities. 3. Data Cleaning & Modeling: Ensuring data integrity through standardization, deduplication, and other essential techniques for high-quality analysis. Reporting & Dashboarding: Developing dashboards that tell a story, making data engaging and informative. 4. Business Intelligence: Enabling strategic decisions through data-driven insights. Data Storytelling: Transforming numbers into narratives that resonate with audiences.
    Featured Skill Pyspark
    Analytical Presentation
    PySpark
    Microsoft Power BI
    Data Visualization
    Data Analysis
    Business Analysis
    Power Query
    Data Extraction
    ETL
  • Want to browse more freelancers?
    Sign up

How hiring on Upwork works

1. Post a job

Tell us what you need. Provide as many details as possible, but don’t worry about getting it perfect.

2. Talent comes to you

Get qualified proposals within 24 hours, and meet the candidates you’re excited about. Hire as soon as you’re ready.

3. Collaborate easily

Use Upwork to chat or video call, share files, and track project progress right from the app.

4. Payment simplified

Receive invoices and make payments through Upwork. Only pay for work you authorize.

Trusted by

How do I hire a Pyspark Developer near Chennai, on Upwork?

You can hire a Pyspark Developer near Chennai, on Upwork in four simple steps:

  • Create a job post tailored to your Pyspark Developer project scope. We’ll walk you through the process step by step.
  • Browse top Pyspark Developer talent on Upwork and invite them to your project.
  • Once the proposals start flowing in, create a shortlist of top Pyspark Developer profiles and interview.
  • Hire the right Pyspark Developer for your project from Upwork, the world’s largest work marketplace.

At Upwork, we believe talent staffing should be easy.

How much does it cost to hire a Pyspark Developer?

Rates charged by Pyspark Developers on Upwork can vary with a number of factors including experience, location, and market conditions. See hourly rates for in-demand skills on Upwork.

Why hire a Pyspark Developer near Chennai, on Upwork?

As the world’s work marketplace, we connect highly-skilled freelance Pyspark Developers and businesses and help them build trusted, long-term relationships so they can achieve more together. Let us help you build the dream Pyspark Developer team you need to succeed.

Can I hire a Pyspark Developer near Chennai, within 24 hours on Upwork?

Depending on availability and the quality of your job post, it’s entirely possible to sign up for Upwork and receive Pyspark Developer proposals within 24 hours of posting a job description.