Hire the best Pyspark Developers in Chennai, IN
Check out Pyspark Developers in Chennai, IN with the skills you need for your next job.
- $35 hourly
- 4.8/5
- (9 jobs)
Microsoft Azure Ecosystem: 1. Expertise in Azure Functions for real-time data processing, automation, and seamless integration with various Azure services. 2. Developed Azure Data Factory (ADF) pipelines to orchestrate ETL workflows, enabling efficient data ingestion, transformation, and movement across cloud environments. 3. Built scalable data processing solutions using Azure Databricks, optimizing large-scale analytics and AI-driven insights. 4. Implemented Logic Apps to automate complex business workflows, integrating with third-party APIs, SharePoint, and Microsoft Teams. 5. Managed and optimized Azure SQL Database, Azure Data Lake, and Dataverse, ensuring secure, efficient, and scalable data storage and processing. Power Tools: 6. Developed Power Automate workflows to streamline business processes such as email automation, document management, task scheduling, and notifications. 7. Integrated Power Automate with SharePoint, Microsoft Teams, and Dataverse to enable seamless data movement, approvals, and workflow automation. 8. Designed custom Power Automate workflows for manual process triggers, automated scheduling, and enhanced operational efficiency. 9. Proficient in Power BI for data visualization, report automation, and interactive dashboards, leveraging Power Query and Power Pivot for advanced data transformations. 10. Experienced in Power Apps for building custom applications, enabling users to interact with Azure and Power BI data sources dynamically.Pyspark
Azure App ServiceAzure Service FabricMicrosoft Azure SQL DatabaseMicrosoft AzureMicrosoft OfficeMicrosoft Dynamics 365Microsoft Power Automate AdministrationMicrosoft Power AutomatePySparkMicrosoft PowerAppsMicrosoft Excel PowerPivotMicrosoft SharePointMicrosoft SharePoint DevelopmentPower QueryData Analysis - $35 hourly
- 5.0/5
- (32 jobs)
Seasoned data engineer with over 11 years of experience in building sophisticated and reliable ETL applications using Big Data and cloud stacks (Azure and AWS). TOP RATED PLUS . Collaborated with over 20 clients, accumulating more than 2000 hours on Upwork. 🏆 Expert in creating robust, scalable and cost-effective solutions using Big Data technologies for past 9 years. 🏆 The main areas of expertise are: 📍 Big data - Apache Spark, Spark Streaming, Hadoop, Kafka, Kafka Streams, Trino, HDFS, Hive, Solr, Airflow, Sqoop, NiFi, Flink 📍 AWS Cloud Services - AWS S3, AWS EC2, AWS Glue, AWS RedShift, AWS SQS, AWS RDS, AWS EMR 📍 Azure Cloud Services - Azure Data Factory, Azure Databricks, Azure HDInsights, Azure SQL 📍 Google Cloud Services - GCP DataProc 📍 Search Engine - Apache Solr 📍 NoSQL - HBase, Cassandra, MongoDB 📍 Platform - Data Warehousing, Data lake 📍 Visualization - Power BI 📍 Distributions - Cloudera 📍 DevOps - Jenkins 📍 Accelerators - Data Quality, Data Curation, Data CatalogPyspark
SQLAWS GluePySparkApache CassandraETL PipelineApache HiveApache NiFiApache KafkaBig DataApache HadoopScalaApache Spark - $40 hourly
- 4.4/5
- (2 jobs)
I have over 8 years of experience in building graph based solutions both in OLTP and OLAP and also in predictive analytics. I have collaborated with IT Project Managers to understand their requirements, and independently delivered over 30 dashboards (each comprises more than 100 KPIs and viz) which is to analyze the productivity of teams.Pyspark
Spring BootDockerGrafanaRegex WritingPySparkApache SparkApache LucenePython Scikit-LearnOracleAmazon NeptuneNeo4jTypeScriptNode.jsJavaPython - $70 hourly
- 0.0/5
- (1 job)
I Specialize in building Data Warehouses, Data lakes, Data Integration, Data Modeling, Data lifecycle Management, Data Governance, and architecture solutions with an extensive experience of more than 18 years. I am an expert in building ETL/ELT data pipelines on distributed infrastructure applications, both on-prem and cloud. As an expert in Dimensional Modeling, I have built several Data marts over on-prem and cloud Datawarehouses.Pyspark
SnowflakeOrchestrationContainerizationMicroserviceData ManagementData IngestionData ModelingData LakeInformatica Data QualityInformaticaPySparkBig DataData EngineeringPython - $150 hourly
- 0.0/5
- (0 jobs)
Hello! I'm Vishnu Vyas, a seasoned data scientist and full-stack developer with a deep passion for leveraging the power of Python, machine learning, and NLP to drive insights and actionable results. With my proficiency in Python/Pyspark, AWS cloud services, and my intimate knowledge of Healthcare technologies such as FHIR and Medicare Advantage, I bring a holistic approach to solving complex problems in the healthcare sector. In addition to my data science skills, I'm also experienced in web development, with a toolkit that includes React, Node.js, and other frontend technologies. My goal is to create comprehensive solutions that are scalable, efficient, and user-friendly.Pyspark
Node.jsReactMachine LearningData SciencePySparkPython - $35 hourly
- 0.0/5
- (0 jobs)
With 7+ years of Total experience, worked as Technical Lead, Data analyst, Data Engineer, Business process lead, Hadoop Developer, Oracle Developer, Application & Technical Support.Pyspark
PySparkApache HadoopETLETL PipelineData Extraction - $40 hourly
- 0.0/5
- (0 jobs)
Data Analyst | Power BI, Azure, PySpark & Data Automation I am a data analyst with a strong background in data visualization, automation, and pipeline management, helping businesses transform raw data into actionable insights. With expertise in Power BI, Azure Databricks, PySpark, and Azure Data Factory, I specialize in: ✔ Building and enhancing Power BI dashboards for real-time insights and seamless decision-making. ✔ Automating data workflows using PySpark in Azure Databricks, improving efficiency for large-scale datasets. ✔ Executing and monitoring Azure Data Factory pipelines to ensure smooth data integration and transformation. ✔ Conducting exploratory data analysis (EDA) with Pandas, resolving anomalies, and standardizing datasets. ✔ Designing impactful data visualizations with Seaborn & Tableau, making complex data easy to understand. ✔ Leading cross-functional projects, optimizing workflows, and delivering data-driven solutions. With a results-driven approach, I ensure clean, efficient, and insightful data that empowers businesses to make informed decisions. 🚀 Let’s connect and turn your data into strategic value!Pyspark
Microsoft Power BIData VisualizationDatabricks PlatformAlteryx, Inc.Microsoft ExcelTableauPySparkPythonData AnalysisETL PipelineETLData Extraction - $20 hourly
- 0.0/5
- (0 jobs)
Data Scientist: I am a Data Scientist proficient in leveraging statistical analysis, machine learning techniques, and domain expertise to derive actionable insights from complex datasets. With a strong foundation in mathematics, statistics, and programming languages like Python, I excel in manipulating large datasets and building predictive models. Collaborating closely with stakeholders, I identify business problems, analyze data, and develop algorithms to drive strategic initiatives and optimize processes. My expertise in data visualization and communication enables me to convey findings effectively to non-technical audiences, facilitating informed decision-making within organizations. Data Engineer: As a Data Engineer, I specialize in designing, building, and maintaining robust data infrastructure to support storage, processing, and analysis requirements. Proficient in programming languages such as SQL, Java, and Scala, I architect scalable data pipelines and integrate disparate data sources to ensure data quality and reliability. With expertise in big data technologies like Hadoop, Spark, and Kafka, I collaborate with cross-functional teams to understand data needs and implement solutions that meet organizational objectives. I also prioritize data governance and security practices to safeguard sensitive information, enabling businesses to derive actionable insights and make informed, data-driven decisions.Pyspark
MongoDBApache SparkPySparkApache HadoopData EngineeringData Science ConsultationData ScienceStatistical AnalysisJupyterLabPyTorchpandasPython Scikit-LearnSciPyNumPyPython - $25 hourly
- 0.0/5
- (0 jobs)
As a Data Engineer, I specialize in designing and implementing ETL/ELT pipelines across various platforms to create robust Data Warehouses and Data Lakes. Expertise in PySpark, AWS Glue, Databricks, AWS Lambda, Amazon S3, SnapLogic, ETL processes, PL/SQL, SQL, and Amazon Redshift.Pyspark
Databricks PlatformSQL Server Reporting ServicesMicrosoft SQL ServerAmazon RedshiftApache AirflowSnapLogicAWS LambdaAWS GlueSQLPySparkETL PipelineETLData Extraction - $5 hourly
- 0.0/5
- (0 jobs)
I’m a Data Engineer with 3+ years of experience in data transformation, visualization, and automation. I specialize in building scalable data pipelines, integrating APIs, and developing dashboards to empower businesses with data-driven decision-making. 💡 What I Can Do for You: ✅ Data Engineering & ETL – Extract, transform, and load (ETL) data from multiple sources into structured formats for analysis. ✅ Python Development – Automate data processes, develop APIs, and work with JSON for seamless data exchanges. ✅ Data Visualization – Build interactive Power BI dashboards to uncover insights and track KPIs. ✅ Airtable Customization – Design workflow automation and database structures to streamline business operations. ✅ AI Model Evaluation – Train and assess models for accurate forecasting and data predictions. ✅ API Integration – Connect internal and external systems to automate workflows and ensure smooth data flow. 🛠 Tech Stack & Tools: 🔹 Python (Pandas, NumPy, JSON, APIs) 🔹 SQL (Data Queries & Optimization) 🔹 Power BI (Dashboards & Data Modeling) 🔹 Airtable (Database Structuring & Automation) 🔹 Jupyter Notebooks (Analysis & Documentation) 🔹 Data Warehousing & Transformation 🌟 Why Work with Me? ✔️ Detail-Oriented – I ensure clean, structured, and meaningful data. ✔️ Business-Focused – I align technical solutions with business needs. ✔️ Automation-Driven – I optimize workflows and eliminate manual bottlenecks.Pyspark
Big DataDatabricks MLflowDatabricks PlatformPySparkData CleaningPythonSQLTableauZapierMachine Learning ModelData AnalysisAnalytical Presentationn8nAirtableMicrosoft Power BI - $20 hourly
- 0.0/5
- (1 job)
🚀 Expert Data Engineer | Databricks | PySpark | ETL | AWS | Big Data | SQL | Python | Power BI 🔹 8+ years of experience designing and building scalable data pipelines and ETL workflows in Databricks, transforming raw data into structured insights. 🔹 Proven expertise in data movement, transformation, and reporting—from ingesting data from APIs, databases, and cloud storage (AWS S3, Salesforce, JSON, etc.) to building Delta Lake architectures for analytics. 🔹 Hands-on experience optimizing incremental and full-load ETL processes for performance, ensuring secure, GDPR-compliant data handling. 🔹 Strong in SQL, Python (PySpark), and Databricks, automating complex data workflows and enabling seamless data integration across platforms. 🔹 Experience in building AI-powered solutions, including NLP-based chatbots for SQL generation and data-driven insights using machine learning models. 🔹 Data Visualization & Reporting – Designed interactive Power BI dashboards for executives, tracking KPIs, sales performance, and product analytics. 💡 Services I Offer: ✅ Databricks ETL Pipeline Development (Delta Lake, PySpark, SQL) ✅ Data Engineering & Cloud Integration (AWS, Azure, APIs, JSON, S3) ✅ Optimized SQL Query Writing & Performance Tuning ✅ Automated Data Pipelines & Workflow Orchestration ✅ Data Modeling & Warehouse Design ✅ Power BI Dashboards & Reports (for business insights) 📩 Looking for an experienced Data specialist to streamline your data workflows and visualise actionable insights? Let’s connect and experience never-seen-before professionalism and craftsmanship in every aspect of your project!Pyspark
Microsoft ExcelDatabase ReportData VisualizationAWS LambdaAWS GlueSQLMicrosoft Power BIPythonPySparkDatabricks PlatformETL - $10 hourly
- 0.0/5
- (0 jobs)
I am highly motivated and dedicated to gain more experience and work hard towards completing my task perfectly within the deadlines. Interested in Data Analytics, Data Engineering, Data Science, Finance, Management, and other roles that can add value to both parties.Pyspark
TutoringManagement SkillsTypingData EntryData EngineeringPySparkFinancial AnalysisOperational PlanningBusiness Process ManagementDatabricks PlatformPythonSQLMicrosoft Power BIETLData Analysis - $14 hourly
- 0.0/5
- (0 jobs)
I am a data engineer specialising in ETL. I have experience in python, spark and ETL tools like Informatica. I have more than six years of experience in the field.Pyspark
Data ExtractionPySparkPythonDatabaseETL PipelineETL - $30 hourly
- 0.0/5
- (0 jobs)
Full Stack Data Engineer | End-to-End Data Solutions | Reliable. Scalable. Insightful. Hi! I’m a versatile Full Stack Data Engineer with a strong foundation in both backend development and data infrastructure. I help businesses make smarter decisions by building reliable data pipelines, managing cloud-based data systems, and delivering clean, actionable insights. What I bring to the table: ✔️ Data pipeline design & automation ✔️ Database development & optimization ✔️ Cloud & on-premises data integration ✔️ Scalable backend services ✔️ Interactive dashboards & data apps ✔️ Strong focus on performance, quality, and results Whether you're starting from scratch or need to level up your existing data systems, I can jump in and drive your project forward—end to end. Let’s connect and turn your data into a powerful assetPyspark
Data ExtractionGenerative AIMongoDBTalend Open StudioETL PipelinePySparkBig DataData MigrationLinuxUnix ShellOracleSQLSnowflakePython - $15 hourly
- 5.0/5
- (2 jobs)
Overall 6 years of experience in IT industry with 4 years of relevant experience in Big Data Engineer, handling and transforming heterogeneous data into Key information using Hadoop ecosystem. - Expertise with the tools in Hadoop Ecosystem – HDFS, Hive , Sqoop, Spark, Kafka, Nifi. - Experience working with Elastic Search, Kibana and good knowledge on Oozie, Hbase, Phonix. - Good understanding of distributed systems, HDFS architecture, internal working details of MapReduce, Yarn and Spark processing frameworks. - More than two year of hands on experience using Spark framework with Scala. - Expertise in Inbound and Outbound (importing/exporting) data form/to traditional RDBMS using ApacheSQOOP. - Extensively worked on HiveQL, join operations, writing custom UDF’s and having good experience in optimizing Hive Queries. - Experience in data processing like collecting, aggregating, moving from various sources using Apache Nifi and Kafka. - Worked with various formats of files like delimited text files , JSON files, XML Files - Having basic knowledge on Amazon Web Services.Pyspark
Data LakeAWS CloudFormationAWS GlueElasticsearchKibanaSqoopApache NiFiPySparkScalaSQLApache HadoopApache KafkaApache HiveApache Spark - $40 hourly
- 0.0/5
- (0 jobs)
With over 16 years of experience in Information Technology and having played a Technical Specialist/Architect role for various assignments with cutting edge technologies involving Application Design, Development, Functional Studies and component development on projects for clients such as Nokia-Siemens (Velocix), Verizon Business, Walgreens, and Qatar Airways, Citi. Currently working as Product Architect for Media product company (Velocix) with exposure to SAAS based product on public cloud (AWS).Pyspark
Apache SolrKubernetesRancherSpring BootJavaApache KafkaSpring AMQPPySpark - $50 hourly
- 0.0/5
- (0 jobs)
Intermediate Data Engineer I'm a mid level Data Engineer who finds data is the new gold landscape intriguing and exciting. I have been a python developer and is into ETL. I use native python coding, pyspark for any distributed processing and well versed with some of the services offered by AWS. I have mid level knowledge with snowflake Data warehouse and some good hands on experiene with postgresql.Pyspark
Machine Learning ModelETLPySparkData EngineeringPostgreSQLSnowflakePython - $15 hourly
- 5.0/5
- (15 jobs)
6+ years of experience in Data Science and Machine learning 6+ years of experience as a Python and R Developer - Data Analysis, data cleansing, Exploratory Data Analysis(EDA), Dimensionality reduction, Feature scaling, Modeling, Model evaluation, and Hyper-Parameter tuning. - data analysis and visualizations using Tableau, Python, and R. - NLP(Natural Language Processing) with packages like NLTK, Spacy, Bag of Words, Scikit Learn, etc. - mining PDF using PDF Plumber, PYPDF2, Spacy PDF reader, OCR, and Tesseract. - PDF extraction through LLM and Generative AI like ChatGPT - Time series analysis, Neural networks, Deep learning. - Experience in Content writing on Data Science and AI. - Strong written and oral communication skills, Permanent Freelancer. Always available 24x7.Pyspark
PySparkScalaRWeb DesignSEO ContentLLM Prompt EngineeringGenerative AIMachine LearningData SciencePython - $30 hourly
- 0.0/5
- (0 jobs)
I am a Microsoft certified Data Engineer and Data Analyst, with Experience in Python , LLM, AI,Machine learning ,deeplearning, AI agents SQL,Power Bi, Microsoft Fabric and Tableau. i can work with data end to end, whether it can be building data pipelines or data flows or data warehousing, or building AI models all rounder in any thing dataPyspark
Data LakeData EngineeringPySparkPythonSQLLLM PromptDeep LearningMachine LearningTableauMicrosoft Power BIBusiness IntelligenceData Analysis - $10 hourly
- 0.0/5
- (0 jobs)
Your Pocket Data Developer . Anytime whenever you have Emergency Sr. Data Engineer with wide knowledge of MSSQL, SSIS, SSRS and Power BI . Key Skill. Data Integration, Data transformation and warehousing Over the last 9+ years, I am senior consultant for all MS SQL services. I have experience in the following areas: MSSQL Server, SSRS, SSIS, power BI. My core competency lies in complete solution design for Database development project. I am seeking opportunities to build these services from the ground up for you or your business. Also I have hand-on experience with azure in different areas (IAAS, PAAS) Key Skills SSIS ETL development and improvement API integration and delta load Database development Database Performance tuning Database 24/7 Monitoring Support. Database Consulting Power BI Management Technologies MS SQL server SSIS SSRS Power BIPyspark
PySparkAWS GlueAmazon S3Amazon RedshiftDatabricks PlatformUnixMicrosoft Power BISQL ProgrammingSQL Server Reporting ServicesSQL Server Integration ServicesDatabaseGitHubPython - $15 hourly
- 0.0/5
- (1 job)
As a Senior Data Engineer my skills in AWS, SnowFlake, and SQL to design, develop, and optimize data solutions for the automotive industry. I have over 5 years of experience in data engineering, working with IT OPENDOORS PVT LTD as an Azure Data Engineer and a Power BI Developer . I have contributed to multiple projects involving data integration, transformation, visualization, and analysis, supporting business intelligence, analytics, and reporting needs. I am passionate about leveraging data to drive innovation, efficiency, and value for the automotive sector, and I am always eager to learn new technologies and best practices. I am a team player, a problem solver, and a customer-focused professional, who strives to deliver high-quality data solutions that meet or exceed expectations.Pyspark
PySparkData VisualizationMachine LearningpandasSQLTalend Open StudioSnowflakeDatabricks PlatformAWS GlueMicrosoft Power BIData EngineeringMicrosoft Azure - $15 hourly
- 0.0/5
- (0 jobs)
I'm a DataOps and Data Platform Engineer experienced in building Data Platform environment, Databricks workspace on Azure and AWS platform. Experienced in ETL Pipeline using Azure Data Factory, Azure Databricks and Azure Synapse Analytics. Experienced in Devops(Azure and AWS) using Github, Azure devops, Github actions.Pyspark
Architectural DesignLinux System AdministrationPySparkPythonCI/CDMicrosoft AzureDockerKubernetesAzure DevOpsTerraformDatabricks PlatformDevOps - $20 hourly
- 0.0/5
- (0 jobs)
Big Data developer specializing in Performance tuning and writing efficient code particularly in apache sparkPyspark
Data ExtractionETLClouderaScripting LanguageUnix ShellApache SparkPySpark - $3 hourly
- 0.0/5
- (0 jobs)
I'm Venkatesh Anantharaman, an Experienced data analyst with a dynamic background in engineering and a knack for uncovering insights that drive impactful decisions. I thrive on solving complex problems with data-driven strategies. What I Bring to the Table? 1. Data Visualization: Crafting visuals that distill complex datasets into clear, actionable insights. 2. Data Analytics: Diving deep into numbers to uncover trends, patterns, and opportunities. 3. Data Cleaning & Modeling: Ensuring data integrity through standardization, deduplication, and other essential techniques for high-quality analysis. Reporting & Dashboarding: Developing dashboards that tell a story, making data engaging and informative. 4. Business Intelligence: Enabling strategic decisions through data-driven insights. Data Storytelling: Transforming numbers into narratives that resonate with audiences.Pyspark
Analytical PresentationPySparkMicrosoft Power BIData VisualizationData AnalysisBusiness AnalysisPower QueryData ExtractionETL Want to browse more freelancers?
Sign up
How hiring on Upwork works
1. Post a job
Tell us what you need. Provide as many details as possible, but don’t worry about getting it perfect.
2. Talent comes to you
Get qualified proposals within 24 hours, and meet the candidates you’re excited about. Hire as soon as you’re ready.
3. Collaborate easily
Use Upwork to chat or video call, share files, and track project progress right from the app.
4. Payment simplified
Receive invoices and make payments through Upwork. Only pay for work you authorize.
How do I hire a Pyspark Developer near Chennai, on Upwork?
You can hire a Pyspark Developer near Chennai, on Upwork in four simple steps:
- Create a job post tailored to your Pyspark Developer project scope. We’ll walk you through the process step by step.
- Browse top Pyspark Developer talent on Upwork and invite them to your project.
- Once the proposals start flowing in, create a shortlist of top Pyspark Developer profiles and interview.
- Hire the right Pyspark Developer for your project from Upwork, the world’s largest work marketplace.
At Upwork, we believe talent staffing should be easy.
How much does it cost to hire a Pyspark Developer?
Rates charged by Pyspark Developers on Upwork can vary with a number of factors including experience, location, and market conditions. See hourly rates for in-demand skills on Upwork.
Why hire a Pyspark Developer near Chennai, on Upwork?
As the world’s work marketplace, we connect highly-skilled freelance Pyspark Developers and businesses and help them build trusted, long-term relationships so they can achieve more together. Let us help you build the dream Pyspark Developer team you need to succeed.
Can I hire a Pyspark Developer near Chennai, within 24 hours on Upwork?
Depending on availability and the quality of your job post, it’s entirely possible to sign up for Upwork and receive Pyspark Developer proposals within 24 hours of posting a job description.