Hire the best Apache Spark Engineers in Hyderabad, IN

Check out Apache Spark Engineers in Hyderabad, IN with the skills you need for your next job.
Clients rate Apache Spark Engineers
Rating is 4.7 out of 5.
4.7/5
based on 283 client reviews
  • $35 hourly
    ════ Who Am I? ════ Hi, nice to meet you! I'm Ajay, a Tableau and SQL Specialist, Business Intelligence Developer & Data Analyst with half a decade of experience working with data. For the last few years I've been helping companies all over the globe achieve their Data Goals and making friends on the journey. If you're looking for someone who can understand your needs, collaboratively develop the best solution, and execute a vision - you have found the right person! Looking forward to hearing from you! ═════ What do I do? (Services) ═════ ✔️ Tableau Reports Development & Maintenance - Pull data from (SQL Servers, Excel Files, Hive etc.) - Clean and transform data - Model relationships - Calculate and test measures - Create and test charts and filters - Build user interfaces - Publish reports ✔️ SQL - Build out the data and reporting infrastructure from the ground up using Tableau and SQL to provide real time insights into the product and business KPI's - Identified procedural areas of improvement through customer data, using SQL to help improve the probability of a program by 7% - Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs and Scala. ═════ How do I work? (Method) ═════ 1️⃣ First, we need a plan; I will listen, take notes, analyze and discuss your goals, how to achieve them, determine costs, development phases, and time involved to deliver the solution. 2️⃣ Clear and frequent communication; I provide frequent project updates and will be available to discuss important questions that come up along the way. 3️⃣ Stick to the plan; I will deliver, on time, what we agreed upon. If any unforeseen delay happens, I will promptly let you know and provide a new delivery date. 4️⃣ Deliver a high-quality product. My approach aims to deliver the most durable, secure, scalable, and extensible product possible. All development includes testing, documentation, and demo meetings.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Apache Hive
    Python Script
    Scala
    Machine Learning
    Hive
    SQL Programming
    Business Intelligence
    Microsoft Excel
    Microsoft Power BI
    Tableau
    SQL
    Python
  • $60 hourly
    Nikhil is a Microsoft certified azure data engineer with 5+ years of experience in data engineering and big data. Have worked for couple of fortune 500 companies for developing and deploying their data solutions in azure and helped them find business insights out of their data. Coding: - SQL, Python, Pyspark Azure: - Azure Data Factory - Azure Databricks - Azure Synapse Analytics - Azure Datalake - Azure Functions and other azure services Reporting: - Power BI - Microsoft Office
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    ETL
    Microsoft Azure
    Data Lake
    Data Warehousing
    Microsoft SQL Server
    Big Data
    PySpark
    Databricks Platform
    SQL
    Apache Hive
    Python
    Microsoft Excel
    Data Engineering
    Data Integration
  • $125 hourly
    A Computer Science Engineering leader with 12+ yrs. of experience in gathering requirements and translating them into functional specifications and developing/designing code. I'm experienced in creating microservices/REST API's using spring boot. I can also create data engineer pipeline using apache spark, kafka. I can also create pipelines to analyze your data using OLAP.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Docker Compose
    Snowflake
    Apache Kafka
    Data Engineering
    Microsoft Azure
    Kubernetes
    Apache Maven
    Spring MVC
    Docker
    Spring Framework
    Gradle
    Java
    Microservice
    Spring Boot
  • $60 hourly
    I am an expert python developer with 13 years of experience in building machine learning models , data visualization , web and rest api applications. I have worked with Investment Banks , Startups and consulted for Technology services companies. I Specialize in : - Data Analytics and Machine Learning with Apache Spark and TensorFlow - Web Scrapping and Data Mining - Data Visualization - Back end Development with Python / Django/ Flask - Mysql , MongoDB With Computer Science and Math Masters degree and extensive experience in agile development environment , I have the necessary skill set and problem solving abilities to get your job done and deliver on the expectations.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    MongoDB
    Django
    Flask
    Machine Learning
    Python
    Deep Learning
    Keras
    PyTorch
    TensorFlow
  • $70 hourly
    • A creative hands-on with around 12 years of experience, exceptional technical skills, and a business-focused outlook. Adept in analyzing information system needs, evaluating end-user requirements, custom designing solutions for complex information systems management • Vast experience in data driven applications ,creating data pipe lines, creating interfaces between up-stream and down-stream applications and tuning the pipe lines. • Interacting with business team to discuss and understand the data flow and designing the data pipelines as per the requirements. • Experience in driving the team to meet the target deliverables. Strong experience in creating scalable and efficient big data pipelines using Spark, Hadoop, Hive, Pyspark, Python,Snowflake,DBT and Airflow • Commendable experience in cloud data warehousing SNOWFLAKE. . Experience in snowflake development , data sharing, advanced features of snowflake Strong experience in integrating snowflake with DBT and creating data layers on the Snowflake warehouse using DBT • Expertise skill in SQL • Have strong exposure in PYTHON. • Strong experience on Hadoop • Strong experience in implementing ETL pipelines using SPARK. • Strong experience in tunings the SPARK applications. • Extensively used SPARK SQL to clean the data and to perform calculations on datasets. • Have strong experience in HIVE. • Strong experience in HIVE query tuning. • Worked on different big data file formats such as Parquet, ORC, etc.. • Familiar with AZURE Databricks. Decent exposure on Airbyte, Bigquery, Terraform • Expertise in Analytical functions. • Have strong exposure in converting the data into business insights. • Decent knowledge on data lake and data marts concepts. • Experience in Creation of Tables, Views, Materialized Views, Indexes using SQL and PL/SQL. • • In-depth knowledge of PL/SQL with the experience in constructing the tables, joins, sub queries and correlated sub queries In SQL * Plus. • Proficient in Developing PL/SQL Programs Using Advanced Performance Enhancing Concepts like Bulk Processing, Collections and Dynamic SQL • Sound knowledge in using Oracle materialized views • Effectively made use of Indexes, Collections, and Analytical functions • Sound knowledge in using Oracle SQL Loader and External Tables. • Has good knowledge and exposure in designing and developing user defined stored procedures and user defined functions. • Having experience in using Packages UTL_FILE, DBMS_JOB and DBMS_SCHEDULE. • Skilled in handling critical application and business validation oriented trigger logic. • Has good Knowledge in trapping runtime errors by providing Suitable Exception Handlers.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Apache Airflow
    Apache Hive
    Databricks Platform
    Python
    Apache Hadoop
    PySpark
    Snowflake
    Amazon S3
    dbt
    Database
    Oracle PLSQL
    Unix Shell
  • $20 hourly
    I am a Big data engineer with over 7 years of professional experience in the industry. I am very proficient in Apache Spark, NiFi, HDFS ecosystem, Kafka , Java, Python Design. I have a lot of background in both software development, building ETL pipelines and data lake.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Apache NiFi
    Apache Hadoop
    Apache Kafka
    Python
    MongoDB
    Spring Boot
    Django
  • $10 hourly
    Technical Experience * Hands on experience in Hadoop Ecosystem including Hive, Sqoop, MapReduce and basics of Kafka * Excellent knowledge on Hadoop ecosystems such as HDFS , Resource Manager, NodeManager , Name Node, Data Node and Map Reduce programming paradigm * Expertise in managing big data processing using Apache Spark and its various components * Load and transform large sets of structured, semi-structured and unstructured data from Relational Database Systems to HDFS and vice-versa using Sqoop tool. * Data ingestion and refresh from RDBMS to HDFS using Apache Sqoop and processing data through Spark Core and Spark SQL * Proficiency in Scala and Pyspark required for high level data processing and have end-to end knowledge for implementation of a project * Designing and creating Hive external tables, using shared meta-store instead of Derby, and creating Partitions and Bucketing
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Amazon Web Services
    Visualization
    Apache Kafka
    Apache Hive
    SQL
    Apache Hadoop
  • $30 hourly
    Data Engineer with 4 years of experience handling Client's healthcare data by migrating & transforming it from Teradata to AWS RDS by building ETL Pipelines and Providing insights by generating multiple reports using PySpark/Python and SQL. * Expert in Python scripting, PySpark & SQL
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Databricks Platform
    Amazon S3
    Teradata
    Git
    MySQL
    Amazon Web Services
    PostgreSQL
    Data Engineering
    PySpark
    AWS Glue
    AWS Lambda
    ETL Pipeline
    ETL
    Python
  • $10 hourly
    • IT Professional around 6.1 years of experience in Software Development and Maintenance of Big Data projects • Possess in-depth working knowledge in all the areas of development of Big Data • Worked extensively on Technologies like Apache Spark, Databricks, Hive, Sqoop, Map Reduce, Apache Kafka applications.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Sqoop
    Hive
    Apache Kafka
    SQL
    Python
    PySpark
  • $120 hourly
    With over 12 years of experience, of which about 8yrs I have worked with different Bigdata technologies(Hadoop, Spark) and the remaining time I mostly worked on writing python scrappers, scripts, API services and also built iOS applications using Objective-C - Experience in building data pipelines to process Petabyte scale data and optimise them for cost and performance - Experience in fine tuning the Spark jobs to the most optimal level and thereby cutting down infrastructure costs by 50-80% - Experience with building Data lakes for major e-commerce and fintech companies - Worked at different startups throughout my career and highly adaptable to different working methodologies like Agile and Kanban
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Big Data
    Apache Hadoop
    PySpark
    Scala
    Python
  • $16 hourly
    Microsoft certified Data engineer. My Area of expertise is as follows - Azure synpase, ADF, Stream analytics, Pyspark and spark streaming - Python, Flask, MySQL, PostgreSQL Web scraping using scrapy - Jupyter notebook, ML libraries like pandas, matplotlib, numpy - Data pipelines using Apache airflow - Backend APIs using express, sequelize, javascript - Containerization of Applications using Docker, Docker swarm and Docker compose - Monitoring Dashboards using Prometheus and Graphana It's important to me to build long term relationships with clients, so I'm primarily looking for long term projects. I'm flexible with my working hours and am happy to work closely with any existing freelancers you work with. I look forward to hearing from you!
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Microsoft Azure
    MongoDB
    Node.js
    Docker
    MySQL
    ETL Pipeline
    Apache Airflow
    SQL
    Scrapy
    Python
  • $10 hourly
    PROFESSIONAL SUMMARY Having 3 Years 3 months of Total Experience in IT Industry with 3 Years of project experience in Data warehouse development and enhancement using Microsoft SQL Server, Azure Data Factory, Azure Data Bricks . Good Knowledge in analyzing data and writing SQL queries according to the requirement. Skilled in mapping clients' business requirements and translating the same into functional specifications, services, implementations and custom designing solutions. Experienced in resolution of production tickets related to Data Issues, Design changes, ETL issues and performance tuning on SQL server. Handled many Change Requests/BUGS which involves Data analysis, Problem identification, code change, Testing and finally deploying the code. Willing to accept responsibility and commit to working as part of a team to utilize my experience and skills in contributing
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Microsoft Azure
    Microsoft Azure SQL Database
    Microsoft SQL Server
    Microsoft SQL Server Programming
    SQL
    ETL Pipeline
  • $6 hourly
    As Azure Data Engineer I work on building data pipelines utilizing Azure Data Bricks + Pyspark + SQL + Azure Data Factory+ Azure synapse Analytics, and ingesting data to and from ADLS, Oracle DB, and Azure SQL DB after necessary data transformations. Track and control the daily loads in the data pipeline. As a developer, it is my responsibility to participate in the project by gathering business clients' requirements. analyzing those specifications in order to prepare and provide the clients with a data solution. Making plans for the creation of data pipeline solutions. implementing UAT and transferring code to the live environment. Maintaining contact with clients, providing them with emergency support, and supporting the team I am working on the retail and finance project, which involves importing CSV and Excel files containing invoice data into the ADLS. storing the data in a staging area known as an Azure SQL Database, transforming it with Pyspark + SQL in Azure Data Bricks, and then loading it into the designated target location—such as an Azure SQL Database or Dedicated SQL Pool—as needed by the business. Azure Data Factory is used to orchestrate and structure all of the aforementioned data pipelines. I'm working on a big data project that uses Azure data bricks to access all of Spark's capabilities.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    ETL Pipeline
    Azure Cosmos DB
    Data Ingestion
    Data Lake
    Unix
    pandas
    Microsoft Power BI
    Microsoft Azure SQL Database
    IBM DataStage
    Data Warehousing & ETL Software
    SQL
    Python
    PySpark
    Data Engineering
  • $15 hourly
    I am a cloud specialist with expertise in AWS and GCP. I build data pipelines using Airflow. I come with 11 years of experience into Bigdata and Databases. I can help you solve your Big data problems or create your growing businesses ready for scaling using AWS, GCP solutions. AWS: EMR, Glue, Redshift, lambda, s3 GCP: dataproc, bigquery, google buckets, dataproc others: Airflow, python, Nifi, Airbyte, SQL, docker
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    AWS Lambda
    ETL
    AWS Glue
    Apache NiFi
    Apache Airflow
    PySpark
    Python
  • $30 hourly
    About Me: Hello! I am Anirvan Sen, I am a data engineer with over 7 years of IT experience, and I bring a proactive 'get stuff done' attitude to every project. Passionate about tackling challenging assignments and specializing in engineering end-to-end solutions. 🔥 Skills: • Languages: Proficient in Python and Shell Scripting • Databases: Skilled in MySQL, PostgreSQL, and Elastic Search • Tools & Frameworks: Experienced with Spark, Hive, Airflow, Databricks, Elastic search, Logstash, Terraform, DBT, GIT, MLflow, and more • Database & Analytics: Expertise in Data Modelling, Data Warehousing, ETL, SQL, and MLOps • AWS: Deep understanding of AWS services including S3, Lambda, Step function, App Sync, API gateway, Sage maker, Cloudformation, and others • Visualization: Proficient in Kibana for data visualization 🔥 Highlighted Projects: • Established Real-time Monitoring System: Spearheaded the development of a centralized real-time logging and alerting system using the ELK stack. Monitored critical systems such as Airflow, Databricks, and various applications within the data platform. • Databricks Disaster Recovery Solutions: Led the design and implementation of comprehensive disaster recovery strategies for Databricks, covering data replication, workflows, S3 backup, and job replication. • Automated Onboarding Processes: Revolutionized onboarding procedures by automating the creation of Databricks notebooks, Lambda workflows, and container workflows. Reduced onboarding time from one day to just 15 minutes. • Rule-Based Framework Development: Developed a rule-based framework using PySpark, significantly reducing the time required for data quality and processing pipeline development from weeks to days. • MLOps Strategy and Deployment: Led MLOps strategy and deployment initiatives, specializing in the seamless implementation of models for real-time inference on Amazon SageMaker. 🟢 Why Me? You'll benefit from my diverse skill set encompassing data engineering, DevOps, and full-stack engineering. I'm committed to delivering top-notch results for your data engineering projects, ensuring efficiency and excellence every step of the way. 🔥 Certifications: AWS Certified Solutions Architect-Associate AWS Certified Developer HashiCorp Certified: Terraform Associate Databricks Certified Spark Associate Developer CCA Spark and Hadoop Developer – Cloudera Feel free to reach out to discuss how I can contribute to your project's success!
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Amazon SageMaker
    MLflow
    MLOps
    CI/CD
    Elasticsearch
    Apache Airflow
    Data Warehousing
    Data Modeling
    Terraform
    Amazon Web Services
    ETL Pipeline
    SQL
    Python
    Databricks Platform
  • $30 hourly
    Professional Summary * Having around 9 years of experience.3 years in DataBricks and Azure DataFactory and 5 years in Db2 UDB,Oracle and Sql Server Database Development and Support. * Experience of working on Azure Data Factory, Data Bricks,Pyspark,Sparksql * Created Linked Services for multiple source system (i.e.: Oracle, SQL Server, ADLS, BLOB, File Storage and Table Storage). * Created Pipeline's to extract data from on premises source systems to azure blob storage; Extensively worked on copy activities. * Exposure on Azure Data Factory activities such as Lookups, Stored procedures, if condition, for each, Set Variable, Append Variable, Get Metadata, Filter and wait. * Extensively used ETL to load data from flat file and also from the relation database. * Experience on played key technical roles in developing the Business Intelligence solutions using Microsoft Azure services of Azure DataFactory ,Azure Databricks, Azure Data Lake, Azure SQL
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    SQL
    SQL Programming
    Python
    Python Script
    PySpark
    Apache Spark MLlib
    Databricks Platform
  • $25 hourly
    i have total experience 2.8 experience in bigdata developer in sql, spark, hive ,hadoop,aws and snowflake.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    HDFS
    Apache Hadoop
    PySpark
    Hive
    Snowflake
    SQL
  • $25 hourly
    Data Engineer : +91-9003128991 Offering 11 years of experience can be headhunted for a Lead level position across any functional sector within an IT organization of repute. Experienced professional with a strong background in both Azure Data Engineering and MSBI development. Skilled in designing scalable data architecture, implementing ETL processing, and Data warehousing by utilizing Azure Cloud services for efficient data pipelines. Seeking a challenging role where I can leverage my experience and technical skills to drive data-driven decision-making and contribute to the organization's success.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Microsoft Azure
    ETL
    Data Lake
    Data Extraction
    ETL Pipeline
  • $10 hourly
    I am a data science enthusiast where iam has been working in the industry for a couple of years and have done a few projects to gain some good insights here are a few Projects __________________________________________________________________________________________ * Easy Tool for File Management: A file management system using Python and MySQL can provide a secure and scalable solution for managing files. The system will provide key features such as file upload, download, deletion, and search, and will be developed using a technology stack that is both efficient and easy to work with. Proper attention must be paid to security, scalability, and user experience to ensure the system's success. (04/2022) * Terror attack Prediction: "Designed, developed, and deployed a binary classification model to predict the severity of terror attacks using [chosen algorithm]. Implemented data preprocessing techniques to enhance model accuracy. Successfully deployed the model for real-time predictions, demonstrating proficiency in end-to-end machine learning project development. Achieved [mention key metrics, e.g.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Apache Administration
    Apache HTTP Server
    Microsoft Project
    Analytical Presentation
    Business Management
    Management Skills
    Data Analysis
    Information Analysis
    Apache Spark MLlib
    Project Management
  • $15 hourly
    Professional Summary: * Over all 12 years of professional IT experience in ETL with tools like SQL, Informatica, Databricks, Python, PySpark, Hadoop, Hive, AWS * 2+ years of experience in working as a Scrum master. * 4+ Years of experience in AWS (S3, EC2, SNS, SQS, Lambda). S3 as the intermediate storage layer while extracting data. SQS & Lambda to read the events from queue and check ingestion status with endpoint being SNS. * 5.5+ years of experience in Databricks & PySpark used for data & code migration from Db2/Oracle to Data lake. Informatica mappings converted PySpark notebooks. * 3+ years of experience in Python, created notebooks for to handle parallel execution of extraction, transformation & ingestion. * 5+ years of IT experience in development and design of ETL methodology for supporting data transformations & processing in a corporate wide ETL Solution using Informatica PowerCenter
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    SQL
    Python
    PySpark
    Databricks Platform
  • $5 hourly
    Experienced freelance data engineer skilled in designing and implementing scalable data pipelines and infrastructure. Proficient in ETL development, data architecture design, and cloud platform management. Proven track record of optimizing performance, ensuring data quality, and maintaining security standards. Strong collaboration and communication abilities, adept at understanding client requirements and delivering tailored solutions. Continuously updating skills and staying abreast of industry trends. Dedicated to empowering organizations by unlocking the value of their data through reliable and efficient data engineering solutions.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    PySpark
    Python
  • $25 hourly
    Core Expertise and offerings: Cloud Architiect with consulting experience to offer Operational Excellece, Security, Reliability, Performance Efficiency and Cost Optimization for Business IT. Expert in ETL tool Matillion and Cloud warehouse Snowflake. My role as Cloud Architect include to pick right architecture, evaluate cost, planning failover/High Availability/Disaster Recovery, designing models for End to End Migrations, Load Balancing for traffic management , Backup and Archival strategy for new and old infra, monitoring tools to keep track of infra health. Other Offerings: Project Management: Tracking of project activities and timely mis report to senior management with KPIs Spread Sheet/Grid Solutions: Expert in Excel VBA /Google Sheets App Scripts Certifications: I am a certified "Mastering Data Analyst" from Duke University. I am a certified "Software Engineer" From Indian Institute of Technology, Madras" in IOT, Cloud and Block Chain. Skills: Cloud: Architecting applications for the Cloud, Containers & Microservices, DevOps for the Cloud IOT: IoT Data Sources & Networking. IoT on Cloud, IoT Big Data Processing, IoT Analytics Block Chain: Fundamentals of Blockchain, Blockchain Development
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Data Extraction
    Spreadsheet Software
    English
    Blockchain
    Accuracy Verification
    ITIL
    Smart Contract
    Macro Programming
    MongoDB
    Amazon Web Services
    Google Cloud Platform
    Python
  • $40 hourly
    PROFILE Data Engineer with 3 Years of Experience in building scalable ETL pipelines , utilizing technologies like Azure Data Factory, Databricks, Azure Data Lake, Apache Spark, Snowflake, Hive, Hadoop, Scala, Python, SQL etc. -Developed an Automation Frameworks using python to streamline Data Quality and comparing the billions of records data from the two different environment. - Experience working with a variety of data types, including Retail, Shopper and Advertisement data. - Optimized the ETL pipelines and achieving over a 50% speed improvement. - I had a experience in working Agile software development environment.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Agile Software Development
    Data Warehousing
    ETL Pipeline
    Snowflake
    Databricks Platform
    SQL
    Hive
    Python
    PySpark
  • $30 hourly
    Data Analytics consultant. Strng Data Architecture, Designing, MVP expert. Expert skills in python, PySpark, Data migration, Advisory & Consulting, Machine Learning and cloud & Hybrid experience. TechStacks: Python, Pyspark,Scala,Kaftka,AWS,Azure,Docker,Kubernetes and NoSQL - Architect, Designer with Strong Data Anlytics experinece. - Building Data pipeline, Data migration skills. - Worked with ETL tools and Big data, hadoop with Spark experience. - Build insights using Python programming with various tools, platforms, and systems such as: Apache Spark, Kafka, HBase, Redis, Akka, etc. - Working in Containerzation Docker, Kubernetes, CI/CD. - Development using a variety of languages which may include Python, Scala, Java Script, Hive, Spark, sqoop, node.js, Shell Scripting - Strong experience on HBase, Casandra & Elastic search - Design, develop, and implement complex reports and analytics - Use advanced data and coding techniques to provide concise and compelling summary of analysis findings in reports and presentations - Excellent programming skills on the JVM using Java/Scala. - Comfortable working with SQL/NoSQL. - Comfortable working in Unix CLI and with cloud infrastructure. - An advocate of agile practices for rapid development of quality software, such as CI, TDD and automated deployment. - Write scala programs that scales to petabytes of data and supports millions of transactions a second - Write and review pull requests in Git - Leverage various data and technology resources to augment analysis - Experience with Spark or the Hadoop ecosystem and similar frameworks is ideal for this role. - Experience with Kafka is ideal for this role.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Amazon Web Services
    Apache Hadoop
    Microsoft Azure
    AWS Glue
    Akka
    Snowflake
    Looker Studio
    BigQuery
    Google Analytics
    Big Data
    Apache Hive
    Cloudera
    Scala
  • $20 hourly
    Happy Day. My name is Rajkiran K. Over 11+ years of comprehensive hands-on Big Data Capacity with Hadoop Eco System access internal and cloud-based platforms. I like to go into customer or client problems and make an effort to use cutting-edge technology stacks to solve them. The programming approach is mostly focused on scalable, effective, and efficient solutions. I frequently contribute to the application's effective setup and layout. I adore taking on challenging jobs combining big data and cloud computing technology. Proficiency with the Hadoop architecture, including its different parts (Hive, Pig, Sqoop, HDFS, MapReduce, Spark, Impala, and HBase), as well as cloud computing. Amazon Web Services - My area of expertise is Amazon Web Services (AWS). Regardless of whether your project calls for end-to-end solutions with a seamless effort, a highly scalable, exceptionally cost-effective strategy, and a long-term sustainable project, architecture, development, engineering, or building. I can assist in your success. I've worked with EC2, ECR, ECS, AWS Frigate, RDS, AWS S3, AWS Glue, AWS EMR, EKS, DMS, Lambda, IAM, VPC, Direct Connect, Route54, CloudWatch, Redshift, Athena, Redshift Spectrum, Kinesis, Amazon Segamaker, and other AWS services. Confluent Kafka - As part of real-time data processing, I have also worked on Confluent Cloud services, constructing Kafka topics using ACL, Schema Registry, KSQL, and Kafka Connect services, as well as building producer & consumer tasks using Python and Pyspark Structure Streaming Solutions. Having extensive expertise creating pipelines for real-time and batch processing solutions using tools like Pyspark, Scala with Spark, Apache Kafka, Confluent Kafka, etc. Project Management - I've used tools like Trello, Monday, Jira, Confluence, etc. CI/CD Pipeline - I have a decent amount of expertise using Docker, Kubernetes, Terraform, Jenkins, Github, BitBucket, etc.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    ETL Pipeline
    Big Data
    Scala
    Amazon Web Services
    Amazon ECS
    Amazon S3
    AWS Glue
    PySpark
    Amazon EC2
    Python
    SQL
    Apache Hadoop
    Databricks Platform
    Apache Kafka
  • $10 hourly
    I am a DATA ENGINEER skilled in transforming business requirements into detailed technical solutions to achieve business goals. I possess diverse experience in planning & executing multiple projects that solve various business problems and deliver excellent results. I am willing to go that extra mile to overcome any hurdles I may face. I like working with data to understand, analyse and transform it in a way that it can be easily consumed. I also like learning new technologies and ways that would help me improve my day-to-day tasks.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Amazon Web Services
    Microsoft Azure
    PySpark
    Apache Airflow
    Scala
    Apache Hadoop
  • $5 hourly
    I'm a Data Engineer with expertise in Python, SQL, and Big Data technologies. Proficient in managing cloud platforms such as Microsoft Azure and Google Cloud Platform, I actively contribute to diverse projects, specializing in the development of robust data solutions for small and medium-sized businesses. My skills include ETL, optimizing data processing, generating insightful reports, and collaborating effectively within cross-functional teams. Committed to continuous learning, I am dedicated to professional growth in the dynamic field of data engineering.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    ETL
    Data Lake
    Big Data
    Looker Studio
    Microsoft Power BI
    Azure Cosmos DB
    Microsoft Azure
    Snowflake
    PySpark
    Databricks Platform
    Python
    Microsoft Excel
    SQL
    Apache Hadoop
  • Want to browse more freelancers?
    Sign up

How hiring on Upwork works

1. Post a job (it’s free)

Tell us what you need. Provide as many details as possible, but don’t worry about getting it perfect.

2. Talent comes to you

Get qualified proposals within 24 hours, and meet the candidates you’re excited about. Hire as soon as you’re ready.

3. Collaborate easily

Use Upwork to chat or video call, share files, and track project progress right from the app.

4. Payment simplified

Receive invoices and make payments through Upwork. Only pay for work you authorize.

Trusted by

How do I hire a Apache Spark Engineer near Hyderabad, on Upwork?

You can hire a Apache Spark Engineer near Hyderabad, on Upwork in four simple steps:

  • Create a job post tailored to your Apache Spark Engineer project scope. We’ll walk you through the process step by step.
  • Browse top Apache Spark Engineer talent on Upwork and invite them to your project.
  • Once the proposals start flowing in, create a shortlist of top Apache Spark Engineer profiles and interview.
  • Hire the right Apache Spark Engineer for your project from Upwork, the world’s largest work marketplace.

At Upwork, we believe talent staffing should be easy.

How much does it cost to hire a Apache Spark Engineer?

Rates charged by Apache Spark Engineers on Upwork can vary with a number of factors including experience, location, and market conditions. See hourly rates for in-demand skills on Upwork.

Why hire a Apache Spark Engineer near Hyderabad, on Upwork?

As the world’s work marketplace, we connect highly-skilled freelance Apache Spark Engineers and businesses and help them build trusted, long-term relationships so they can achieve more together. Let us help you build the dream Apache Spark Engineer team you need to succeed.

Can I hire a Apache Spark Engineer near Hyderabad, within 24 hours on Upwork?

Depending on availability and the quality of your job post, it’s entirely possible to sign up for Upwork and receive Apache Spark Engineer proposals within 24 hours of posting a job description.