Hire the best Apache Spark Engineers in Bengaluru, IN

Check out Apache Spark Engineers in Bengaluru, IN with the skills you need for your next job.
Clients rate Apache Spark Engineers
Rating is 4.7 out of 5.
4.7/5
based on 283 client reviews
  • $15 hourly
    With a Bachelor’s degree in Computer Science, and hands-on experience using JAVA and C++ to create and implement software applications. I work as a Software engineering SDE, in a well known fintech startup , I use JAVA and C++ extensively for my day to day work. Have experience in working with advance BIG DATA frameworks such as Apache Hadoop, Apache Spark and Apace Hive. Works as SME at Chegg where I help students with there doubts and assignments in the field of Computer Science. Have 1yr+ experience in teaching.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    PyTorch
    AWS Development
    Rust
    Golang
    Python
    LLM Prompt Engineering
    Data Engineering
    C++
    Spring Boot
    Apache Hive
    Core Java
    Apache Hadoop
    Data Structures
    MySQL
  • $90 hourly
    I pride myself on achieving a 𝗽𝗲𝗿𝗳𝗲𝗰𝘁 𝗿𝗲𝗰𝗼𝗿𝗱 𝗼𝗳 𝟱-𝘀𝘁𝗮𝗿 𝗿𝗮𝘁𝗶𝗻𝗴𝘀 𝗮𝗰𝗿𝗼𝘀𝘀 𝗮𝗹𝗹 𝗽𝗿𝗼𝗷𝗲𝗰𝘁𝘀. My expertise in 𝗰𝗹𝗼𝘂𝗱 𝗱𝗮𝘁𝗮 𝗲𝗻𝗴𝗶𝗻𝗲𝗲𝗿𝗶𝗻𝗴 and 𝗳𝘂𝗹𝗹-𝘀𝘁𝗮𝗰𝗸 development has been honed through experience with premier institutions like 𝗚𝗼𝗹𝗱𝗺𝗮𝗻 𝗦𝗮𝗰𝗵𝘀, 𝗠𝗼𝗿𝗴𝗮𝗻 𝗦𝘁𝗮𝗻𝗹𝗲𝘆, a member of the 𝗕𝗶𝗴 𝗙𝗼𝘂𝗿 and a 𝗙𝗼𝗿𝘁𝘂𝗻𝗲 𝟱𝟬𝟬 company. With over 9 years of experience in Data Engineering and Programming, I bring a commitment to excellence and a passion for perfection in every project I undertake. My approach is centered around delivering not just functional, but 𝗵𝗶𝗴𝗵𝗹𝘆 𝗲𝗳𝗳𝗶𝗰𝗶𝗲𝗻𝘁 𝗮𝗻𝗱 𝗼𝗽𝘁𝗶𝗺𝗶𝘇𝗲𝗱 code, ensuring top-quality outputs that consistently impress my clients. My expertise combined with extensive experience on both GCP and AWS Cloud platforms, allows me to provide solutions that are not only effective but also innovative and forward-thinking. I believe in going beyond the basics, striving for excellence in every aspect of my work, and delivering results that speak for themselves. 𝗖𝗵𝗼𝗼𝘀𝗲 𝗺𝗲 𝗶𝗳 𝘆𝗼𝘂 𝗽𝗿𝗶𝗼𝗿𝗶𝘁𝗶𝘇𝗲 𝘁𝗼𝗽-𝗻𝗼𝘁𝗰𝗵 𝗾𝘂𝗮𝗹𝗶𝘁𝘆 𝗶𝗻 𝘆𝗼𝘂𝗿 𝗽𝗿𝗼𝗷𝗲𝗰𝘁𝘀 𝗮𝗻𝗱 𝗮𝗽𝗽𝗿𝗲𝗰𝗶𝗮𝘁𝗲 𝗮 𝗳𝗿𝗲𝗲𝗹𝗮𝗻𝗰𝗲𝗿 𝘄𝗵𝗼 𝗮𝘂𝘁𝗼𝗻𝗼𝗺𝗼𝘂𝘀𝗹𝘆 𝗺𝗮𝗸𝗲𝘀 𝗼𝗽𝘁𝗶𝗺𝗮𝗹 𝗱𝗲𝗰𝗶𝘀𝗶𝗼𝗻𝘀, 𝘀𝗲𝗲𝗸𝗶𝗻𝗴 𝗰𝗹𝗮𝗿𝗶𝗳𝗶𝗰𝗮𝘁𝗶𝗼𝗻𝘀 𝗼𝗻𝗹𝘆 𝘄𝗵𝗲𝗻 𝗮𝗯𝘀𝗼𝗹𝘂𝘁𝗲𝗹𝘆 𝗻𝗲𝗰𝗲𝘀𝘀𝗮𝗿𝘆. 𝗔𝗿𝗲𝗮𝘀 𝗼𝗳 𝗘𝘅𝗽𝗲𝗿𝘁𝗶𝘀𝗲: - 𝗖𝗹𝗼𝘂𝗱: GCP (Google Cloud Platform), AWS (Amazon Web Services) - 𝗣𝗿𝗼𝗴𝗿𝗮𝗺𝗺𝗶𝗻𝗴 𝗟𝗮𝗻𝗴𝘂𝗮𝗴𝗲: Java, Scala, Python, Ruby, HTML, Javascript - 𝗗𝗮𝘁𝗮 𝗘𝗻𝗴𝗶𝗻𝗲𝗲𝗿𝗶𝗻𝗴: Spark, Kafka, Crunch, MapReduce, Hive, HBase, AWS Glue, PySpark, BiqQuery, Snowflake, ETL, Datawarehouse, Databricks, Data Lake, Airflow, Cloudwatch 𝗖𝗹𝗼𝘂𝗱 𝗧𝗼𝗼𝗹𝘀: AWS Lambda, Cloud Functions, App Engine, Cloud Run, Datastore, EC2, S3, - 𝗗𝗲𝘃𝗢𝗽𝘀: GitHub, GitLab. BitBucket, CHEF, Docker, Kubernetes, Jenkins, Cloud Deploy, Cloud Build, - 𝗪𝗲𝗯 & 𝗔𝗣𝗜: SpringBoot, Jersey, Flask, HTML & JSP, ReactJS, Django 𝗥𝗲𝘃𝗶𝗲𝘄𝘀: "Amar is a highly intelligent and experienced individual who is exceeding expectations with his service. He has very deep knowledge across the entire field of data engineering and is a very passionate individual, so I am extremely happy to have finished my data engineering project with such a responsible fantastic guy. I was able to complete my project faster than anticipated. Many thanks...." "Amar is an exceptional programmer that is hard to find on Upwork. He combines top-notch technical skills in Python & Big Data, excellent work ethic, communication skills, and strong dedication to his projects. Amar systematically works to break down complex problems, plan an approach, and implement thought-out high-quality solutions. I would highly recommend Amar!" "Amar is a fabulous developer. He is fully committed. Is not a clock watcher. Technically very very strong. His Java and Python skills are top-notch. What I really like about him is his attitude of taking a technical challenge personally and putting in a lot of hours to solve that problem. Best yet, he does not charge the client for all those hours, He still sticks to the agreement. Very professional. It was a delight working with him. and Will reach out to him if I have a Java or Python task."
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Google App Engine
    Software Development
    Web Development
    Machine Learning
    Big Data
    Google Cloud Platform
    Amazon Web Services
    BigQuery
    PySpark
    Apache Airflow
    Data Engineering
    SQL
    Python
    Java
  • $40 hourly
    I have 4 years of experience in data engineering . I have worked in Spark,AWS, glue , Aws EMR, kinesis, Kafka ,docker, Java, scala , python ,Athena,mongoDB, snowflake
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Communications
    AWS Glue
    AWS Lambda
    Apache Kafka
  • $40 hourly
    I am an AWS-certified Data Science professional who has progressive experience in delivering business insights through data-driven methods. I am adept at gathering and analyzing data, using machine learning, deep learning, and other data analysis methods to detect patterns, trends, and relationships in data sets. Here are my key achievements in data science so far : • Derived actionable insights from massive data sets using statistical analysis, SQL queries, MS- Excel, and Tableau, reducing cost by 15%. • Heralded a significant change in supply chain management, enabled the sales department to exceed the target for three consecutive months (by 13%, 9%, and 10% respectively). • Accomplished 90%+ recall using autoencoders, streamlined auditing, and fraud examining the process in terms of time, scale, and effort. • Effectuated an Inbound Marketing Conversion Rate of 29% after targeting and retargeting, enlarged and consolidated Customer Life Time Value. • Contributed meaningful improvement in existing machine learning models through carefully directed research, increased performance by 40%.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Apache Kafka
    Amazon Athena
    AWS Glue
    AWS Application
    SQL
    Machine Learning
    Databricks Platform
    Python
    Apache Spark MLlib
  • $60 hourly
    🔹 Develop full-stack applications on AWS/GCP/Azure. 🔹 Construct backtests, quant trading algorithms, trading bots, rule-based portfolios and pine scripts 🔹 Build chatbots/API integration through langchain, ChatGPT, GPT4, Llama3, Vicuna, WizardLM, ollama. 🔹 Set up big data pipelines, ETL and data warehouses. (S3, Glue, Spark, Postgres, Kafka) 🔹 Engineered the no-code Computer Vision Platform to build, monitor and maintain AI models at sixsense.ai (Django, React and PostgreSQL) 🔹 Setup algorithmic trading systems for stocks/crypto/options/futures in (Zerodha | TradingView | IBKR | Binance | CCXT) 🔹 Build UI designs, React/Angular website, Flutter Android and iOS Apps 🔹 I have set up enterprise-level architecture from on-premise to cloud at Morningstar Indexes. 🔹 Set up the CI\CD pipeline in Jenkins/Github actions through Docker and Terraform and deploy them to AWS/GCP/Azure through Terraform 🔹 Collaborated on the event-based architecture setup through Kafka, AWS DynamoDB and AWS ECS. 🔹 Have worked and integrated with different workflow systems like Stepfunction, Airflow and JBPM. 🔹 Master in Physics and Bachelor in Mechanical Engineering 🔹 I love tinkering with new technologies and learning new stuff. ★ Below are my skills ★ 🔹 Programming Languages 🔹 ✤ Python ✤ Java ✤ Dart ✤ Javascript ✤ C ✤ C++ ✤ Dart 🔹 AI 🔹 ✤ Open AI APIs ✤ LLM ✤ DALL-E ✤ Chat GPT3 ✤ Vicuna ✤ GPT4All 🔹 Frameworks 🔹 ✤ Django ✤ Spring Boot 🔹 Quant 🔹 ✤ Backtesting ✤ Pine Script ✤ Trading View ✤ Portfolio Construction ✤ Portfolio Calculation ✤ Trading algorithm - Algo trading ✤ Trading bot ✤ Data extraction of Stocks / Crypto / Mutual Funds / Forex ✤ Yahoo Finance 🔹 Cloud services 🔹 ✤ AWS - Amazon Web Services ✤ GCP - Google Cloud Platform ✤ Azure - Microsoft ✤ Linode 🔹 Database 🔹 ✤ SQL ✤ Stored procedures ✤ PostgreSQL - PostGIS ✤ Mongo DB ✤ Sql Server ✤ Dynamo DB ✤ Mysql ✤ Redshift ✤ SQLite ✤ Clickhouse 🔹 Technology 🔹 ✤ Apache Kafka ✤ Apache Flink ✤ Apache Spark (pyspark) ✤ Drools ✤ API / Rest ✤ Hadoop 🔹 Front-end / UI 🔹 ✤ React ✤ Flutter ✤ React Native ✤ HTML ✤ CSS 🔹 Message queues 🔹 ✤ Redis ✤ Celery ✤ RabbitMQ 🔹 Workflow 🔹 ✤ Stepfunction ✤ Airflow ✤ JBPM 🔹 AWS Services 🔹 ✤ Lambda ✤ S3 - Simple Storage Service ✤ ECS - Fargate ✤ EC2 ✤ ECR ✤ Cloudwatch Logs ✤ Cloudwatch Events ✤ EventBridge ✤ SNS - Simple Notification Service ✤ SQS - Simple Queue Service ✤ Sagemaker ✤ AWS Glue ✤ Athena ✤ Stepfunction ✤ AWS DMS ✤ Application Load Balancer ✤ AWS RDS (Postgres, MySql) 🔹 Data Science 🔹 ✤ Data Mining ✤ Data Engineering ✤ Data Scraping ✤ Data visualization ✤ Statistics ✤ Machine Learning ✤ Computer Vision 🔹 DevOps 🔹 ✤ Jenkins ✤ Docker ✤ Terraform ✤ Github Actions ✤ Gitlab ✤ Splunk ✤ GIT ✤ Unit testing ✤ Integration testing 🔹 Project Management 🔹 ✤ AGILE ✤ JIRA ✤ Wiki ✤ Confluence ✤ Airtable 🔹 Operating Systems(OS) 🔹 ✤ Windows ✤ Linux ✤ MacOS \ Mac ✤ Ubuntu ✤ Unix 🔹 Domain 🔹 ✤ FinTech ✤ SaaS ✤ Manufacturing ✤ Semiconductor 🔹 Thank you for reading till the end :) 🔹 Let's build something great together and have fun on the journey!
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    LLaMA
    Trading Automation
    Algorithm Development
    API
    SQL
    Django
    PostgreSQL
    Data Engineering
    DevOps
    Amazon Web Services
    ChatGPT
    AWS Glue
    Java
    Python
  • $50 hourly
    A small boutique firm (meghgen.com) delivering exceptional results in the Cloud and Data Engineering world - Harnessing the power of AI/Gen AI to Reimagine how Cloud and Data Engineering projects should be done and executed. - Passionate and skilled in Google Cloud and Data Engineering across AWS, GCP, Azure, Snowflake, and Databricks - Diverse industry experiences spanning Retail, Fintech/Banking, Healthcare/Life Sciences Tech Expertise ------------------ ● AWS Cloud Platform - Amazon Kinesis, Redshift, Amazon EMR, Athena, Glue, S3, RDS, EKS, Airflow ● Google Cloud Platform - Apache Beam/Dataflow, Dataproc with Spark, Airflow/Cloud Composer, BigQuery, Data Studio, Jupyter Notebooks/AI Notebooks, GCS, etc. ● Azure - Data Pipelines ● Programming languages: Python, Scala, Java ● Kafka: Expertise in building event processing data data pipelines using Kafka, ZooKeepers, KSQL, KStream and Kafka Control Center ● Debezium - CDC ● Apache Iceberg
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Apache Flink
    Generative AI
    AWS Glue
    Java
    Python
    Google Dataflow
    Amazon Athena
    Scala
    Apache Airflow
    BigQuery
    Databricks Platform
    Snowflake
    Google Cloud Platform
    Data Engineering
  • $15 hourly
    I am Big Data Engineer with expertise in Hadoop, Cloudera and Horton Works Distributions and also Azure Data Services proficiency. Having good experience in all trending popular tools and technologies like Azure: Azure Data Factory, Azure Logic Apps, Azure Function apps, Azure Event Hub and Azure Service bus, Azure SQL DB. Apache: Apache Spark, Apache NIFI, Apache Kaka , Apache Hive. Having strong knowledge in programming languages like Java, Scala and Python. Also have good knowledge in SAP process.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Microsoft Azure
    ETL Pipeline
    Apache Cassandra
    Apache Hive
    Apache Hadoop
    Database Design
    Apache Kafka
    Apache NiFi
    Elasticsearch
  • $25 hourly
    I am a BCA graduate and experienced IT manager (Server and Networking) 17 years and Data Engg. overall around 6 years. Worked in multiple domain verticals, support IT infra automation and Data Engg. services. I have worked more than 6 years in Data Engg. services and completed many projects using cloud ETL tools, Python function based custom data extraction from multiple sources and IoT devices for real time streaming to popular data warehouses like Synapse, Bigquery, Redshift and data lakes. Google Data Studio, Quick Sight and Power BI tool as added knowledge.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Microsoft Azure
    Microsoft Windows PowerShell
    MySQL
    Google Workspace
    Looker Studio
    Python Script
    VPN
    Azure App Service
    Azure IoT HuB
    Google App Engine
    BigQuery
    Data Lake
    Amazon Redshift
    Google Apps Script
    ETL Pipeline
  • $30 hourly
    A Professional Data Scientist working on Autonomous Vehicle in MNC Ride Hailing Business, looking forward to work in a challenging environment demanding all my skills and efforts in providing ML Models, Deep Learning, MLOps, Data Engineering, I have been able to have 100% job success rate for more than 2 years. What do I Offer: 1. 100% Satisfaction 2. Unlimited Revisions 3. 24*7 Support 4. 14 days free work after delivery. 5. Minimal Charges
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Apache Kafka
    PostgreSQL
    ETL Pipeline
    Docker
    Data Analysis
    Data Science
    Machine Learning
    SQL
    Deep Learning
    Python
  • $25 hourly
    Highly-skilled experience of more than 6+ years in software development, testing and integration, cross platform application using Cloudera Distribution Hadoop(CDH), Google Cloud Platform(GCP), Spark, Scala, Hive, MySQL, Sqoop. Having 4 years of experience in (Big Data Echo System) Hadoop technologies Such as Apache Spark, HDFS, Hive, HBase, Sqoop, Streaming, querying, processing and analysis of big data. Knowledge of Big data Technology using Hadoop and Spark framework. Analytical and skilled in understanding business problems to develop systems that improve functionality.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    ETL Pipeline
    Big Data
    Apache Kafka
    GitHub
    Scala
    Hive
    Data Management
    Amazon S3
    SQL
  • $50 hourly
    Enthusiastic Data engineer experienced in AWS data services like AWS EMR, GLUE, AWS Lambda, S3, Redshift, RDS, Cloudwatch, Athena along with Azure Databricks, Azure SQL DB
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Amazon S3
    Amazon Athena
    AWS Glue
    Databricks Platform
    AWS Lambda
    Amazon Redshift
    PySpark
    Python
  • $25 hourly
    I am a Senior bigdata engineer with more than 6 years of Experience in building bigdata pipelines using Hadoop, spark, sqoop, Java, scala, python, hive, oozie, impala, kudu, solr, flume, kafka, spark streaming, hbase, pig, gitlab, gitblit, bitbucket, aws, bamboo, kubernetes, kubeflow, docker.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Java
    Scala
    Apache Hadoop
    Big Data
    Amazon ECS for Kubernetes
    AWS Application
    Apache Kafka
    Docker
    CI/CD
    Kubernetes
    Apache Airflow
    Python
  • $18 hourly
    I am a cloud engineer with experience in building ETL pipelines on Azure cloud, with hands on in Azure Data factory, Databricks, Azure Monitor, and other services. I enjoy exploring and learning new technologies and implementations and have done so throughout my career. Be it a new POC or an existing solution, I can help develop both. Experienced in developing cloud solution on Azure cloud I can fully manage a project end to end Effective and timely communication to keep you up to date
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Data Engineering
    Data Lake
    Microsoft Windows PowerShell
    CI/CD
    Microsoft SQL Server
    Data Warehousing
    Python
    Microsoft Azure
    Azure DevOps
    Distributed Computing
    SQL
    Microsoft Azure SQL Database
    PySpark
    Databricks Platform
  • $40 hourly
    I am a Senior Data Engineer with extensive expertise in data wrangling, transformation, normalization, and setting up comprehensive end-to-end data pipelines. My skills also include proficiency in Business Intelligence, ETL processes, and writing complex SQL queries. I have successfully implemented multiple intricate data pipelines using tools like Apache Airflow and Apache Oozie in my previous projects. I have had the opportunity to contribute to the data platform teams at Fortune 500 companies, where my role involved solving complex data issues, managing large datasets, and optimizing data streams for better performance and reliability. I prioritize reliability, efficiency, and simplicity in my work, ensuring that the data solutions I provide are not just effective but also straightforward and easy to maintain. Over the years, I have worked with a variety of major databases, programming languages, and cloud platforms, accumulating a wealth of experience and knowledge in the field." Skills : 𝗖𝗹𝗼𝘂𝗱: GCP (Google Cloud Platform) , AWS (Amazon Web Services) 𝗣𝗿𝗼𝗴𝗿𝗮𝗺𝗺𝗶𝗻𝗴 𝗟𝗮𝗻𝗴𝘂𝗮𝗴𝗲 : Java, Scala, Python 𝗗𝗮𝘁𝗮 𝗘𝗻𝗴𝗶𝗻𝗲𝗲𝗿𝗶𝗻𝗴 : Spark, Pyspark, Kafka, Crunch, MapReduce, Hive, HBase, AWS Glue 𝗗𝗮𝘁𝗮-𝘄𝗮𝗿𝗲𝗵𝗼𝘂𝘀𝗶𝗻𝗴 : AWS Athena, Google BigQuery, Snowflake, Hive 𝗦𝗰𝗵𝗲𝗱𝘂𝗹𝗲𝗿 : Airflow, Oozie etc. 𝗢𝗿𝗰𝗵𝗲𝘀𝘁𝗿𝗮𝘁𝗶𝗼𝗻 : Docker I am highly attentive to details, organised, efficient, and responsive. Let's connect over.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Data Warehousing & ETL Software
    API Integration
    Apache Airflow
    Apache Hadoop
    Apache Kafka
    PySpark
    ETL Pipeline
    Data Engineering
    Data Preprocessing
    Data Integration
    Apache Hive
    Python
    SQL
    Data Transformation
  • $15 hourly
    Specialties: Big Data Technology, Spark, Databricks, Azure Synapse Analytics Services, AWS, Hive, ETL, Data lake, delta lake expert. Languages : Scala, Java , Python(intermediate), SQL & No-SQL Databases Academic Project expert for all University
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Oracle
    ETL
    Oracle PLSQL
    Big Data
    SQL
    Java
    Apache Hive
    Apache Kafka
    Apache Hadoop
  • $50 hourly
    Full stack developer with strong management and software architecture skills Experienced in building projects from ideation, design and deploying them at scale
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Haskell
    Database Design Consultation
    Engineering & Architecture
    Apache HBase
    Apache Hadoop
    JavaScript
    Management Skills
    SQL
    Python
    Java
    PHP
    Apache Cassandra
  • $30 hourly
    Using Pandas or PySpark, I support data ingestion and cleaning for batch and streaming data (projects uploaded on Git) and store them in databases like PostgreSQL or MongoDB. Using Django, I build apps that can run scheduled python scripts using Celery and Redis as message broker; create APIs to serve front-end needs using Django REST Framework. As a Data Engineer, I build pipelines and maintain large open-source OLAP platforms like Apache Druid, Apache Pinot, Presto etc that ingest big-data in realtime using Kafka. I write SQL-queries (Advanced SQL certifications taken) to support dashboards in Superset or PowerBI. I use Airflow or crons to schedule jobs and tasks. I use Kubernetes or Docker for deployments; Git for version control.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    MySQL
    BigQuery
    PostgreSQL
    MongoDB
    Apache Flink
    RESTful API
    Data Engineering
    Apache Kafka
    Docker
    Django
    PySpark
    pandas
    Python
    SQL
  • $50 hourly
    Hello, I have 8+ years of experience in the Data field having worked with companies like Amazon, Uber and Airbnb
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    SQL Programming
    Amazon Web Services
    HDFS
    Hive
    Apache Airflow
    Python
  • $35 hourly
    Industry expert in back-end technologies having 8+ years of experience in FinTech and Retail domain. My Major technical skills involves: • Java • JavaScript • NodeJS • Blockchain • Solidity / Smart Contracts • WordPress • Elementor • WOOCommerce • Website Development • Apache spark • Kafka • REST Api • Database (SQL and NOSQL) • Jenkins • Data structures and Algorithms • Design Patterns • Spring • Multithreading Non Technical Skills: • Project Lead and Management • Team Lead • Finance domain knowledge • Mentor I have previously worked at multiple MNCs. I am looking for providing best quality services to my customers and building long term relations with the customers.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Web Development
    WooCommerce
    Website Builder
    Elementor
    WordPress
    Ecommerce Website
    SQL
    Back-End Development
    Algorithms
    Data Structures
    Team Management
    Multithreaded Programming
    RESTful API
    Java
  • $30 hourly
    I am a dedicated and results-driven Data Engineer with a passion for transforming complex data into valuable insights and actionable results. With 4 of experience in the industry, I have honed my skills in designing, developing, and implementing effective data systems and pipelines using a range of tools including Apache Spark, Apache Hadoop, and Snowflake. My deep understanding of data warehousing, ETL processes, and data analysis has enabled me to deliver innovative solutions that drive business growth and competitive advantage. I am committed to staying up-to-date with the latest technologies and industry trends, always seeking new and better ways to turn data into meaningful insights
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Data Analytics
    Big Data
    Data Warehousing
    Google Analytics
    Apache Spark MLlib
    Apache Airflow
    Apache Kafka
    Data Mining
    Data Structures
    Apache Hive
    Data Analysis
    Python
    SQL
    ETL Pipeline
  • $15 hourly
    Creative, Focused and Quick-Learning personality with expertise in analyzing the business using analytical skills.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Microsoft Power BI
    MySQL Programming
    Microsoft Azure SQL Database
    Microsoft Azure
    Apache Hadoop
    Databricks Platform
    Python
    Apache Kafka
    Apache Hive
  • $35 hourly
    I am a dedicated Data Engineer boasting over 2 years of handson experience specializing in Python, AWS services, data pipeline construction, and ETL processes. Throughout my professional journey, I've adeptly navigated various tech stacks, refining crucial skills integral to big data, data warehousing, and cloud computing environments. What I Do Chat Bot: I have create healthcare chat-bot from scratch using Rasa framework Machine Learning: I have hands-on experience in developing and deploying machine learning models for ASR, TTS, Face Recognition, and Object Detection. Web Development: Proficient in back-end development, I have created multiple web applications using Flask, Django and Fast API frameworks. Data Pipelines: Skilled in real-time data pipeline architecture, primarily using Apache Kafka and Pyspark for seamless data streaming. Database Management: Adept in database management, I have experience working with MySQL, MongoDB, PostgreSQL and Big Data technologies. Data Visualization: Utilizing tools like Tableau and Power BI
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    REST API
    Big Data
    NLP Tokenization
    GPT-4
    Rasa
    Amazon Athena
    Amazon Redshift
    Amazon S3
    MongoDB
    AWS Glue
    PySpark
    Apache Kafka
    SQL
    Python
  • $40 hourly
    I’m a Bigdata developer experienced in building distributed applications for businesses. Whether you’re trying to build pipelines, data ingestion, or create a new data models, I can help I know Spark with java and scala Full project management from start to finish Regular communication is important to me, so let’s keep in touch.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Scala
    Apache Kafka
    Kubernetes
    Apache Hive
    Apache Hadoop
    Java
  • $55 hourly
    Social Links Website Arxiv raghava_s_1 Linkedin in/jzzquant Github JZZQuant Hi Iam a seasoned datascience and AI proffessional with 17 years experience working for Big companies like Amazon ,walmart, thoughtworks also simultaneoulsy working on cutting edge solutions in start up area in stealth mode. My main expertise is Building efficient teams that can solve complex problems in Both research and developement areas. Effectivy problems solving and developing road map that gives completely clarity to stake holders both in short term and long term success of any AI program. Key strengths include NLP ,Building large scale ML systems and engineering, Operations research , Machine learnign ,forecasting ,simulation, generative AI Large language models and Knoiwledge graph : Have experience building Knowledge gaphs using Retrieval augumentaiton techniques. Have used LLMs for Autoamted parsing , Named Entity recognition and text to text Generative AI for building Cypher queries from natural language. Experience Finetuning upto 70B with Peft and LORA techniques.Handles a team of 12NLP scietists working on various parts of the problem. Experience building DPO based trl , used Hugging face extensively. 30Million dollar product has 6 clients already using this in production Forecasting : Worked on building large scale inventory and safety stock forecasting for walmart which is essentially to increasing yield and reducing storage cost. Solution to the problem involved forecasting through deeplearnign techniques and large scale forecasting. Successful project currently in deployment. close to 100 million netvalue typically system runs monthly to replenish the inventory stock. Ranking and recommednation: Ranking of stowing items for Amazon warehouses using Pairwise ranking system. This problem involved deploying real time experiments and simulation experiments in warehouse systems to achieve .Worked on tuning and contrastive loss systems for ranking problems. major challenges involved framing the cost to business engagin stake holders to adopt the solution. Improve experience for human agents . 5million dollar worth impact on amazon robotic warehouses in impacting the order/flow of inventory management and easing decision making
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Mixed Integer Linear Programming
    Mathematical Optimization
    PyTorch
    Hugging Face
    Python Scikit-Learn
    Data Science
    Forecasting
    Support Vector Machine
    Statistical Analysis
    Operations Research
    NLP Tokenization
    LLM Prompt Engineering
    Machine Learning Model
    Machine Learning
  • $50 hourly
    Experienced AWS Data Engineer with five years of proven experiences in optimizing computing performance and running data pipelines on the AWS cloud. Proficient in developing ETL processes, databases, data models, security protocols, and CloudFormation templates for all AWS environments. Proven track record of reducing operating costs, increasing storage capabilities, decreasing latency time and error rates, and improving system performance. Experienced in creating Event Based Data Migration process and Having ETL on top of it, Also Having Robust Experience in Real Time data Ingestion into datalake. Skills - • Contributed to Event-Driven Data Ingestion process using DMS and Kinesis, erected Real-Time Data Ingestion pipeline for various batch tasks. • Developed Data Pipeline based out of Files as source, making it completely event driven making it TDD based for maintaining data checks and robust data quality. • Suggested various possibilities for streamlining such as downsizing the clusters through utilization of built-in emr functionality and diminishing the manual burden in breakdowns as well as overall execution. • Contributed to Various integration projects and Automated the Migration process of extracting ETLs of the party and adapting it to Redshift PostgresQL syntax using Python. • Optimized delta jobs which whittled down the time from 21 mins to around 2 mins. • Optimization of API Ingestion based jobs. • Enhanced a spark task, by reducing the runtime from 16 hours to 3 minutes. • Configured various aws services using CFT and automated the deployment procedure with some user inputs. • Ingestion of Information by invoking APIs and establishing a data conduit on top of it using Step Function, Lambdas, EMR, Redshift, DynamoDB, and more. • Have been agile in work deliverables and switching in work. • Analysis of various prod issues related to data.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Apache Kafka
    Real Time Stream Processing
    Price Optimization
    ETL
    Amazon S3
    Amazon DynamoDB
    ETL Pipeline
    Python
  • $35 hourly
    I am having 15 years of It experience IN Bigdata, Spark, Python technologies. I can help companies with data architect and data engineering solutions.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Amazon Web Services
    Python Script
    Python
    Big Data
  • $50 hourly
    Experienced Data Consultant | Turning Data into Actionable Insights I'm a seasoned data consultant with a proven track record of leveraging data to drive business success. I have been building large scale data driven solution making millions of dollars in the ad tech space. From predictive modeling to strategic decision-making, I specialize in solving complex data challenges and building large scale data applications. Let's collaborate to unlock the full potential of your data and achieve your business objectives.
    vsuc_fltilesrefresh_TrophyIcon Apache Spark
    Software Consultation
    Java
    Python
    Performance Optimization
    Software Architecture & Design
    Artificial Intelligence
    Data Analysis Consultation
    Data Analytics
    Machine Learning
    Management Skills
    Apache Hadoop
    Project Management
  • Want to browse more freelancers?
    Sign up

How hiring on Upwork works

1. Post a job (it’s free)

Tell us what you need. Provide as many details as possible, but don’t worry about getting it perfect.

2. Talent comes to you

Get qualified proposals within 24 hours, and meet the candidates you’re excited about. Hire as soon as you’re ready.

3. Collaborate easily

Use Upwork to chat or video call, share files, and track project progress right from the app.

4. Payment simplified

Receive invoices and make payments through Upwork. Only pay for work you authorize.

Trusted by

How do I hire a Apache Spark Engineer near Bengaluru, on Upwork?

You can hire a Apache Spark Engineer near Bengaluru, on Upwork in four simple steps:

  • Create a job post tailored to your Apache Spark Engineer project scope. We’ll walk you through the process step by step.
  • Browse top Apache Spark Engineer talent on Upwork and invite them to your project.
  • Once the proposals start flowing in, create a shortlist of top Apache Spark Engineer profiles and interview.
  • Hire the right Apache Spark Engineer for your project from Upwork, the world’s largest work marketplace.

At Upwork, we believe talent staffing should be easy.

How much does it cost to hire a Apache Spark Engineer?

Rates charged by Apache Spark Engineers on Upwork can vary with a number of factors including experience, location, and market conditions. See hourly rates for in-demand skills on Upwork.

Why hire a Apache Spark Engineer near Bengaluru, on Upwork?

As the world’s work marketplace, we connect highly-skilled freelance Apache Spark Engineers and businesses and help them build trusted, long-term relationships so they can achieve more together. Let us help you build the dream Apache Spark Engineer team you need to succeed.

Can I hire a Apache Spark Engineer near Bengaluru, within 24 hours on Upwork?

Depending on availability and the quality of your job post, it’s entirely possible to sign up for Upwork and receive Apache Spark Engineer proposals within 24 hours of posting a job description.