Hire the best Apache Spark Engineers in Kathmandu, NP
Check out Apache Spark Engineers in Kathmandu, NP with the skills you need for your next job.
- $45 hourly
- 4.9/5
- (10 jobs)
Reliable data engineer with 10 years of proven industry experience in data lake development, data analytics, real-time streaming, and back-end application development. My work is used by millions of people in the legal and entertainment industries. I have built exceptionally stable solutions for high-traffic, high-visibility projects, and understand what it takes to ensure products are robust and dependable. I also have expertise in the Apache Spark ecosystem, Elastic Search, ETL, AWS Glue, DMS, Athena, EMR, Data Lake, AWS Big Data, Apache Kafka, Java, and NoSQL. Specific Experience 1. Databricks : 5+ years of experience 2. Unity Catalog: 2+ years of experience 3. Apache Spark: 8+ years of experience 4. ETL: 8+ years of experience 5. SQL: 9+ years of experience 6. AWS: 8+ years of experience 7. Azure and GCP: 5+ years of experience I am a data professional, worked with many companies, and delivered some of the enormous data engineering and data science projects in the past. My focus is always on scalable, sustainable, and robust software building. As a data scientist, I will use data modeling, programming, analysis, visualization, and writing skills to help people have the insight to develop products, customers, and impact. As a data scientist, I care deeply about the data from beginning to end—I am actively involved in all aspects of data analysis, from data modeling tasks to writing reports and making visualizations. Python/Scala Programming, Linux Admin, Data Wrangling, Data Cleansing & Data Extraction services utilizing Python 3 or Python 2 Programming or Scala/Spark on Linux or Windows. I slice, dice, extract, transform, sort, calculate, cleanse, collect, organize, migrate, and otherwise handle data management for clients. Services Provided: - Big data processing using Spark Scala - Building large Scale ETL - Could Management - Distributed platform development - Machine learning - Python Programming - Algorithm Development - AWS glue - Pyspark - Data Conversion (Excel to CSV, PDF to Excel, CSV to Excel, Audio) - Data Mining - Data extraction - ETL Data Transformation - Data Cleansing - Linux Server Administration - Website & Data Migrations - DevOps (AWS,AZURE)Apache Spark
Amazon EC2Data Warehousing & ETL SoftwarePySparkETL PipelineRedisAWS GlueDatabricks MLflowDatabricks PlatformPython - $50 hourly
- 5.0/5
- (24 jobs)
With over 5+ years of experience, I've expertise in the following areas:- - Architecting Distributed Database clusters & Data pipelines for Big Data Analytics and Data Warehousing using tech stacks which include but are not limited to Citus Data, Timescale, Spark, Kafka, Kinesis, Trino/PrestoSQL, Airflow, DBT, Athena, Glue, Hadoop, Hive,Iceberg, Minio, wasabi,S3 Data lake . - Database Administration, Setup, Maintenance, Data migration, Backup&Recovery, Monitoring, Replication, Performance Tuning, and Query Optimization of Postgres, MySQL, Oracle,Starrocks, Ignite, Dorris, Clickhouse, and MongoDB Databases. - Python, Bash, and SQL scripting for database management and automation. - Architecting your next enterprise-level software solution - Web application and API development with Python/Django/Flask | Nodejs/Expressjs - Configure and maintain cloud services like AWS, GCP, Azure, Digital Ocean - Linux Server administration for setup and maintenance of services on cloud and on-premise servers. - Creating scripts to automate tasks, web scraping, and so on.Apache Spark
DjangoDatabase ArchitectureKubernetesLinux System AdministrationDevOpsBig DataDatabase AdministrationPythonMySQLETLPostgreSQLSQLDatabase OptimizationDatabase DesignData Migration - $40 hourly
- 5.0/5
- (1 job)
Experienced Big Data Engineer / Development Manager seasoned to develop BI infrastructure. With proficiency in Python and Scala, I have developed ETL pipelines to automate ingestion of data for data analysis, configured big data technologies using Apache Spark on kubernetes and EMR, Trino/ presto as well as fine tuned processes for performance at scale. Currently, my major responsibilities include, building data infrastructure, developing data pipelines, establishing coding conventions, defining processes and mentoring sub ordinates to the follow the same.Apache Spark
ReactBusiness IntelligenceData Warehousing & ETL SoftwareJasperReportsPHPAmazon RedshiftPostgreSQLData LakeMysteryTalend Open StudioScalaPythonETL Pipeline - $35 hourly
- 5.0/5
- (1 job)
As an experienced Data Engineer specializing in Azure, I am dedicated to helping you design and implement scalable and efficient data solutions tailored to your specific needs. My expertise spans across ETL pipelines, real-time data processing, Spark-based transformations, and comprehensive data architectures, making me the ideal partner for your data challenges. I can provide expert assistance in the following areas: Azure Data Expertise: I excel in developing and optimizing data pipelines using Azure Data Factory, Fabric Data Factory, Spark, and Azure Synapse Analytics. Whether you need incremental data loads, real-time Change Data Capture (CDC), or batch processing, I can deliver solutions that reduce latency and enhance your data processing capabilities. Data Integration & ETL: I can integrate data from various sources, including on-premise SQL Server and Azure SQL Server, into cloud-based environments like OneLake and Azure Data Lake Storage (ADLS). My expertise in ETL processes, along with leveraging Spark and notebooks for complex data transformations, ensures that your business logic requirements are met with precision. Data Modelling & Architecture: I can implement a complete medallion architecture within Azure or Microsoft Fabric, creating data models and optimizing resource usage to ensure efficient data management across your domains and environments. Real-Time and Full Load Pipelines: I specialize in the design and implementation of real-time and full-load data pipelines in Azure or Fabric, dramatically improving your data refresh rates and ensuring timely access to critical information. Spark & Notebooks for Transformations: Utilizing Spark for data transformations within notebooks, I ensure efficient processing and optimization of large datasets, enabling you to derive meaningful insights quickly and accurately. With a deep understanding of SQL, Azure, Spark, and hands-on experience in real-time data processing, Change Data Capture, and pipeline optimization, I am committed to delivering high-quality, tailored data solutions that meet your unique challenges and drive your success.Apache Spark
Web Scraping PluginWeb Scraping FrameworkData MiningMicrosoft Azure SQL DatabaseAzure DevOpsMicrosoft AzureETL PipelineReal Time Stream ProcessingData EngineeringPythonData WarehousingData Warehousing & ETL SoftwareSnowflakeETLData IngestionMicrosoft ExcelData AnalysisSQL - $25 hourly
- 5.0/5
- (10 jobs)
Data Engineer specialized in Python, Big Data and AI. Data Warehousing, Datalake, ETL, Automation, RAG Skills and Expertise: - Programming Languages: Python, Scala, SQL. - Frameworks: Flask, Spark, OpenAI, Langchain. - Database: MySQL, PostgreSQL, AWS Redshift, AWS RDS. - Cloud: AWS - Scripting and REST APIs. - ETL and ML. I will be very glad to work with you. Lets collaborate.Apache Spark
APIAWS GlueAmazon AthenaDjangoAmazon RedshiftData ScrapingAmazon S3SQLFlaskPythonComputer VisionDeep Learning - $15 hourly
- 5.0/5
- (1 job)
Looking to transform your raw data into meaningful insights? I’m here to help! With over 2 years of experience as a Data Engineer, I specialize in building scalable data pipelines, architecting cross-cloud infrastructures, and delivering real-time analytics that empower businesses to make smarter decisions. My expertise lies in integrating AWS and Azure ecosystems, leveraging tools like Databricks, PySpark, and Delta Lake to ensure data flows seamlessly from source to insights. Whether it's automating ETL processes, setting up incremental data loading, or building Power BI dashboards, I make complex data projects simple and efficient. Here’s what I bring to the table: • Cross-cloud Data Architecture: Seamlessly connect AWS and Azure to ensure smooth data movement and storage. • End-to-End Data Pipelines: Automate and optimize your data workflows, from ingestion to transformation. • Real-time & Batch Processing: Build data systems that handle both real-time and historical data, giving you the flexibility you need. • Advanced Analytics: Create sales, user, and RFM analysis data marts to unlock deep insights. Power BI & Tableau Dashboards: Develop interactive, real-time reports that drive data-driven decisions. • Security & Compliance: Implement Role-Based Access Control (RBAC) to keep your data secure and compliant. • Consulting & Cost Optimization: Get expert advice on cloud infrastructure design and cost-effective data solutions. What sets me apart is my commitment to delivering high-impact results. I work closely with clients to understand their business needs, and I build tailored data solutions that are both scalable and cost-effective. No project is too complex—whether you’re looking for a robust cloud architecture or seamless data integration, I’ll help you bring your vision to life.Apache Spark
Azure Service FabricETLData WarehousingMicrosoft AzureTableauData AnalysisData LakeData IngestionDatabricks PlatformSQLPySparkPythonData Engineering - $20 hourly
- 5.0/5
- (1 job)
1. Experienced AI engineer working as an ML team lead in Computer vision and NLP projects and other data related tasks a. Data Extraction b. Clustering c. Dimensionality Reduction d. Classification/Regression 2. Doing research in Transformer Architecture in Computer Vision and NLP 3. Skills a. Machine Learning and Deep Learning techniques b. Computer Vision c. Natural Language Processing e. PyTorch f. SciPy stack 4. I build end-end running AI systems using Python ecosystem Feel free to reach out to me for anything related to building anything related Machine Learning and Deep Learning. If you are interested in developing content or performing independent research, I am ready for that too.Apache Spark
Play FrameworkReduxApache HadoopCore JavaReactPythonScalaMySQL - $10 hourly
- 0.0/5
- (2 jobs)
I am from Nepal .I am new to upwork you can test me I can deliver a quality result .I have a bachelors degree in computer engineering. I am fluent in Python,databases engines such as (Mysql,mssql,mariadb,sqlite),snowflake,dbt,kafka,airflow,redshift,aws.Apache Spark
Data Warehousing & ETL SoftwarePentahoMicrosoft Power BITableauApache KafkaSnowflakeDatabricks PlatformSQLiteMicrosoft SQL ServerMySQLLaravelPHPPythonC - $30 hourly
- 0.0/5
- (0 jobs)
Hello! I'm Kiran Bhandari, an enthusiastic and skilled Data Engineer with a Bachelor's degree in Information Technology. Over the past year, I have built a solid foundation in data engineering, specializing in AWS, ETL pipelines, and database management. My goal is to leverage data to drive informed decision-making and help businesses unlock actionable insights through seamless data processing and management. What I Do Best: Data Engineering Excellence: I excel at designing, developing, and optimizing robust ETL pipelines that streamline data flow, improve performance, and ensure data integrity. I work with large-scale data processing systems to ensure timely and reliable delivery. AWS Cloud Expertise: With in-depth experience in Amazon Web Services (AWS), I specialize in utilizing the cloud to implement scalable, efficient data processing, storage, and warehousing solutions. My skills in AWS Glue, S3, and Redshift enable seamless integration across platforms, ensuring high-performance workflows. Database Management & SQL: I am proficient in SQL and experienced in managing relational databases, creating data models, and ensuring optimal performance through efficient query design and indexing. I focus on data quality and performance in every stage of the pipeline. Python Scripting for Automation: I leverage Python to create custom scripts that automate data processing tasks, build ETL workflows, and enhance overall efficiency. Whether it's parsing large datasets or integrating with APIs, I have the tools and expertise to get the job done. CI/CD Pipeline & Automation: I have hands-on experience designing and implementing CI/CD pipelines for AWS Glue workflows, utilizing AWS CloudFormation for automated deployment and management. This ensures continuous integration and delivery of high-quality data engineering solutions. Data Quality & Monitoring: I implement data quality checks and monitoring to identify and resolve inconsistencies or errors in datasets, ensuring that the data is accurate and reliable for analysis. Data Visualization & Reporting: I specialize in creating impactful data visualizations and dashboards using Amazon QuickSight and Tableau. By delivering clear, actionable insights, I empower stakeholders to make data-driven decisions that lead to business growth. Continuous Learning & Knowledge Sharing: I am deeply passionate about staying up-to-date with the latest technology trends. I regularly share insights and provide training sessions for junior team members on best practices in ETL development, cloud technologies, and data engineering methodologies. Key Achievements: Designed and optimized ETL processes for large datasets, enabling seamless extraction, transformation, and loading of data into centralized data warehouses. Implemented delta load processes and data partitioning techniques, significantly improving data refresh efficiency and reducing processing time. Enhanced CI/CD pipelines for AWS Glue, automating deployment and improving data processing workflows. Successfully collaborated with data scientists, analysts, and cross-functional teams to meet evolving data processing requirements. Delivered intuitive, data-driven dashboards and reports, empowering stakeholders to make informed decisions. If you're looking for a dedicated, results-driven data engineer who can transform your data workflows, ensure the accuracy of your data, and help you leverage AWS cloud technologies for scalable solutions, I'm here to help. Let's collaborate and bring your data initiatives to life!Apache Spark
Apache Spark MLlibPython ScriptpandasETLPythonSQLPySparkData Engineering - $15 hourly
- 0.0/5
- (0 jobs)
Data Engineer with specializing in data pipeline design, ETL, and data modeling. Skilled in Spark, Python, Hadoop, and cloud platforms (AWS, Azure, GCP). Expertise in building data lakes/warehouses and visualizing data with Power BI and Tableau.Apache Spark
dbtMachine LearningData VisualizationETL PipelineData WarehousingData ModelingSnowflakeApache HadoopApache KafkaApache AirflowSQLPythonCloud ComputingPySpark - $5 hourly
- 0.0/5
- (0 jobs)
Reliable data engineer with proven industry experience in data lake development, data analytics, real-time streaming, and back-end application development. I also have expertise in the Apache Spark ecosystem, Elastic Search, ETL, AWS Glue, DMS, Athena, EMR, Data Lake, AWS Big Data, Apache Kafka, Java, and NoSQL. As a data engineer, I care deeply about the data from beginning to end—I am actively involved in all aspects of data engineering, from data modeling tasks to writing ETL and developing pipeline. Python/Scala Programming, Linux Admin, Data Wrangling, Data Cleansing & Data Extraction services utilizing Python 3 or Python 2 Programming or Scala/Spark on Linux or Windows. I slice, dice, extract, transform, sort, calculate, cleanse, collect, organize, migrate, and otherwise handle data management for clients. Services Provided: - Big data processing using Spark Scala - Building large Scale ETL - Could Management - Distributed platform development - Python Programming - AWS glue - Pyspark - Data Conversion (Excel to CSV, PDF to Excel, CSV to Excel, Audio) - Data Mining - Data extraction - ETL Data Transformation - Data Cleansing - Linux Server Administration - Website & Data Migrations - DevOps (AWS,AZURE)Apache Spark
Couchbase ServerAWS LambdaAmazon EC2AWS GlueData AnalyticsDatabaseBig DataAmazon Web ServicesMicrosoft Power BIDevOpsSQLPython Want to browse more freelancers?
Sign up
How hiring on Upwork works
1. Post a job
Tell us what you need. Provide as many details as possible, but don’t worry about getting it perfect.
2. Talent comes to you
Get qualified proposals within 24 hours, and meet the candidates you’re excited about. Hire as soon as you’re ready.
3. Collaborate easily
Use Upwork to chat or video call, share files, and track project progress right from the app.
4. Payment simplified
Receive invoices and make payments through Upwork. Only pay for work you authorize.
How do I hire a Apache Spark Engineer near Kathmandu, on Upwork?
You can hire a Apache Spark Engineer near Kathmandu, on Upwork in four simple steps:
- Create a job post tailored to your Apache Spark Engineer project scope. We’ll walk you through the process step by step.
- Browse top Apache Spark Engineer talent on Upwork and invite them to your project.
- Once the proposals start flowing in, create a shortlist of top Apache Spark Engineer profiles and interview.
- Hire the right Apache Spark Engineer for your project from Upwork, the world’s largest work marketplace.
At Upwork, we believe talent staffing should be easy.
How much does it cost to hire a Apache Spark Engineer?
Rates charged by Apache Spark Engineers on Upwork can vary with a number of factors including experience, location, and market conditions. See hourly rates for in-demand skills on Upwork.
Why hire a Apache Spark Engineer near Kathmandu, on Upwork?
As the world’s work marketplace, we connect highly-skilled freelance Apache Spark Engineers and businesses and help them build trusted, long-term relationships so they can achieve more together. Let us help you build the dream Apache Spark Engineer team you need to succeed.
Can I hire a Apache Spark Engineer near Kathmandu, within 24 hours on Upwork?
Depending on availability and the quality of your job post, it’s entirely possible to sign up for Upwork and receive Apache Spark Engineer proposals within 24 hours of posting a job description.