Hire the best Pyspark Developers in Ireland

Check out Pyspark Developers in Ireland with the skills you need for your next job.
  • $150 hourly
    Hi! My name is Alex and I am from Ireland. I have a passion for everything involved in coding, AWS and learning new technologies. * Bachelors Degree in Digital Technology & Design * Masters Degree in Software Engineering - Won multiple awards including top student overall in whole university (8000+ students) * 6x AWS Certified. * 10+ years experience * Always looking to learn and improve * My main skills revolve around 3 distinct fields ** Full Stack Development ** AWS Architecture & DevOps ** Data Engineering I am proficient in the following languages & technologies Everything AWS! Serverless Framework Serverless Application Model Terraform PySpark Java Javascript & ES6 TypeScript Python Vue React ReactNative MySQL MongoDB ElasticSearch PHP Laravel Node.js A Strong advocate for "serverless" design on AWS using AWS Lambda, API gateway & DynamoDB. I have real work experience with 70+ services on AWS.
    Featured Skill Pyspark
    Terraform
    Big Data
    AWS CloudFormation
    AWS AppSync
    React Native
    Solution Architecture
    AWS Lambda
    PySpark
    DevOps
    AWS Glue
    TypeScript
    AWS Amplify
    JavaScript
    Node.js
    Amazon Web Services
  • $19 hourly
    About Me Greetings! I'm Syed Ahmed Shah, a dedicated and results-driven Data Analyst with a passion for transforming raw data into actionable insights. With 10 years of experience in the field, I have honed my skills in data manipulation, analysis, and visualization to help businesses make informed decisions and drive growth. My Expertise Data Analysis: I excel at cleaning, organizing, and analyzing complex datasets to extract valuable information. Whether it's working with SQL, Python, or R, I have the technical proficiency to uncover meaningful patterns and trends. Statistical Modeling: I am well-versed in statistical techniques, including regression analysis, hypothesis testing, and clustering, allowing me to provide data-driven solutions for various business problems. Data Visualization: I have a keen eye for creating visually appealing and informative dashboards and reports using tools like Tableau, Power BI, and matplotlib. This ensures that data is not just understood but also easy to communicate to stakeholders. Machine Learning: I have experience in applying machine learning algorithms for predictive modeling and classification tasks, enhancing business processes and decision-making. Data Interpretation: My ability to translate complex data into clear, non-technical language enables effective communication between technical and non-technical teams, facilitating informed decision-making at all levels.
    Featured Skill Pyspark
    Apache Hadoop
    Microsoft Power BI
    Data Analytics
    Azure DevOps
    Apache Flink
    PySpark
    SQL Programming
    ETL Pipeline
    Artificial Intelligence
    Big Data
    Python Scikit-Learn
    TensorFlow
    Apache Spark
    Python
    Machine Learning
  • $35 hourly
    SUMMARY * Business Intelligence and Data Project Management: * Led the successful delivery of six BI projects, overseeing end-to-end development from ETL/ELT processes to reporting and dashboard creation. * Utilised DBT and the Pentaho platform to design and develop BI solutions tailored to analytics teams, integrating with Postgres, MySQL, SQLServer, Vertica, Redshift, and Oracle for schema design. * Implemented data visualization with Power BI, Metabase, and Google Data Studio. * Data Integration and Analytics Expertise: * Demonstrated proficiency in extracting and transforming data from diverse sources, including APIs and webhooks, employing formats like JSON and text. * Engineered comprehensive analytics environments by orchestrating the transition
    Featured Skill Pyspark
    PySpark
    SQL
    Python
    Java
    ETL Pipeline
    Data Mining
    Data Extraction
    ETL
  • $45 hourly
    I’m a Lead Data Engineer with 10+ years of experience, specializing in building scalable, high-performance data solutions. With expertise in Spark, Airflow, and AWS Cloud, I help businesses design and optimize data pipelines, architectures, and systems. My focus is on delivering end-to-end solutions that drive efficiency, scalability, and real-time insights. Throughout my career, I’ve designed complex ETL pipelines, developed real-time data processing systems, and provided high-level architecture solutions for businesses of all sizes. I have a passion for graph-based solutions and constraint-based executions, ensuring optimized and intelligent data workflows. I also bring deep expertise in SQL optimization to ensure that your data access and processing are as efficient as possible. Core Skills & Expertise: Data Engineering Technologies: Expertise in Apache Spark (batch/streaming), Apache Airflow, ETL processes, and cloud data solutions like AWS (S3, Glue, Lambda, RDS, Redshift). Cloud Architecture: Proficient in designing and implementing cloud-based data solutions using AWS for scalable storage, processing, and analytics. Graph-Based Solutions & Constraint-Based Execution: Skilled in creating graph-based solutions to represent complex data relationships and constraint-based execution to optimize data workflows. SQL Performance Tuning: Specialized in optimizing SQL queries and database structures, ensuring fast data retrieval and improved performance in large-scale systems. Solution Architecture: Extensive experience designing end-to-end data architectures and leading the technical direction of projects, ensuring solutions align with business goals and technical requirements. Experience Highlights: Data Pipeline Design & Optimization I design ETL pipelines that handle complex data flows, transforming raw data into actionable insights. I focus on performance and scalability, ensuring the solutions can grow with the business. Cloud Solutions (AWS) I build cloud-native solutions using AWS, including data storage (S3), processing (Lambda, Glue), and data warehousing (Redshift). I ensure solutions are cost-effective, scalable, and optimized for high performance. Graph-Based & Constraint-Based Solutions I create graph-based data models to represent complex relationships in data, allowing for more intelligent queries and optimized decision-making. My experience includes constraint-based execution models to maximize resource utilization and improve processing times. SQL Query Optimization I specialize in optimizing SQL queries, improving performance in high-volume environments. I utilize indexing, query rewriting, and execution plan analysis to ensure fast data retrieval. Leadership & Project Management: As a Lead Data Engineer, I have had the privilege of leading diverse teams and managing complex data projects. I’ve been responsible for: Designing end-to-end data solutions from scratch, focusing on scalability, maintainability, and performance. Leading teams in building data architectures, managing cloud deployments, and ensuring the success of data-centric projects. Mentoring junior engineers and guiding them in best practices for database design, coding standards, and optimizing workflows. Stakeholder Communication: I’m skilled at communicating complex technical concepts to non-technical stakeholders, ensuring alignment between technical solutions and business needs. Why Choose Me? Extensive Technical Expertise: I have a comprehensive understanding of the entire data stack and cloud infrastructure, ensuring I can handle everything from data ingestion to advanced optimization. Scalable Solutions: I design systems that can scale with your business, ensuring that data workflows can handle increasing complexity without compromising performance. Proven Leadership: I have led multiple teams, helping businesses design and implement data solutions that drive results. I ensure projects are delivered on time, within budget, and meet all technical requirements. Results-Oriented: My approach is always focused on delivering solutions that solve real business problems. Whether it’s optimizing data workflows, reducing costs, or providing actionable insights, I’m here to help you succeed. Types of Projects I Can Help With: Data Pipeline Design & Optimization: Designing, building, and optimizing data pipelines for batch and real-time processing. Cloud-Based Data Solutions: Architecting and implementing cloud-native data systems using AWS. SQL Performance Tuning & Optimization: Improving query performance and optimizing database operations. ETL/ELT Solutions: Building and optimizing ETL processes for data extraction, transformation, and loading into storage/warehouses. Data Warehouse Design & Optimization: Creating and optimizing data warehouses to ensure efficient storage and fast analytics.
    Featured Skill Pyspark
    Data Engineering
    Problem Solving
    Architecture
    Graph
    Database Management
    Database
    Apache Hadoop
    Apache Airflow
    PySpark
    Apache Kafka
  • $30 hourly
    I am highly analytical and process-oriented data engineer having in-depth knowledge of modern data engineering techniques and data analysis methods. Proven knowledge in data warehousing, databases, data quality and data cleaning techniques. My objective is to improve your business by providing Scalable, Reliable, and Secure solutions. To ensure the Client’s utmost satisfaction, I provide Consultation and brainstorming sessions to fathom a piece of the problem. I never compromise on the quality of services I provide to my clients or disavow my word. Do you want to drive your business forward using DATA to help you along the way? This is what I can do to help you: - Data Cleaning and Transformation for your dataset using Python,SQL and PySpark. - Build Big Data pipelines on AWS Cloud using ✔ S3 ✔ Glue ✔ Athena ✔ EMR ✔ Event Bridge ✔ Lambda ✔ DynamoDB ✔ Redshift ✔ Terraform ✔ SNOWFLAKE - Data Migration from Old to New Systems. - Build ETL Processes for your business case using tools like Pentaho,Talend and SSIS. - Working with Datasets in Excel Format, JSON, XML,Parquet and Different DB Files. - Working with - Simple - Data Visualizations using Tableau. - Build Databases in SQL/MySQL,PostgresSQL,MSSQL Server and Design the Database Schema. - Working with Data Warehouses, Data Cube, Star Schema, Snowflake Schema. - Can work on Programming Tasks using Python. - Scrap publically available data for your business. For deadlines, Do not worry. I'm so flexible to client requirements and aware of time differences across countries. You can be sure that i will be able to complete your tasks on time! My goal is to build long-term relationships with clients because completing tasks is my mission. Looking forward to getting you valuable insights that can help you make data-driven business decisions. - Certified Data Engineer from Data Camp - Certified in Data Manipulation from Data Camp - Certified in Importing & Cleaning of Data from Data Camp
    Featured Skill Pyspark
    Microsoft SQL Server
    PySpark
    Amazon S3
    Kubernetes
    AWS Glue
    Data Scraping
    Docker
    AWS Lambda
    Data Analytics
    SQL
    Data Science
    R
    Python
    Machine Learning
  • $12 hourly
    I am a business analyst experienced in managing data resources to optimise processes and minimise costs. My educational background comprises of a bachelor's in applied economics, and an MSc in Business Analytics from a top 1% business school. I offer the following services: - All data related tasks: Master Excel sheets, large datasets, automation of data extraction and entry. - Data cleaning and preparation for machine learning applications. - Unsure what data you need to improve your processes, cut costs, or even remain competitive? I offer business process re-engineering services for small businesses. - Data management consulting (small businesses) - you most likely already generate all the data you need to excel in your industry, all you need is adequate management and purposeful deployment. - Analytics reports, market research, and lead generation using innovative web scraping tools. - SQL database design & management. - Proficiency in using statistical analysis and tools to uncover insights and generate hypotheses. Or else, reach out with your task and let's discuss how I can add value to your business. Please note I offer most of these services in both French and Arabic as well.
    Featured Skill Pyspark
    PySpark
    Apache Spark
    Apache Airflow
    Apache Kafka
    Tableau
    Microsoft Power BI
    SQL Server Integration Services
    SQL
    Cost Management
    Arabic
    French
    Machine Learning
    Artificial Intelligence
    Data Mining
    Data Analysis
  • $20 hourly
    I’m a Data Engineer experienced in building scalable data pipelines, ETL workflows, and cloud-based data solutions. Whether you need to optimize data processing, migrate to the cloud, or enhance data governance, I can help. -> Proficient in Azure (Data Factory, Databricks, Key Vault), Snowflake, AWS Glue, Airflow, and Kubernetes -> Expertise in SQL, PySpark, Scala, and data modeling for efficient storage and retrieval -> Skilled in data warehousing, ETL/ELT, and real-time data processing -> Experience with CI/CD pipelines, GitHub Actions, and performance optimization I believe in clear communication and collaborative problem-solving, ensuring seamless execution from requirement gathering to deployment. Let’s discuss how I can help streamline your data workflows!
    Featured Skill Pyspark
    Snowflake
    AWS Glue
    Microsoft Azure
    Apache Spark
    SQL
    Databricks Platform
    Apache Airflow
    Scala
    Python
    PySpark
    ETL Pipeline
    Data Extraction
  • $15 hourly
    Result-driven data professional with international experience in data analytics, ETL development, programming, and project execution. Skilled in collaborating with teams to create data-driven solutions that align with organizational goals. Detail-oriented and an effective communicator, adept at building strong relationships with clients, stakeholders, and cross-functional teams.
    Featured Skill Pyspark
    Microsoft SQL Server Programming
    Neo4j
    Microsoft Power BI Data Visualization
    CI/CD
    GitLab
    Apache Impala
    Hive
    PySpark
    Amazon Redshift
    Amazon EC2
    Google Cloud Platform
    Interactive Data Visualization
    ETL
    Oracle PLSQL
    Bash Programming
    Big Data
    Python
    Tableau
  • Want to browse more freelancers?
    Sign up

How hiring on Upwork works

1. Post a job

Tell us what you need. Provide as many details as possible, but don’t worry about getting it perfect.

2. Talent comes to you

Get qualified proposals within 24 hours, and meet the candidates you’re excited about. Hire as soon as you’re ready.

3. Collaborate easily

Use Upwork to chat or video call, share files, and track project progress right from the app.

4. Payment simplified

Receive invoices and make payments through Upwork. Only pay for work you authorize.