Hire the best Pyspark Developers in China
Check out Pyspark Developers in China with the skills you need for your next job.
- $25 hourly
- 5.0/5
- (5 jobs)
I’m a developer with experience in NLP、LLM 、 CV and recommendation system and bigdata. 1. I’m experienced in RAG、langchain、deepspeed、llm、 object dection、video generation 2.I’m experienced in hadoop/hdfs/yarn/hbase/redis/kafka/hive, 2. I’m experienced in tensorflow/kubeflow/tf serving/trtion serving 3.I’m experienced in k8s/docker/html/jqurey/spring boot/mybatis 4.I’m experienced in aws componet, such as emr/ec2/s3/code deployPyspark
jQueryLangChainLLM Prompt EngineeringPySparkAWS ApplicationBig DataArtificial IntelligenceApache HadoopSpring BootKubernetesApache FlinkJavaTensorFlowApache Spark - $80 hourly
- 0.0/5
- (0 jobs)
A highly skilled and motivated Web Full-Stack Engineer with a strong focus on Big Data, I am able to excel in both regular web project development and data analysis. With extensive experience in creating advanced reporting systems, I am confident in my ability to deliver high-quality results that exceed client expectations. Skills 1. Programming Languages: Proficient in JavaScript, HTML5, CSS3, Python, Java, and C++. 2. Frameworks and Libraries: Familiar with React, Angular, Node.js, Express, Bootstrap, jQuery, and D3.js for front-end development; and Django, Flask, Spring Boot, and Hibernate for back-end development. 3. Big Data Technologies: Experience with Hadoop, Spark, Kafka, Elasticsearch, and NoSQL databases such as MongoDB and Cassandra. 4. Cloud Computing: Knowledge of AWS, Azure, and Google Cloud Platform for deploying and managing cloud-based applications. 5. Database Management: Skilled in using MySQL, PostgreSQL, Oracle, and SQL Server for database design, optimization, and maintenance. 6. Reporting Systems: Experienced in developing and implementing advanced reporting systems using tools like Tableau, Power BI, and SSRS. 7. Version Control: Proficient in using Git for version control and collaboration with team members. 8. Agile Methodology: Familiar with Agile development methodologies, such as Scrum and Kanban, and able to work effectively in a fast-paced environment. 9. Communication and Collaboration: Excellent communication and interpersonal skills, able to work well in a team and collaborate with cross-functional teams.Pyspark
Docker ComposeLinuxClickHousePyQtPySparkReact NativeNext.jsJavaScriptKubernetesDockerPostgreSQLMySQLSpring BootJavaPython - $25 hourly
- 5.0/5
- (1 job)
1. more than 8 years experience in data analytics field 2. worked 10 years in international projects, fluent communication in English & German 3. good knowledge in the following technologies, including Tableau, PowerBI, Python, R, SQL, ExcelPyspark
AngularTableauAI Agent DevelopmentApache FlinkPySparkScalaPythonJavaData ExtractionETL - $45 hourly
- 0.0/5
- (0 jobs)
I’m a senior technology professional specializing in full-stack development, cloud infrastructure, and data engineering, helping businesses accelerate digital transformation. Whether you need industrial IoT solutions, cloud migrations, or data governance systems, I deliver scalable, secure, and cost-effective outcomes. Skills: Languages: Go, Python, Java, HTML/CSS/JS, Vue Cloud: AWS, Azure, Aliyun, K8s, Docker, Terraform Data: Apache Spark, Airflow, Databricks, MySQL, MongoDB Security: Nessus, vulnerability automation, compliance frameworks Services: End-to-End Project Leadership: Manage teams of 50+ to design, develop, and deploy complex systems (e.g., cloud-based CMDB, data pipelines for 10B+ daily records). Digital Transformation: Migrate legacy systems to cloud-native architectures, automate workflows, and optimize manufacturing processes (e.g., reduced BMW’s data latency by 70%). Security & Compliance: Build automated threat detection systems (achieved 95% vulnerability scanning efficiency for BMW’s defense center). Why Work With Me? Proven Track Record: Delivered multi-cloud projects for Fortune 500 clients like BMW and Huawei. Technical Depth: Expert in full-stack development, containerization, and data governance. Collaborative Approach: Prioritize clear communication and align technology with business goals. Let’s connect to discuss how I can help your business leverage cutting-edge technology for growth.Pyspark
Data EngineeringDatabricks PlatformAlibaba CloudAWS DevelopmentAzure DevOpsPySparkDockerKubernetesReactNodeJS FrameworkGolangPythonETLData Extraction - $25 hourly
- 0.0/5
- (0 jobs)
As a data engineer with 3+ years of experience, I specialize in data middle platform architecture, ETL optimization, and data governance, leveraging technologies like Hadoop, Spark, Python, and AWS to deliver scalable solutions. My strengths lie in translating complex business needs into robust technical frameworks while ensuring data security and efficiency. Key Projects & Achievements Data Middle Platform Design: Spearheaded the architecture of a high-availability platform using Hadoop and Spark, processing 10G+ daily incremental data. Reduced query latency by 40% via optimized indexing and caching, enabling real-time analytics for 90% of business units. Insurance Customer Analytics: Built logistic regression models (Python/Pandas) to identify high-potential insurance buyers, achieving 15% accuracy improvement and informing product innovation strategies. Cost Optimization: Slashed storage costs by 30% by implementing AWS Redshift Spectrum for S3 cold data querying and lifecycle management.Pyspark
Machine LearningData CleaningData Analytics & Visualization SoftwareTableauAmazon RedshiftMongoDBMySQLData PreprocessingData EngineeringApache HadoopPySparkApache KafkaApache AirflowElasticsearchPython - $20 hourly
- 0.0/5
- (0 jobs)
Proficient in Java development, with exper4se in Spring, MyBa4s, and related architectures. • Proficient in designing distributed high-concurrency systems. • Strong knowledge and experience with Redis and Ka`a. • Skilled in Java memory tuning, mul4-threaded programming, and solving online problems. • Proficient in using tools such as jmap, jstat, jstack, mat, jprofiler, greys, and other related tools. • Experienced in database development, including SQL development, par44oning, and database op4miza4on. • Familiar with Linux development environment and shell programming. • Knowledgeable about Hadoop and Spark, with exper4se in Spark for data processing. • Proficient in NLP technologies such as HMM, CRF, LSTM, CNN, Transformer, BERT, etc., including hands-on experience with TensorFlow. • Experienced in TensorFlow model deployment and TensorFlow serving. • Familiarity with Naive Bayes, Logis4c Regression, Decision Trees, Ensemble Learning, and other algorithms.Pyspark
PySparkNatural Language ProcessingPythonJava - $30 hourly
- 0.0/5
- (1 job)
I am Chang, currently pursuing a Master's degree at HEC Paris. Being an experienced campus ambassador at HEC Paris, I have an in-depth understanding of the application processes for business schools. I can bring both a business perspective as a management student, as well as an academic perspective as an engineering student, into your essay. I will work to make sure that your resumes and essays are creative, professional, and precise. I can help with: - Polishing business school application materials (Essays, CV, RL, etc.) - Web Scraping - Web development - Data Engineering with Spark/Hive - Data Analytics with R/Python/VBA - Data Visualization with Tableau/PowerBI/Neo4jPyspark
EssayNatural Language ProcessingMachine LearningWeb ScrapingJavaScriptFigmaHiveTableauJavaCSSHTMLSQLPythonPySparkR - $30 hourly
- 0.0/5
- (0 jobs)
- Experienced data professional with 15+ years in high-tech and global corporations, spanning content, mobility, retail, and finance sectors. Proven expertise in the end-to-end data lifecycle, including predictive modeling, data mining, and data warehousing. Collaborative strategist adept at aligning data science solutions with business challenges, delivering quantifiable results, and driving impactful decisions.Pyspark
Microsoft ExcelGitPySparkHiveIBM Cognos TM1CognosMicrosoft Power BITableauPythonData Science - $11 hourly
- 4.3/5
- (1 job)
I'm a Data analyst experienced in big data analysis and data science. - Knowns python, MySQL, Java and Spark.Pyspark
RPySparkJavaMySQLPython - $80 hourly
- 0.0/5
- (0 jobs)
I am proficient in all Python-related technologies. Such as Pygame,Pyspark,PyQt,PyQtgraph and so on.Pyspark
PyQtGraphPyQtPyCharmPygamePySparkPyTorchPython - $100 hourly
- 0.0/5
- (0 jobs)
I am a highly skilled and motivated Data Engineer with a proven track record of delivering high-quality data solutions that drive business value. I specialize in extracting, transforming, and loading (ETL) data from various sources into structured data warehouses and analytics platforms. I have expertise in designing and implementing scalable data pipelines that can handle large volumes of data efficiently, ensuring data integrity and quality throughout the process.Pyspark
Agriculture & MiningETLETL PipelineData ExtractionData AnalysisWarehouse ManagementPySparkApache Hadoop - $3 hourly
- 0.0/5
- (0 jobs)
Professional Summary Seasoned Java Developer with over 10 years of experience in software development, specializing in big data technologies, cloud computing, and distributed systems. Proficient in designing, developing, and optimizing high-performance applications using a wide range of modern frameworks and tools. Adept at working with large-scale data processing platforms and real-time computing environments. Technical Skills Programming Languages: Java, Python, Shell Frameworks: Spring Boot, Spring Cloud, FastAPI, Netty Big Data Technologies: Spark, PySpark, Flink, Kafka, Hive, Hadoop, Doris Cloud & Containerization: Kubernetes, Rancher, Harbor Databases: MySQL, Hive Development Tools: Git, Maven, Jenkins Others: Real-time computing, Data engineering, Performance tuningPyspark
Data Warehousing & ETL SoftwareSpring BootDockerKubernetesNettyApache KafkaHiveFastAPIApache FlinkPythonSpring CloudPySparkApache SparkWeb Development - $56 hourly
- 0.0/5
- (0 jobs)
I am an experienced data developer who is good at building data warehouses for small and medium-sized enterprises and extracting potential value from them. I am proficient in spark, flink, hive, scala, java, python. I attach great importance to regular communication, so please keep in touch.Pyspark
PythonJavaScalaElasticsearchHiveApache HBaseApache KafkaApache FlinkPySparkETL PipelineData ExtractionETL - $30 hourly
- 0.0/5
- (0 jobs)
Al and Data Engineer specializing in GenAl and big data solutions to drive innovative business outcomes.Pyspark
PySparkPythonData AnalysisArtificial Intelligence - $5 hourly
- 0.0/5
- (0 jobs)
⚡️ Deliver 100% satisfaction on your project is most important to me. ⚡️ My area of expertise: Frontend: 🔸 React 🔸 Javascript, TypeScript 🔸 HTML5, CSS Backend: 🔸Java, Python 🔸Redis, MySQL, MongoDB DevOps: 🔸Git 🔸Docker Big Data: 🔸Components: Hadoop, Hive, Spark, Flink, Kafka, Hbase, Clickhouse, Doris, Iceberg 🔸Others: Data warehouse model design, ETL design and develop I am looking forward to work with you and your company, to provide reliable, consistent, and High-Level solutions to ensure your business and ideas succeed!Pyspark
Data Warehousing & ETL SoftwareJavaPySparkSQLPythonJavaScript Want to browse more freelancers?
Sign up
How hiring on Upwork works
1. Post a job
Tell us what you need. Provide as many details as possible, but don’t worry about getting it perfect.
2. Talent comes to you
Get qualified proposals within 24 hours, and meet the candidates you’re excited about. Hire as soon as you’re ready.
3. Collaborate easily
Use Upwork to chat or video call, share files, and track project progress right from the app.
4. Payment simplified
Receive invoices and make payments through Upwork. Only pay for work you authorize.