Big Data Lead Hadoop Developer
Over 15 years of experience in BigData/Java Stack as Manager/Architect, a collaborative engineering professional with substantial designing and executing solutions for complex business problems involving large scale data warehousing, real-time analytics and reporting solutions. Known for using the right tools when and where they make sense and creating an intuitive architecture that helps organizations effectively and process terabytes of structured and unstructured data.
Proven history of building large-scale data processing systems and serving as an expert in data warehousing solutions while working with a variety of database technologies. Experience architecting highly scalable, distributed systems using different open source tools as well as designing and optimizing large, multi-terabyte data warehouses. Able to integrate state-of-the-art Big Data technologies into the overall architecture and lead a team of developers through the construction, testing and implementation phase.
Databases and Tools: MySQL, MS SQL Server, Oracle, DB2, NoSQL: HBase, MongoDB, Cassandra, HDFS
Consulted with business partners and made recommendations to improve the effectiveness of Big Data systems, descriptive analytics systems, and prescriptive analytics systems. Integrated new tools and developed technology frameworks/prototypes to accelerate the data integration process.
Tools: Hive, Pig and Hadoop Streaming, MapReduce, YARN, Bedrock, Kafka, Apache Spark
Experience designing, reviewing, implementing and optimizing data transformation processes in the Hadoop and ETL ecosystems. Able to consolidate, validate and cleanse data from a vast range of sources – from applications and databases to files and Web services.
Tools: Tivoli, Custom Core Framework and Scripting; Linux/UNIX Commands.
Capable of extracting data from an existing database, Web sources or APIs. Experience designing and implementing fast and efficient data acquisition using Big Data processing techniques and tools. Tools: APIs and SDKs, RESTful Interfaces.
Knowledge of Apache Spark, Storm, Apache Kafka and Amazon EC2.
Hands on experience on installing, configuring Hadoop Cluster of Apache Hadoop distribution; with knowledge of Apache Lucene, Map R, Splunk, Mahout Tools.
Instrumental in architecture of data warehouses, OLTP and OLAP solutions.
Strong track record in Enterprise IT systems, JEE frameworks and SOA design patterns for building stable and scalable systems.
Capable of processing large sets of structured, semi-structured and unstructured data and support system application architecture.
Extensive experience in Software SDLC and Agile processes, managed cross geography location scrum teams (up to 40 members) with producing high quality end to end development and delivery with [On/Off]shore model.
Experience in HealthCare, Financial, Banking and Manufacturing domains.
Extensive experience in OOAD, UML, JEE and Core Java design patterns.