Data Scientist , Deep learning , Reinforcment and control developer.
I have been working with Data Science , and AI from 3 years with predictive modeling in hard environment to designing of Deep learning Architectures for vision, control and Natural language processing. Major goal is to influence the society with growth and potential of human intelligence in augmentative AI to singularity in AI.
Result driven Analyst with over 5 years experience of analyzing current and potential business processes to identify clear opportunities for improvement
Comfortable manipulating, transforming, and analyzing complex, high-volume, high-dimensionality data from varying sources using tools such as SQL, R, Hive, Pig and Spark
Experience in applying Machine Learning and statistical modeling techniques to solve business problems
Big Data Developer with extensive experience in Hadoop Ecosystem
Big Data Applications developer with extensive hands-on IT experience (5 years) in application Design, Development, Testing and Support using Hadoop Ecosystem (Spark, Scala, MapReduce, HBase, Cassandra, Hive, Oozie, Sqoop, Kafka and Apache Nifi).
She brings an experience of manipulating/analyzing large datasets and finding patterns and insights within structured and semi-structured data. Has an expertise in gathering technical requirements and creating technical design documents with a strong working experience using agile methodologies like Scrum and Kanban.
Tools: Eclipse, SQL Server, IntelliJ, Maven, GitHub, Swagger, Postman API, Apache Tomcat, SVN and Jenkins.
Other Technology & Languages: SQL, Java, Linux, Shell Scripting, REST and JSON.
Holds a Permanent Residency of Canada and a valid B1 visa for USA with a keen interest in proving her potential across geographies. Has an experience of working in Canada, UK and India, both in large teams and as an individual contributor.
Experienced and driven data scientist with eight years of experience in the industry. Strong background in engineering research, computer programming, soft computing and knowledge of various types of databases and data warehouse. Solid skills in mathematics, statistics and algorithms. Commitment to providing support and essential information about trends to companies in a variety of industries.
I am looking for projects that involve signal/image processing and/or data analysis and machine learning. Projects with a medical imaging component would be ideal, but I would be very happy to work on any project that has any of the above-mentioned aspects.
I am a Research Analyst with experience in signal/image processing and machine learning. My major expertise is in MATLAB (10 years) and Python (4 years) and their libraries, though I am comfortable programming in Java and C# too. I have developed algorithms and full blown programs for real time medical signal processing and analysis which are in use in a major Toronto hospital. My work has resulted in multiple publications and even a couple of patents.
Over 8+ years of diversified experience in IT industry and with extensive exposure to Hadoop Ecosystem: Map Reduce, Spark, Sqoop, Hive, Kafka, Pig and HDFS.
Hands-on knowledge of Big Data environments: Hadoop (CDH, HDP and EMR), MapReduce in Java, Spark and RDDs for data science, Sqoop ingests, Flume, Hive, NoSQL databases such as HBase.
Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and vice versa.
Well versed with Spark processing Framework such as Spark and RDDs, Spark SQL, DataFrames and Spark Streaming.
Expertise in delivering various project types and end to end software implementation such as requirements gathering, documentation, analysis, design, development, debugging, testing, deployment, porting and support/maintenance.
First hand exposure to global clientele/projects across geographies (USA, Canada, UK, Sweden, Liechtenstein, India) along with worldwide Customer Engineering experience working at Qualcomm.
Profound understanding of various software development execution models like Agile & Scrum, traditional Waterfall and Adaptive Iterative model.
Skilled in executing projects with strategic global delivery model. Managed and trained team (both co-located and virtual) & ensured successful delivery.
A unique blend of experience ranging from exposure to hadoop ecosystem to mobile apps development including implementation of build enhancements, new interfaces and support provision on both Android and Brew mobile platforms. Along with exposure to windows apps development, and IoT and M2M devices.
Proficient in English communication (Scored 8.0 on IELTS), exhibits adept interpersonal skills and a quick learner, possessing an ability to understand and adapt to new technologies and environments faster.
Confident decision maker with fine analytical, problem solving and technical skills with a strong desire to achieve specified goals.
I am a Data Scientist and an applied Science Researcher and have extensive experience in solving various data driven engineering challenges. Some of my key attributes are:
- have developed solutions big companies like GE, SUEZ and developed proprietary algorithms for machine learning and data analytics for various other clients
-have advanced programming skills using Python (e.g., numpy, pandas, scikit-learn, keras/TesnsorFlow (deep learning)), R, Matlab, and VBA Excel Macro
-Have advanced knowledge in descriptive and inferential statistics and various Machine Learning techniques and tools including predictive and classification algorithms
-Proficient in using various advanced tools such as PostgreSQL/ MySQL, InfluxDB, Firebase, Grafana
-have advanced level skills in Hadoop and Spark
I can guarantee you the best solution for your data challenges
Big Data professional with over 10 years of IT experience and 7 years in Data Engineering. Solved many use cases for American retail clients like Target and Lowes. Also implemented data lakes and data pipelines for Mercedes Benz and Citco group of companies using Hadoop and AWS technologies. Following are my technical skills:
Hadoop, Hive, MapReduce, Pig, Spark(PySpark), Oozie, Sqoop, SQL, RDBMS, Data warehousing, Redshift, AWS,Blur Prism, BigQuery, Data Engineering and reporting.
I'm looking to do some simple work to cover expenses while I am searching for a full-time job. It does not have to be relevant to my experiences. I have just completed my undergrad with a double major in mathematics at the University of Waterloo and I have completed work placements at technology companies doing software development and data engineering.
My first internship was at the Toronto Transit Commission (TTC) where I did work under a professional electrical engineer, where I designed and evaluated plans and procedures for maintenance work done on the subway rail vehicles. My second internship was as a research assistant under where I did work on implementing an algorithm that computed a consistency metric for distributed databases using Spark and Hadoop. My third internship and fourth internships were at separate startups doing data engineering.
I also have some experience with data science and am currently working on some data science projects on Kaggle while studying relevant data science topics like machine learning and data visualization through textbooks and online courses.
Working knowledge of Econometric Models, Decision Choice Analysis, User Behavioral Models, and Pricing
Working knowledge of database, SQL, MySQL, relational and non-relational database
Proficiency in GIS geoprocessing and spatial data analysis
Proficiency in image processing and remote sensing
Proficiency in stochastic and risk analysis and prediction models
Working knowledge of big data mining, statistical and spatial and network analysis, and quantitative analysis techniques
Extensive working knowledge of big data and applied machine learning and statistical regression models (e.g., Supervised and Unsupervised Classification, Clustering Algorithm, Optimization, Neural Networks, Deep Learning, CNN, RNN, Regression, Naïve Bays, Prediction Models), Natural Language Processing
Mobile and web application developing, Android, iOS
We are a shop that is exclusively focused on Apache Spark, Apache Storm, and Apache Storm. You can check out flag ship product at the web at streamly.io. Our team members have PHDs and Masters and have worked for such companies as Cisco, Verizon, Overstock, ATT, etc.
HADOOP BIG DATA:
Designed and developed ELK stack to collect Hadoop cluster logs, customized reports generation and analysis. Designed alerting and Kibana dashboards for Hadoop cluster health check
Designed and developed solution for developers Hadoop log access and provided easy means of reporting and analysis by combining Python+Painless scripting with ELK stack
Clickfox integration with Big data cluster and data ingestion pipeline configuration for data analytics using Clickfox
Apache zeppelin Installation on Hadoop cluster and configuration of Zeppelin with Livy Interpreter. Configuration of Kerberos security on Zeppelin to provide developer access to Hadoop cluster.
Ranger Installation, configuration and active directory integration, policy administration, audit logging and plugin maintenance. HDFS, HBASE, HIVE, YARN, KNOX, STORM and KAFKA access control through ranger policies.
HDP 18.104.22.168/Ambari 22.214.171.124 upgrade planning, designing and implementation across different regions
Dr Elephant configuration for yarn and spark jobs performance measuring and tuning
Hadoop stress testing using Terasort, Teragen, NNBenchmark, TestDFSIO
Hortonworks Hadoop installation through Ambari and manual process.
In memory fast data processing design using Spark and live data streaming using Storm topologies
Data ingestion optimization design by replacing traditional message processing with Apache Kafka design.
HDFS file system optimization by better YARN queue design, preemptive memory design, maintenance, data node and name node maintenance with high availability and replication
Hadoop security and access maintenance using Kerberos, Ranger, Knox, ACLs etc
Storm topology designing, architecture, integration with Active MQ broker, Kafka, HDFS data lake
Cluster monitoring using Ambari metrics, logstach, elastic search, Kibana dashboard etc
Hadoop migration, upgrades and other maintenance activities
Rack awareness setup, auditing and other activities
DB2 LUW Administration:
Installation of DB2 on AIX system
Fixpack Upgrade and database migration
Instance & DB configuration analysis & tweaking as per performance
Database monitoring for troubleshooting the application level performance issues
AIX System monitoring for analyzing database server level performance issues which impacts database performance
Database Server Capacity Planning in terms of CPU utilization, Storage, Memory
Understanding of Power Systems & AIX Operating systems
Handling all the maintenance jobs on the database servers
HMC administration and controlling power systems through HMC
Understanding of SAN storage to allocate database physical layout for better performance
Understanding of TSM storage to handle the database level and system level backups & its recovery
Ability to Identify the Poorly performing queries from database on-line monitoring
SQL tuning using query explain to identify bottleneck area for query performance
Shell, PERL, AWK and SED scripting
Migrating DB instances, installing DB2 patches, creating instances, maintaining server disc, RAM memory
AIX memory analysis with tools like VMSTAT, NMON, IOSTAT, etc
Q-replication implementation, automation and knowledge transfer to project teams and worked on SQL replication
Sybase Adaptive Server Enterprise administration:
Monitor the databases to optimize db performance, checking for db consistency, fixing DBCC errors, monitored entire db for performance metrics & backup/recovery process implementation.
Sybase server Installation : Installed Sybase servers on versions starting from 10.
Shell Scripting : Created scripts for Monitoring and Maintenance task, automated Database and transaction dumps, monitoring data received at 15 min interval was received and populate reporting tables.