Proficient data engineer with 5+ years of experience in implementing a diverse range of complex projects using Python, Hadoop, AWS, SQL, and Web Crawling. Experienced, results-oriented, resourceful, and problem-solving data engineer with leadership skills including two years of managerial experience, looking to build robust databases that lead to client revenue increases. Enjoy creative problem solving and getting exposure to multiple projects, and I would excel in the collaborative environment on which your company prides itself. Data Scientist with strong math background and experience in machine learning techniques such as predictive modeling, data processing, and data mining algorithms to solve challenging business problems and professional experience in advanced analytics, machine learning, natural language processing, text analytics, and web scraping techniques.
As a Data Scientist, I have worked on Data Science lifecycle concepts from Data Collection, Data Extraction, Data Cleansing, Data Exploration, Data Transformation, Feature Engineering, Data Integration, Data Mining, building Prediction models, Data Visualization and deploying the solution to the customer. Proficient data engineer with 3 years of experience in the financial industry with extensive experiences in data pipelining, web scraping and automation technology. I have worked on Python, Beautiful Soup, Request, UrlLib, MongoDB, SQL to fetch data from various data sources for the purpose of model building and predictive modelling.
Used predictive analytics such as machine learning and data mining techniques to forecast output with a 95% accuracy rate.
Wrote Python scripts, using Pandas, to parse and load the data into the database from the JSON files.
Used the Beautiful Soup Python Library to extract desired data from a given URL an automated way.
Performed in agile methodology, interacted directly with entire team provided/took feedback on design, Suggested/implemented optimal solutions, and tailored application to meet business requirement and followed Standards.
Re-designed and developed a critical ingestion pipeline to process over 300 TB of data
Performed DBA activities like performing Vacuum and Analyse for tables, creating tables, views, recovery and cluster monitoring and maintenance.
Identified areas for improving efficiency, shortening of test cycles, and automation of test scripts and recommend solutions for improvement.
Used Git version control system for collaborating with teammates and maintaining code versions.
As Data Architect designed and maintained high performance ELT/ETL processes.