1) Processing high volume of data and tuning [login to view URL] assigning cluster nodes ,fine tuning pyspark jobs, various optimizations used at source table
2) Demonstration of handling data skewing , data spilling , data compression & data serialization and other best practices 3) various cost optimizations designs implemented in AWS from data standpoint ( Glue , S3, EMR, DMS , Athena, Redshift or any other services that you have worked/come arcoss)
If you are comfortable with anyone of the above points do reach out.
11 freelancer chào giá trung bình $19/giờ cho công việc này
HI I am experienced in Amazon Web Services Hadoop Ubuntu etc I can start right now but i have few doubts and questions lets have a quick chat and get it started waiting for your reply