Hi, I have reviewed the project details and understood your requirement for a Spark Expert who can help to optimize your Spark Streaming Application. Along with below tasks:
- Continuously monitoring and processing data from Kafka using Spark.
- Storing the results into a Hadoop Hive table.
I am capable of completing this task in a timely and cost-effective manner. Let's connect and discussion more on it.
Short Into about Me:
Currently working in a Top Financial Institution of USA & previously worked for one of the Big4 Companies and solving Big Data Business problem using AWS Cloud and PySpark Engine.
I have expertise on the following Tools/Technologies:
- Apache Kafka, Spark Structured Streaming, Glue Streaming
- Apache Spark (PySpark), Hive
- AWS Cloud Services (Step function, Lambda, Kinesis, DynamoDB, RDS, Aurora, EMR, S3, Glue Lambda, EBS, Athena, IAM, Step function, EC2, ECS, secret manager (kms), s3, sns, ses, sqssqs, cloudwatch etc)
- Python3, pydbc, unit test script, NumPy, Pandas
- Creation of cloud formation templates for various AWS Services
- Databricks, Delta Table, Data Lake, Data pipeline, ETL, ELT
- CI/CD Pipelines
- SQL/MySQL
- Git, GitHub, Bit Bucket
- Big Data Pipelines
- IBM DB2
- MS Excel, MS Office
- PyCharm, VS Code, Databricks, Jupyter Notebook, MS Excel, Putty
Please provide more details if it's more than what you have given in the project details. Thank you!
Regards,
Rajeev