Descripción del empleo: Contribute to the design of scalable Big Data solutions across the entire data supply chain with a focus on ensuring the functionality delivered can be monitored for health and the design is extensible Monitoring the health of existing processes to ensure they are delivering output as expected and proactively raising issues when they are not Modify existing data integration and data warehouse processes to account for requirement changes Participate in 2nd level production support Verify accuracy of data, testing methods and the maintenance and support of the Analytics Data Platform Collaborate with management, business partners, analysts, developers, architects, and engineers to support all data quality efforts Create and review technical and user-focused documentation for data solutions (data models, data dictionaries, business glossaries, process and data flows, architecture diagrams, etc.) Operational functions You don't just learn how things work, you learn why. Understanding how systems work at a fundamental level is a passion of yours Be open and willing to learn new skills!
WHAT SKILLS ARE WE LOOKING FOR IN AN IDEAL CANDIDATE?
In data management, data access (Big Data, traditional Data Marts and Data Warehousing).
In Advanced programming (python, Shell scripting, and Java)
With interactive and batch processing using Spark SQL and spark scripting.
In applied data technologies: Hadoop Spark Kafka, Spark Streaming Pig Hive MongoDB Oozie EMR Lambda SQL
Current data warehousing concepts and technologies like Redshift, Spark, Hadoop, web services etc. to support business-driven decisioning
In data architecture and data assembly
In Data Governance and Data Security
Functional requirements, detailed technical specifications, and test cases for new or modified projects
Understanding of data sources (e.g., 3rd party RDBMS, MS access, SQL server, Oracle, and MySQL)
Data integration tools (Talend preferred)