Job Responsibilities
* Develop, test and support future-ready data solutions for customers across industry verticals
* Develop, test and support end-to-end batch and near real-time data flows/pipelines
* Demonstrate understanding in data architectures, modern data platforms, big data, ML/AI, analytics, cloud platforms, data governance and information management and associated technologies
* Design and implement data models according to business requirements for specific use cases and/or client’s business domains
* Demonstrate understanding in data modelling schema design approaches and techniques such as Third Normal Form (3NF), Dimensional Modelling, Entity-Relationship (ER) Models and Data Vault
* Demonstrate understanding in Data Asset concept or Data Mesh Architecture and its Domain Data Products
* Develop and demonstrate Proof of Concepts and Working Demos
* Lead or collaborate with other internal/external consultants in consulting, workshops and delivery engagements
* Mentor junior IBM consultants in the practice and delivery engagements
Required Professional and Technical Expertise
* Minimum of 10+ years of total work experience as a solution designer, developer, tester or support role in IT Consulting or other technology business units across the industry
* Minimum of 3+ years of hands-on development, testing and administration experience with Big Data/Data Lake services in cloud and on-premise
* Minimum of 3+ years of designing and implementing data models in Big Data/Data Lake or Data Warehouse environment using modelling approaches and techniques such as Dimensional Modelling on top of Star or Snowflake Schemas
* Experience in implementing data model products (IFW/BDW/FSDM) from vendors such as IBM and Teradata
* Hands-on experience in data model design tools such as Erwin and Sparx Systems Enterprise Architect
* Experience in applying industry best practices, design patterns and first-of-a-kind technical solutions as developer or administrator on data and non-data specific platforms and applications
* Experience in implementing near real-time data flows/pipelines and distributed streaming applications
* Experience in implementing traditional ETL, data warehousing and BI solutions
* Hands-on development experience and working skill level in Python, Scalar, SQL, shell scripting and other programming/scripting languages
* Hands-on implementation of services sensitive to performance SLAs with high availability, fault-tolerance, automatic fail-over and geographical redundancy
* Working knowledge and hands-on experience with data services (Azure Synapse, Cosmos DB etc) in Cloud Platform such as Azure, AWS, GCP and IBM and other modern data platforms
* Solid understanding of containerisation, virtualisation, infrastructure and networking
* Diverse experience in software development methodologies and project delivery frameworks such as Agile Sprints, Kanban, Waterfall etc
* Experience in presenting to and influencing stakeholders and senior managers
* Team leadership and people manager experience
* Degree in Computer Science, Information Technology or related Engineering courses
Desired Attributes and Skills
* One or more industry-recognised technology certifications or badges
#J-18808-Ljbffr