Title: Data Modeler (Azure Databricks)
Location: Jersey City, NJ
Type: Hybrid, 2 days a week
Duration: 6+ months
Visa: All except H1B
Interview Process: Onsite interview.
Key Responsibilities:
• Design, implement, and optimize data models for the data warehouse, ensuring alignment with business requirements.
• Work extensively with **IBM IDA (InfoSphere Data Architect) ** to create logical and physical data models.
• Ensure that the data models are scalable, performant, and aligned with the organization’s data strategy.
• Collaborate with data engineers to implement and maintain data models in **Azure** and **Databricks** environments.
• Translate business requirements into conceptual, logical, and physical data models, ensuring data quality and consistency across systems.
• Participate in data architecture and design discussions to ensure data modeling best practices are followed.
• Develop and maintain documentation of data models, data flow diagrams, and database design specifications.
• Work with stakeholders to ensure the data models support analytics, reporting, and data integration needs.
• Perform impact analysis of changes to existing models and collaborate with teams to assess the effects on downstream systems.
• Continuously improve and optimize data models to enhance performance and data accessibility.
Qualifications:
• Bachelor’s degree in computer science, Information Systems, or a related field.
• Proven experience as a **Data Modeler** working in data warehouse environments.
• Strong expertise with **IBM IDA (InfoSphere Data Architect) ** or similar data modeling tools.
• Solid understanding of **Azure** and **Databricks** architecture, and experience working in cloud-based data environments.
• Proficiency in relational and dimensional data modeling techniques, including star and snowflake schemas.
• Hands-on experience with SQL and database design, with knowledge of data lakes, data warehouses, and ETL processes.
• Strong analytical and problem-solving skills, with attention to detail.
• Ability to communicate complex technical concepts to both technical and non-technical stakeholders.
• Familiarity with data governance, metadata management, and data quality practices.
• Experience with **Azure Synapse**, **Azure Data Factory**, and **Databricks Delta Lake**.
• Knowledge of big data technologies and principles of data lakehouse architectures.
• Experience in agile methodologies and working in a fast-paced, collaborative environment
Note:
• Must have banking client experience.
• Require a reference from most recent employment (or employment before that if still currently there)