Senior Data Engineer
1 Nos.
109065
Full Time
3.0 Year(s) To 7.0 Year(s)
Not Disclosed by Recruiter
IT Software- DBA / Datawarehousing
Banking/Financial Services
B.Tech/B.E. - Computers
Job Description:
Job Description
As a Senior Data Engineer, you will play a key role in designing and implementing data solutions @Kotak811.
- You will be responsible for leading data engineering projects, mentoring junior team members, and collaborating with cross-functional teams to deliver high-quality and scalable data infrastructure.
- Your expertise in data architecture, performance optimization, and data integration will be instrumental in driving the success of our data initiatives.
Responsibilities
- Data Architecture and Design:
- Design and develop scalable, high-performance data architecture and data models.
- Collaborate with data scientists, architects, and business stakeholders to understand data requirements and design optimal data solutions.
- Evaluate and select appropriate technologies, tools, and frameworks for data engineering projects.
- Define and enforce data engineering best practices, standards, and guidelines.
- Data Pipeline Development & Maintenance:
- Develop and maintain robust and scalable data pipelines for data ingestion, transformation, and loading for real-time and batch-use-cases
- Implement ETL processes to integrate data from various sources into data storage systems.
- Optimise data pipelines for performance, scalability, and reliability.
- Identify and resolve performance bottlenecks in data pipelines and analytical systems.
- Monitor and analyse system performance metrics, identifying areas for improvement and implementing solutions.
- Optimise database performance, including query tuning, indexing, and partitioning strategies.
- Implement real-time and batch data processing solutions.
- Data Quality and Governance:
- Implement data quality frameworks and processes to ensure high data integrity and consistency.
- Design and enforce data management policies and standards.
- Develop and maintain documentation, data dictionaries, and metadata repositories.
- Conduct data profiling and analysis to identify data quality issues and implement remediation strategies.
- ML Models Deployment & Management (is a plus)
- Responsible for designing, developing, and maintaining the infrastructure and processes necessary for deploying and managing machine learning models in production environments
- Implement model deployment strategies, including containerization and orchestration using tools like Docker and Kubernetes.
- Optimise model performance and latency for real-time inference in consumer applications.
- Collaborate with DevOps teams to implement continuous integration and continuous deployment (CI/CD) processes for model deployment.
- Monitor and troubleshoot deployed models, proactively identifying and resolving performance or data-related issues.
- Implement monitoring and logging solutions to track model performance, data drift, and system health.
- Team Leadership and Mentorship:
- Lead data engineering projects, providing technical guidance and expertise to team members.
- Conduct code reviews and ensure adherence to coding standards and best practices.
- Mentor and coach junior data engineers, fostering their professional growth and development.
- Collaborate with cross-functional teams, including data scientists, software engineers, and business analysts, to drive successful project outcomes.
- Stay abreast of emerging technologies, trends, and best practices in data engineering and share knowledge within the team.
- Participate in the evaluation and selection of data engineering tools and technologies.
Qualifications:
- 3-5 years’ experience with Bachelor's Degree in Computer Science, Engineering, Technology or related field required
- Good understanding of streaming technologies like Kafka, Spark Streaming.
- Experience with Enterprise Business Intelligence Platform/Data platform sizing, tuning, optimization and system landscape integration in large-scale, enterprise deployments.
- Proficiency in one of the programming language preferably Java, Scala or Python
- Good knowledge of Agile, SDLC/CICD practices and tools
- Must have proven experience with Hadoop, Mapreduce, Hive, Spark, Scala programming. Must have in-depth knowledge of performance tuning/optimizing data processing jobs, debugging time consuming jobs.
- Proven experience in development of conceptual, logical, and physical data models for Hadoop, relational, EDW (enterprise data warehouse) and OLAP database solutions.
- Good understanding of distributed systems
- Experience working extensively in multi-petabyte DW environment
- Experience in engineering large-scale systems in a product environment
Location – Bangalore
Company Profile
Is an Indian ---ing and financial services company headquartered in Mumbai. It offers ---ing products and financial services for corporate and retail customers in the areas of personal finance, investment ---ing, life insurance, and wealth management.
Apply Now
- Interested candidates are requested to apply for this job.
- Recruiters will evaluate your candidature and will get in touch with you.