Python Data Engineer

Python Data Engineer

Job Title: Data EngineerJob Summary:We are looking for a proficient Data Engineer with expertise in Amazon Redshift, Python, ApacheAirflow, dbt (Data Build Tool), API integration, and AWS. This role will be responsible for developingand maintaining scalable data pipelines, integrating data from multiple sources, and ensuring thatour data architecture supports business intelligence, reporting, and analytics requirements. You willcollaborate with cross-functional teams to build and optimize our data infrastructure and provideclean, high-quality data to the business.Key Responsibilities: Data Pipeline Development: Build and maintain robust ETL/ELT pipelines using Python, ApacheAirflow, and dbt to extract, transform, and load data from various sources into Amazon Redshift. Amazon Redshift Management: Design, optimize, and maintain Amazon Redshift clusters,ensuring the warehouse is capable of handling large-scale data efficiently. API Integration: Develop solutions to integrate external APIs for data ingestion, ensuring properdata extraction, transformation, and integration into our data infrastructure. Data Modeling: Create and maintain scalable data models in Redshift that support analytics andreporting needs, including designing star and snowflake schemas for optimized querying. AWS Infrastructure Management: Leverage AWS services such as S3, Lambda, EC2, andCloudWatch to build and maintain a scalable and cost-efficient data architecture. dbt (Data Build Tool): Use dbt to manage and automate SQL transformations, ensuring modular,reusable, and well-documented data transformation logic. Workflow Orchestration: Utilize Apache Airflow to orchestrate and automate data workflows,ensuring reliable data pipelines and scheduled jobs. Data Quality & Testing: Implement and maintain data validation checks and testing frameworksto ensure data integrity, accuracy, and compliance across all data pipelines. Collaboration: Work closely with data scientists, analysts, and product teams to understand dataneeds and provide technical solutions that meet business objectives. Performance Optimization: Tune SQL queries and manage the performance of Redshift clustersto ensure fast, efficient data access and analysis. Data Governance: Enforce data governance policies to ensure compliance with security, privacy,and data quality standards throughout the data lifecycle.Key Skills & Qualifications: Bachelors/Masters degree in Computer Science, Engineering, Data Science, or a related field. 3+ years of experience in data engineering with expertise in Amazon Redshift, Python, and AWS. Strong experience with Apache Airflow for workflow scheduling and orchestration. Hands-on experience with dbt (Data Build Tool) for managing SQL transformations and datamodels. Proficiency in API development and integration, including the use of RESTful APIs for dataingestion. Extensive experience with AWS services such as S3, Lambda, EC2, RDS, and CloudWatch.
 Expertise in data modeling concepts and designing efficient data structures (e.g., star schemas,snowflake schemas) in a data warehouse environment. Advanced knowledge of SQL for querying and optimizing large datasets in Redshift. Experience building ETL/ELT pipelines and integrating data from multiple sources, includingstructured and unstructured data. Familiarity with version control systems like Git and best practices for code management anddeployment automation. Knowledge of data governance principles, including data security, privacy, and quality control.Preferred Qualifications: Experience with real-time data processing tools such as Kafka or Kinesis. Familiarity with data visualization tools like Tableau, Looker, or Power BI. Knowledge of other data warehousing solutions like Snowflake or Google BigQuery. Experience with DevOps practices for managing infrastructure and CI/CD pipelines (Docker,Kubernetes). Understanding of machine learning pipelines and how data engineering supports AI/ML initiatives.Soft Skills: Strong analytical and problem-solving skills. Ability to work independently and as part of a cross-functional team. Strong written and verbal communication skills, with the ability to explain technical concepts tonon-technical stakeholders. Detail-oriented, proactive, and self-motivated with a focus on continuous improvement. Strong organizational and project management skills to handle multiple tasks


Job Type

Payroll


Categories

Data Engineer (Software and Web Development)

Cloud Architects (Software and Web Development)

DevOps Engineers (Software and Web Development)

Database Administrator (Software and Web Development)

Software Engineer (Software and Web Development)


Must have Skills

  • Python – 3 Years
  • Amazon Redshift – 3 Years
  • AWS – 3 Years
  • Apache Airflow – 1 Years
  • ETL(Extract, Transform, Load) – 1 Years
  • Kubernetes – 1 Years
Job Category: Information Technology
Job Type: Full Time
Job Location: Bengaluru
Job Salary: 12 - 50 Lakh/Year (Annual salary)

Apply for this position

Allowed Type(s): .pdf, .doc, .docx
Scroll to Top