Synechron Logo

Synechron

Data & AI Engineer | Cloud-Native Pipelines, Python, Pandas, Large-Scale Data Processing, Cloud (AWS/Azure/GCP), Machine Learning Frameworks

Reposted 17 Days Ago
In-Office or Remote
Hiring Remotely in Hinjawadi, Pune, Mahārāshtra
Senior level
In-Office or Remote
Hiring Remotely in Hinjawadi, Pune, Mahārāshtra
Senior level
Lead the design and development of scalable data pipelines using Ab-initio and Python, migrating workflows, ensuring data quality, and mentoring team members.
The summary above was generated by AI

Job Summary
Synechron is seeking an experienced Lead Python Data & AI Engineer to architect, develop, and optimize large-scale data pipelines and AI solutions. This role involves utilizing advanced Python, multi-threading, and data processing tools to support enterprise data science and machine learning initiatives. The successful candidate will collaborate across cross-functional teams to deliver scalable, secure data systems, integrating AI/ML workflows supported by containerization and cloud deployment, driving organizational innovation and operational excellence.

Software Requirements

  • Required:

    • Extensive hands-on experience with Python (latest stable version recommended), including proficiency in multi-threading and object-oriented programming (OOP) design patterns

    • Practical knowledge of data processing libraries such as Pandas and familiarity with data storage formats such as Parquet and Delta Lake

    • Experience developing and maintaining data pipelines on cloud platforms (Azure, AWS, GCP) supporting AI/ML workflows

    • Experience with databases such as Oracle and ORM frameworks (SQLAlchemy, Django ORM) for database interaction and management

    • Proficiency with containerization tools like Docker and orchestration platforms such as Kubernetes for deployment

    • Strong understanding of CI/CD practices, automated testing (pytest), and version control (Git/GitHub) workflows

    • Knowledge of AI engineering tools such as GitHub Copilot, code assistants, or similar development aids

  • Preferred:

    • Experience with cloud-native data services, data governance practices, and scalable architecture design

    • Knowledge of distributed systems, data streaming, and real-time processing frameworks

    • Familiarity with ML frameworks like TensorFlow, PyTorch, for integrating AI models into pipelines

Overall Responsibilities

  • Design, build, and optimize scalable data pipelines using Python, Pandas, and cloud services for enterprise analytics and AI deployment

  • Develop and maintain data workflows for large datasets, supporting data cleansing, feature engineering, and model inference processes

  • Collaborate with data scientists, AI architects, and platform teams to facilitate seamless data and model integration

  • Implement automation for data ingestion, processing, and system deployment using CI/CD pipelines and containerization

  • Troubleshoot and resolve performance bottlenecks, optimize data storage/processing, and enforce data security standards

  • Create technical documentation and design standards supporting data and AI workflows

  • Lead initiatives on data governance, quality, and compliance across data pipelines and models

  • Support innovation by evaluating new tools, cloud features, and AI techniques for organizational benefit

Technical Skills (By Category)

  • Programming Languages:
    Required: Python (advanced), multi-threading, OOP design patterns
    Preferred: Additional languages such as Java or Scala for performance-critical components

  • Data Management & Storage:
    Pandas, NumPy, Parquet, Delta Lake, SQL (Oracle, MySQL), ORM frameworks (SQLAlchemy, Django ORM)

  • Cloud Technologies:
    AWS, Azure, or GCP cloud platforms supporting data pipelines, AI workflows, and storage

  • Frameworks & Libraries:
    Pandas, TensorFlow, PyTorch, Spark (PySpark), ML frameworks (preferred for AI model deployment)

  • Data Orchestration & Automation:
    Airflow, Terraform, Docker, Kubernetes, CI/CD tools (Jenkins, GitHub Actions)

  • Security & Governance:
    Data encryption, access controls, compliance with enterprise security standards

Experience Requirements

  • Minimum of 6 years of practical experience supporting data engineering, AI workflows, or large-scale data pipelines

  • Proven expertise in Python automation, multi-threaded data processing, and cloud-based data solutions

  • Strong experience deploying AI/ML models within scalable pipelines supported by containerized and cloud environments

  • Demonstrated success working in agile teams supporting enterprise and data science initiatives

  • Industry experience in finance, healthcare, or large enterprise data environments is preferred but not mandatory

Day-to-Day Activities

  • Develop, deploy, and optimize large-scale data pipelines supporting enterprise analytics and AI workloads

  • Implement batch and streaming data workflows, feature engineering, and model inference pipelines

  • Collaborate with data scientists, ML engineers, and platform teams for seamless data integration and deployment

  • Automate data workflows, infrastructure provisioning, and model deployment using DevOps practices

  • Troubleshoot pipeline performance issues, optimize storage and processing, and ensure data governance compliance

  • Document data architecture, model workflows, procedures, and best practices

  • Evaluate emerging data/AI tools, integrate new frameworks, and contribute to innovation initiatives

Qualifications

  • Bachelor’s or Master’s degree in Data Science, Computer Science, or related fields

  • 6+ years supporting data pipelines, AI models, and large-scale data environments on cloud platforms

  • Certifications in cloud (AWS, Azure, GCP), data engineering, or ML frameworks are advantageous

  • Demonstrated expertise in Python, Pandas, data processing, and cloud-native data architecture support

Professional Competencies

  • Analytical mindset with strong problem-solving skills for complex data and AI system issues

  • Effective communication skills for cross-team collaboration and stakeholder engagement

  • Leadership qualities to mentor junior team members and promote best practices

  • Strategic thinking regarding data governance, security, and scalable architecture

  • Adaptability to new AI/ML tools, cloud features, and data management trends

  • Time management skills to prioritize tasks and deliver impactful solutions within deadlines

S​YNECHRON’S DIVERSITY & INCLUSION STATEMENT
 

Diversity & Inclusion are fundamental to our culture, and Synechron is proud to be an equal opportunity workplace and is an affirmative action employer. Our Diversity, Equity, and Inclusion (DEI) initiative ‘Same Difference’ is committed to fostering an inclusive culture – promoting equality, diversity and an environment that is respectful to all. We strongly believe that a diverse workforce helps build stronger, successful businesses as a global company. We encourage applicants from across diverse backgrounds, race, ethnicities, religion, age, marital status, gender, sexual orientations, or disabilities to apply. We empower our global workforce by offering flexible workplace arrangements, mentoring, internal mobility, learning and development programs, and more.

All employment decisions at Synechron are based on business needs, job requirements and individual qualifications, without regard to the applicant’s gender, gender identity, sexual orientation, race, ethnicity, disabled or veteran status, or any other characteristic protected by law.

Candidate Application Notice

Similar Jobs

2 Hours Ago
Remote or Hybrid
Senior level
Senior level
Fintech • Professional Services • Consulting • Energy • Financial Services • Cybersecurity • Generative AI
The role involves managing cross-program dependencies, optimizing processes, conducting quality assurance, and enhancing resource allocation strategies for financial transformation projects.
Top Skills: Azure DevopsClarityMS OfficeTableau
4 Hours Ago
Easy Apply
Remote
Easy Apply
Senior level
Senior level
Artificial Intelligence • Consumer Web • Digital Media • Information Technology • Social Impact • Software
As a Senior Quality Platform Engineer, you will develop and maintain quality infrastructure, improve developer experience, and implement quality engineering practices to ensure scalable, efficient testing workflows.
Top Skills: AWSAzureCircleCICypressDockerGCPGithub ActionsGitlabJavaJavaScriptJestJunitKubernetesPlaywrightPythonRubyTypescript
4 Hours Ago
Remote or Hybrid
Expert/Leader
Expert/Leader
Fintech • Professional Services • Consulting • Energy • Financial Services • Cybersecurity • Generative AI
Seeking an experienced Full Stack Developer with 8+ years expertise in backend and React UI, overseeing high-quality development, collaborating on modern web solutions, and driving performance optimization.
Top Skills: Context ApiDockerGitGraphQLHTML5JavaJavaScriptJestKubernetesMochaNode.jsPythonReactReduxRestful ApisShellTypescript

What you need to know about the Melbourne Tech Scene

Home to 650 biotech companies, 10 major research institutes and nine universities, Melbourne is among one of the top cities for biotech. In fact, some of the greatest medical advancements were conceptualized and developed here, including Symex Lab's "lab-on-a-chip" solution that monitors hormones to predict ovulation for conception, and Denteric's vaccine for periodontal gum disease. Yet, the thousands of people working in the city's healthtech sector are just getting started, to say nothing of the tech advancements across all other sectors.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account