Skip to main content
I

Software Engineer

InfoCepts

Location

Coopersburg, PA

Salary

Not specified

Type

fulltime

Posted

Today

via linkedin

Job Description

Position Title: DevOps Engineer

Work Location: Coopersburg, PA

Type of Employment: Full time

Role Overview:

We are seeking a hybrid DevOps \+ Data Engineering profile, with a stronger emphasis on DevOps capabilities. The role owner will be required to modify \& manage ETL pipelines as part of platform operations. The role owner should have hands on DevOps experience along with good understanding of data pipelines.

Key Result Areas and Activities:

DevOps requirements:

  • Modify and manage data pipelines in DevOps setup.
  • Good understating of CI/CD pipelines.

Development responsibilities:

Software Asset Maintenance \& Upgrades

  • Upgrade and maintain third-party components such as ingress controllers, service meshes, monitoring agents, infrastructure libraries, and cloud-native tools.
  • Apply the latest versions and security patches to ensure compliance, stability, and performance.

Infrastructure as Code (IaC) Enhancements

  • Update and enhance IaC scripts to support version upgrades across development, QA, and production environments.
  • Validate changes through sandbox testing before deployment to production.

Compatibility \& Dependency Management

  • Ensure upgraded components remain compatible with dependent services and applications.
  • Identify and mitigate potential breaking changes or dependency conflicts.

Application Code Adjustments

  • Implement necessary code changes in supported languages (e.g., Python) to accommodate new versions or configuration requirements.
  • Address minor and moderate changes required for compatibility with upgraded components.
  • Update existing unit tests in response to the application code changes.

Security \& Compliance

  • Apply immediate fixes for vulnerabilities.
  • Maintain adherence to organizational security and governance guidelines.

Testing \& Validation

  • Create and execute test strategies for validating the upgrades.
  • Execute existing unit tests and manual test cases post-upgrade.
  • Conduct functional testing of impacted applications to ensure end-to-end stability.
  • Validate application behavior after code changes and infrastructure updates.

Reporting \& Governance

  • Provide weekly status reports detailing software versions, security posture, upgrade activities, and testing outcomes.
  • Participate in regular reviews and acceptance processes.

Must-Have Skill Set:

  • Terraform.
  • CI/CD Pipelines.
  • Platform upgrades and maintenance.
  • QA exposure – integration \& platform testing.
  • Deep understanding of cloud data services (AWS) and migration strategies.
  • Strong proficiency in ETL/ELT pipelines and framework development using Python. Development may not be needed but understanding is required.
  • Modifying \& monitoring data pipelines for batch and real-time processing.
  • Exceptional communication skills for engaging executives and non-technical stakeholders.
  • Knowledge of containerization (Docker, Kubernetes) and orchestration for data workloads.

Good-to-Have Skill Set

  • Certifications in cloud platforms (AWS) and data engineering.
  • Experience with advanced analytics and machine learning pipelines.
  • Prior consulting experience or leading large-scale data transformation programs.
  • Knowledge of data extraction from SAP OData Services.
  • Experience with multiple relational and NoSQL databases (Must have - Redshift, RDS and Athena).
  • Experience with BI tools integration with enterprise data platforms.

Qualification:

  • Bachelor’s degree in computer science, engineering, or related field (master's degree is a plus).
  • Demonstrated continued learning through one or more technical certifications or related methods.
  • At least 10\+ years of relevant experience; two years may be substituted for a master’s degree.

Looking for more opportunities?

Browse thousands of graduate jobs and entry-level positions.

Browse All Jobs