Overview
Edifecs/Cotiviti/Cotiviti is seeking a Data Integration Engineer to join our innovative software teams.In this position, you will be responsible for onboarding customers to the Risk Adjustment workflow applications. As part of this role, you will work with platform engineering, product, and implementation teams. The ideal candidate has experience in building scalable data pipelines that enable workflow applications, reporting, and customer confidence. You must have strong, hands-on technical expertise in big data technologies and communication abilities with non-tech and technical audiences.
Responsibilities
- Design and develop data flows and extraction
- Integrate client’s data into the Edifecs/Cotiviti/Cotiviti Risk Adjustment
- Work with relational databases and Python
- Implement open-source standards for data
- Develop a new framework for data ELT jobs to scale implementation, monitoring, and
- Build and scale automation that orchestrate complex
- Ability to support existing processes while leading efforts to redefine the data
Qualifications
- Bachelor’s degree (or foreign equivalent) in Computer Science, Computer Information Systems, Computer Engineering, or a related field.
- Minimum 5 years of progressive, post-baccalaureate experience in product support within the healthcare industry, including:
- 5 years of experience reading and writing relational database queries using T-SQL or ANSI-SQL.
- 5 years of experience working with Linux platforms such as RedHat, CentOS, SUSE/SLES, Ubuntu, etc.
- 5 years of strong analytical and troubleshooting experience.
- 4 years of experience modifying and supporting code or scripting development using Bash/Python.
- Experience managing and troubleshooting Linux servers in production.
- Working knowledge of source control systems (Git preferred).
- Familiarity with CI/CD pipelines and Git integration.
- Ability to maintain Git repositories, branching strategies, and version control best practices.
- Collaborate with cross-functional teams to deploy, configure, and troubleshoot applications in a Linux environment.
- Implement and maintain automation scripts to improve efficiency and scalability.
- Monitor system performance and proactively address issues to prevent downtime.
- Plan and execute system upgrades, patches, and migrations.
- Experience designing and developing data flows and extraction processes.
- Experience integrating client data into workflow applications.
- Experience with open-source data quality tools (e.g., Great Expectations).
- Experience developing scalable ELT pipelines using tools such as Argo Workflows, cron, Airflow, dbt.
- Experience building reusable data methods and automation for complex workflows.
- Experience supporting existing processes while contributing to data strategy improvements.
- Plus: Experience with container deployment platforms and tools such as Kubernetes, Docker, Helm, and Terraform.
- Plus: AWS Cloud experience with services like EC2, RDS, SQS, IAM, and S3.
- Excellent spoken and written English communication skills.
- Strong customer interaction and problem-solving abilities.
- Customer-focused with a proactive and responsive approach.
- Ability to diagnose customer-reported problems and recommend solutions within agreed SLAs.
- Creative thinking to understand complex problems and communicate them to non-technical audiences.
- Willingness to participate in on-call rotations for critical system issues.
- Ability to document system configurations, operational procedures, and troubleshooting guides.
Sorry the Share function is not working properly at this moment. Please refresh the page and try again later.
Share on your newsfeed