Data Engineer InterOp COE
Abacus Insights
Abacus Insights is a mission-driven, growth-stage technology company focused on transforming the healthcare industry, ultimately creating a more personalized patient experience, improving health outcomes, and lowering the overall cost of healthcare. Abacus Insights provides a flexible, efficient, and secure platform that organizes and exchanges healthcare data from various sources and formats. Our platform allows our customers to uncover differentiated insights that address their client's needs. In addition, our employees know they play an active role in keeping our customers' data safe and are responsible for ensuring that our comprehensive policies and practices are met.
With our deep expertise in cloud-enabled technologies and knowledge of the healthcare industry, we have built an innovative data integration and management platform that allows healthcare payers access to historically siloed and inaccessible data. Through our platform, these healthcare organizations can ingest and manage all the data they need to transform their business by supporting their analytical, operational, and financial needs.
Since our founding in 2017, Abacus has built a highly successful SaaS business, raising $100 Million from leading VC firms with deep expertise in the healthcare and technology industries. We are solving problems of massive scale and complexity in a sector ready for disruption. We're growing quickly and would love for you to be a part of it!
About the Role:
** This position is based in Pune India.**Come join our team! Help us tackle the data usability challenge for payers. Your expertise and experience will help drive meaningful performance outcomes. You'll also have the chance to advance your career, acquire new skills, and collaborate with some of the most innovative minds in payer data management.
We are seeking an experienced Senior Data Engineer to join our Connector Factory team. This role offers an opportunity to be a key contributor in a critical feature delivery team, where your expertise will guide the evolution of our data pipeline infrastructure. Our team is responsible for the development and operation of data pipelines that handle diverse data sources through both large batch and streaming systems. You'll work extensively with AWS services and play a crucial role in driving the growth and innovation of our platform.
You will:
- Design and Develop Data Systems: Architect, build, and maintain data pipelines and ETL processes utilizing tools such as Databricks, Snowflake, SQL, and PySpark.
- Enhance Data Quality: Play a pivotal role in creating and optimizing data assets to uphold high standards of data quality, performance, and reliability.
- Manage Data Pipelines: Actively monitor and troubleshoot data pipelines to ensure efficient and uninterrupted data distribution.
- Collaborate Across Teams: Partner with the Connector Factory team and cross-functional teams to understand client data requirements and transform these into scalable data solutions.
- Implement Agile Practices: Apply Agile methodologies and best practices to drive incremental improvements and adapt to emerging requirements.
- Communicate Effectively: Keep communication channels open with stakeholders to gather and clarify requirements and provide regular updates on project progress.
- Ensure Data Security: Stay committed to data privacy, security, and regulatory compliance, particularly given the sensitive nature of healthcare data.
What we're looking for:
- Educational Background: Bachelor’s degree in computer science, Engineering, or a related field. Advanced degrees are a plus.
- Extensive Experience: A minimum of 5 years of experience in data engineering and big data architecture.
- Technical Expertise: Deep knowledge of designing and maintaining big data architectures, including data lakes, columnar databases, large batch processing (Spark), and stream processing (Kafka).
- Cloud Proficiency: Strong experience with AWS data services and building scalable, distributed systems on cloud platforms.
- Programming Skills: Proficiency in Python or other object-oriented programming languages.
- Data Analysis Skills: Hands-on experience with data analysis and modeling of large data sets.
- Project Management: Strong organizational skills and experience managing complex projects.
- Root Cause Analysis: Proven ability to perform root cause analysis on data processes to improve efficiency and resolve business queries.
- Adaptability: A willingness to learn about new technologies and adapt to changing environments.
- Independent and Collaborative Work: Ability to self-direct tasks and effectively collaborate within a technical team.
- Infrastructure Automation: Familiarity with tools such as Terraform and GitLab CI/CD for infrastructure automation.
- Business Acumen: Comfort with ambiguity and a keen interest in solving business-related problems.
- Agile Experience: Background working in an Agile delivery framework.
Bonus Points:
- Relevant certifications in data engineering, cloud computing, or specific technologies such as Databricks, Snowflake, or AWS.
Equal Opportunity Employer
As a mission-led technology company that is helping to drive better healthcare outcomes, Abacus Insights believes that the best innovation and value we can bring to our customers comes from diverse ideas, thoughts, experiences, and perspectives. Therefore, we proudly dedicate extensive time and resources to building diverse teams and providing equal employment opportunities to all applicants. Abacus prohibits discrimination and harassment of any type regarding race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state, or local laws.