Data Developer Lead - Military veterans preferred

2024-04-19
AppCast (https://www.appcast.io)
Other

/yr

  full-time   employee


Newtown Square
Pennsylvania
19073
United States

Description:
Your career starts now. We are looking for the next generation of health care leaders.

At AmeriHealth Caritas, we are passionate about helping people get care, stay well and build healthy communities. As one of the nations leaders in health care solutions, we offer our associates the opportunity to impact the lives of millions of people through our national footprint of products, services and award-winning programs. AmeriHealth Caritas is seeking talented, passionate individuals to join our team. Together we can build healthier communities. If you want to make a difference, we would like to hear you.

Headquartered in Newtown Square, AmeriHealth Caritas is a mission-driven organization with more than 30 years of experience. We deliver comprehensive, outcomes-driven care to those who need it most. We offer integrated managed care products, pharmaceutical benefit management and specialty pharmacy services, behavioral health services, and other administrative services.

Discover more about us at www.amerihealthcaritas.com

DUTIES:

Analyze business requirements for the central cloud repository to store data. Architect data and design data patterns and data solutions on Azure Cloud, Azure Synapse and Databricks by using data modeling tools including Erwin. Design and build conceptual, logical, and physical data models in Erwin to solve business problems. Design and develop Azure Data Factory (ADF) pipelines to integrate data to Data Lake 2.0 by integrating data from multiple internal and external sources managed on Sybase, Oracle, SQL server and external files. Build Databricks notebooks using SQL, Python, Scala or PySpark programming languages for data transformation and loading in Azure Data Lake environment. Perform proof of concept on technical feasibility assessment for integrating new technical platform to Azure Data Lake platform. Design, develop, and test technical solutions and ensure they are aligned with Enterprise Data Strategy throughout Software Development Life Cycle (SDLC). Manage and analyze large volumes of unstructured data from internal and external sources using Hive, Impala, Oozie, SPARK (Scala), Sqoop, Flume, Hadoop API, and HDFS to optimize data loads and data transformations. Apply testing techniques, including unit, system, and regression testing, to verify deployed components work as designed. Deploy data science machine learning models to Azure Data Lake environment by using Machine Learning Operations (ML Ops) process. Design and implement data governance framework. Analyze and implement data quality rules for Data Lake tables to monitor different data quality dimensions and health of the data including the Informatica tool IDQ. Harvest and capture technical and business metadata for data cataloging using tools including EDC and Axon. Follow and prepare change control process as part of SDLC for production deployment. Provide first-time support for deployed components. Leverage and analyze EDWH data for regulatory reporting requirements and on-boarding data to Azure Data Lake 2.0 for advance analytical use cases. Remote option up to 100% of the time.

EDUCATION/REQUIREMENTS:

Bachelor’s degree (or foreign equivalent) in Information Systems Management or a related field, as well as the following experience which can be gained prior, during, or after Bachelor’s degree:

  • Five (5) years of experience architecting databases to build analytical solutions for business problems.
  • Three (3) years of experience designing and building conceptual, logical, and physical data models for data integration platforms.
  • Two (2) years of experience designing, building, and orchestrating ETL pipelines in Azure or other cloud- based systems.
  • Two (2) years of experience using Python and Spark programing languages to code Databricks notebooks per business requirements.
  • Two (2) years of experience implementing Machine Learning Modeling to model data science and monitor performance.
  •  And two (2) years of experience utilizing Hadoop (including Hive, Impala and Oozie) to analyze large volumes of unstructured data from internal and external sources.
  • In lieu of a Bachelor’s degree, employer will accept a Master’s degree (or foreign equivalent) in Information Systems Management or a related field plus 2 years of experience in the above which can be gained prior, during, or after Master’s degree.
  • Remote option up to 100% of the time.