BCBS-MI needs multiple roles to fill

We have the following below immediate opportunities with our direct client. Details are below. Please review and send any viable candidates for review. Thanks!

Project Details:

Work will be Remote till COVID-19 lets up and then will have some onsite ? all travel expenses covered

Client: BCBS of Michigan

Start Date: 5/18/20

Duration: 6 Months+

Notes: Candidates MUST HAVE previous healthcare experience

*************************************************************************************************************

Role #1 – Lead Developer ? Spark (Scala)

Job Title: Senior Spark Scala Developer

Prior Healthcare Payer Experience is preferred

Job Description

The Lead Spark Data Engineer will be responsible for designing, developing and deploying ETL data ingestion and data extraction pipelines using Scala & Spark to extract and transform data stored in a Hive Data Lake provisioned on Hadoop Ecosystem to a Distributed Redshift Data Lake for moving the data to a CPDM (Clinical Program Data Mart).

Both Relational and Non-Relational data will have to be extracted from the distributed clusters and the data will be flattened and transformed to extract to Redshift Data Lake.

Project: CPDM (Clinical Program Data Mart) is a vital enterprise data asset.

The Information Management and the Analytics & Insights teams are aligned to prioritize the migration of CPDM from the stand-alone SQL environment to the Informatics Environment (IE).

We will replicate all of the required data within IE.

This will be inclusive of obtaining data directly from EDW, FARE and other data sources as well as bringing in data from CPDM to complete the data sets.

Roles & Responsibilities:

? As a key member of the team, you will be responsible for leading the development activities for designing and developing the ETL Pipelines for Data Ingestion and Replication from Hive to Redshift Data Lake.

? Work on large scale distributed clusters spanning multiple data centers

? Help lead design discussions on architectural tradeoffs including Scalability, Resiliency, High Availability and Security

? Deploy and test out new technologies and additionally will help the platform engineering team in adopting new technologies with sound design, automation, and data-driven decisions.

? Perform capacity planning and solution review Required Experience

? Develop data ingestion and extraction pipelines using Spark Hive Data Lake provisioned on Hadoop Ecosystem

? Coding and testing complex system components

? Working with other teams to deliver and operate large scale, distributed services in the cloud

? Experience working with migrating large data volumes and building high load server-side software on Scala

? Actively participate in and help evolve all adopted Agile/Lean/Kanban practices such as: TDD (Test Driven Development), BDD (Behavior Driven Development), stand-ups, retrospectives, continuous delivery, and knowledge sharing across team members. ? Aggressively and proactively test software throughout the development cycle using a TDD mindset – including unit testing, functional testing, load testing and automation.

? Collaborate with Quality Assurance engineers in test case construction, issue identification, as well as remediation and automation.

? Continuously improve, refine, and refactor software.

Technical Requirements:

? A degree in Computer Science, Information Systems, Engineering, or Mathematics preferred.

? 7+ years of professional software development experience with data engineering and data warehouse projects hosted on-prem and on-cloud environments

? 3+ Years of Apache Spark programming experience using REQUIRED with Hadoop platform including Spark/Scala, SparkSQL, Hbase, Hive, Sqoop and HDFS in multi-tenant environments

? Must have hands on experience of Amazon Redshift Architecture AWS Redshift experience for DB design, data ingestion and writing advanced SQL Queries

? Expert extract/transform/load (ETL) developer in a Linux environment and able to write ETL code for Big Data

? Familiarity with HBase or other NoSQL technologies

? Familiarity with HiveQL to perform queries, summaries, and analysis of the stored data in Hive

? Familiarity with Storing and querying Structured and Non-Structured Data from Redshift.

? Prior Experience with development of Data Ingestion Pipeline and services on Hadoop Eco system.

? Familiarity with HBase or other NoSQL technologies

? Familiarity with HiveQL to perform queries, summaries, and analysis of the stored data in Hive

? Familiarity with Storing and querying Structured and Non-Structured Data from Redshift.

? Good knowledge on AWS environment and Service knowledge with S3 storage understanding

? Previous experience in database architecture, development, or other related fields

? Working knowledge of Hadoop-based infrastructure and security configuration

? Should be able to work independently within a project team with minimum supervision and mentor other developers when needed

? Understand, articulate, and conform to core functional and OOP programming principles.

? Strong analytical and critical thinking skills

? Experience in Agile software development.

? Ability to prioritize and multitask

? Deadline and detail-oriented

? Data modeling skills a plus

? Java Programming Experience is a plus

*************************************************************************************************************

ROLE# 2: Job Title: Data Analyst /Tester ? Big Data

Job Description

The Data Analyst should have extensive experience in Healthcare Data management, well-versed to extract, profile & analyze healthcare data to support corporate initiatives and assisting with data-intensive projects.

Should be able to interact with business and technical stakeholders and communicate findings using data visualization and detailed reports.

The candidate should have prior experience supporting data migration projects from a Data Lake on Big Data Ecosystem and should be able to help project teams with Data migration testing effort, perform data validations, verifying data flows between data end points, creating sample data & sharing data through secure tools, data mapping activities.

Must have a good knowledge on dealing with Relational and non-Relational datasets.

Project: CPDM (Clinical Program Data Mart) is a vital enterprise data asset.

The Information Management and the Analytics & Insights teams are aligned to prioritize the migration of CPDM from the stand-alone SQL environment to the Informatics Environment (IE).

We will replicate all of the required data within IE. This will be inclusive of obtaining data directly from EDW, FARE and other data sources as well as bringing in data from CPDM to complete the data sets.

Roles & Responsibilities:

? Participate in discovery of data entities, their relationships and source for extracting data.

? Participate in discovery and functional requirements documentation phase

? Extract data from source system and organize data into a File/Database Table for analysis.

? Develop and support Conceptual, Logic and Physical data models to support enterprise data warehouse and big data projects and initiatives.

? Familiarity with HBase or other NoSQL technologies

? Familiarity with Storing and querying Structured and Non-Structured Data from Amazon Redshift.

? Profiling and analysis of data resources to understand content, patterns, trends, etc.

? Understand the data flow mechanism and identify data sources and end points.

? Participate in Identifying data elements from source systems and mapping to target attributes

? Create reports and spreadsheets, displaying data in a logical and concise manner.

? Understand data storage and data sharing methods for secure transfer of data.

? Collaborate with report developers and other data consumers to support their data requirements

? Provide fundamental analytical support by developing and promoting consistent analytical methodologies.

Technical Requirements:

? Bachelor’s degree in Computer Science, Mathematics, Economics, or related field; or an equivalent combination of education and experience.

? 5+ years of experience with health data analysis, business analysis, software application, database design, and technical documentation experience.

? 2+ years of minimum experience in the healthcare insurance industry supporting one or more data management projects, includes, data analysis, data migration, data warehouse and data analytics.

? Minimum 1-year Experience working on Big Data / Data Lake solutions (Hadoop, AWS).

? Experience in implementing Data Warehouse solution in the Cloud (Azure, AWS, Google).

? Familiarity with HiveQL to perform queries, summaries, and analysis of the stored data in Hive

? Familiarity with Storing and querying Structured and Non-Structured Data from Amazon Redshift.

? Advanced analytical skills with the ability to interpret and synthesize complex data sets. Experience in advanced analytical techniques (R, Python) is a plus

? Should demonstrate writing advanced SQL Queries, analyzing data using Excel or other data analytics software tools.

? Should have proficiency with profiling structured and unstructured data using an industry standard data profiling tool, like Talend, Informatica, etc.?

? Knowledge of applied statistics, business analytics certification a plus.

? Good communicate on skills

? Good interpersonal skills in working with both onsite and virtual teams

? Prior Healthcare Payer Experience is preferred

*************************************************************************************************************

Role #3 Job Title: Big Data Spark Developer

?Job Description

The Big Data Spark Developer will be responsible for coding, testing and supporting ETL data ingestion and data extraction pipelines using Scala & Spark to extract and transform data stored in a Hive Data Lake provisioned on Hadoop Ecosystem to a Distributed Redshift Data Lake for moving the data to a CPDM (Clinical Program Data Mart. Both Relational and Non-Relational data will have to be extracted from the distributed clusters and the data will be flattened and transformed to extract to Redshift Data Lake.

Project: CPDM (Clinical Program Data Mart) is a vital enterprise data asset. The Information Management and the Analytics & Insights teams are aligned to prioritize the migration of CPDM from the stand-alone SQL environment to the Informatics Environment (IE). We will replicate all of the required data within IE. This will be inclusive of obtaining data directly from EDW, FARE and other data sources as well as bringing in data from CPDM to complete the data sets.

Roles & Responsibilities:

? As a member of the data engineering team, the Scala developer will be involved in the development, testing and deployment activities throughout the full life cycle of the project.

? The developer will work with the development lead and be responsible to code, unit test, code review, follow coding standards and coordinate resolution with other development team members

? Work on large scale distributed clusters spanning multiple data centers

? Develop data ingestion and extraction pipelines using Spark & Scala on Hive

? Coding and testing complex system components

? Working with other teams to deliver and operate large scale, distributed services in the cloud

? Experience working with migrating large data volumes and building high load server-side software on Scala

? Should be able to work independently within a project team to understand business needs and design, build and test integrations, transformations and processes that are production ready

Technical Requirements:

? A degree in Computer Science, Information Systems, Engineering, or Mathematics preferred.

? 5+ years of professional software development experience with data engineering and data warehouse projects hosted on-prem and on-cloud environments

? 2+ Years of Apache Spark programming experience using Scala & Spark REQUIRED with Hadoop platform including Spark/Scala, SparkSQL, Hbase, Hive, Sqoop and HDFS in multi-tenant environments

? Must have hands on experience of Amazon Redshift Architecture AWS Redshift experience for DB design, data ingestion and writing advanced SQL Queries

? Must have good understanding of ETL concepts and able to write ETL code for Big Data

? Familiarity with HBase or other NoSQL technologies

? Familiarity with HiveQL to perform queries, summaries, and analysis of the stored data in Hive

? Familiarity with Storing and querying Structured and Non-Structured Data from Redshift.

? Prior Experience with development of Data Ingestion Pipeline and services on Hadoop Eco system.

? Experience writing automated tests using tools like Junit, Scala Test, or Specs

? Working knowledge of Hadoop-based infrastructure and security configuration

? Good knowledge on AWS environment and Service knowledge with S3 storage understanding

? Should be able to work independently within a project team with minimum supervision

? Strong analytical and critical thinking skills

? Experience in Agile software development.

? Ability to prioritize and multitask

? Deadline and detail-oriented

? Java Programming Experience is a plus

? Prior Healthcare Payer Experience is preferred

Thanks & Regards,

Pushpendra Shukla

Technical Recruiter- Andeo Group LLC

EMAIL: pshukla@andeogroup.net

Hangout :- pshukla111994@gmail.com
O: 301-804-0303

Leave a Reply

Your email address will not be published. Required fields are marked *