Senior Big Data Engineer (Remote) Engineering - Boston, MA at Geebo

Senior Big Data Engineer (Remote)

Job DescriptionJob DescriptionParticular Details Job TitleBig Data EngineerPosition LevelSenior-Level/Experience ProfessionalIndustryInformation TechnologyTotal Position01Job TypeFull TimeCompany Alivia AnalyticsAbout the CompanyAlivia Analytics(TM) is following a mission to bend the healthcare cost curve by using analytic applications and data science to identify cases of fraud, waste, and abuse (FWA).
By turning mountains of data into actionable answers, Alivia Analytics(TM) does the heavy lifting - delivering the accuracy, confidence, and speed you need to solve the healthcare payment integrity challenges.
By putting powerful, easy-to-use, advanced technology into the hands of payment integrity business leaders and experts, Alivia Analytics(TM) empowers users to go beyond recovery to prevention and simulation within days.
Alivia takes pride in stating that Alivia is observing the lowest false positive rate in the industry.
Currently, approximately 8 trillion dollars are spent globally in healthcare annually from which FWA accounts for up to 10% of every dollar spent in the healthcare systems.
That equates to almost 800 billion dollars lost annually.
FWA is a growing problem - as providers find ways to evade traditional forms of rule-based fraud detection, our clients need to enhance their firepower with advanced analytic systems.
Our development team and data science team build applications using JavaScript that leverage algorithms built using Python, R, and SQL to identify these bad actors.
About the PositionWe have an opportunity for a highly motivated senior-level:
Data Engineer to join our rapidly growing team.
You will have broad opportunities to succeed and grow, both technically and non-technically.
Our development team is the keystone of our corporate structure, directly translating business problems into technical solutions for our global clients.
This makes communication skills as important as developing skills.
As a start-up, we want you to be able to grow with us.
You'll be able to learn from a management team with a combined 60
years of technical and medical expertise and a history of successful exits.
Our founder has developed 70
systems in his career, 2 of which are still number 1 in the world today.
About the Role We are seeking a highly skilled Cloud Data Engineer with at least 5 years of experience in designing, developing, documenting, and integrating applications using Big Data platforms like Snowflake, Databricks, Hadoop, and Hive.
The successful candidate will have expertise in deploying these pipelines to cloud infrastructure hosted in AWS or Azure.
Job Description/Key Areas of Responsibilities Gather requirements from business/user groups to analyze, design, develop, and implement data pipelines according to customer requirements.
Process data from Azure/AWS data storage using Databricks and Snowflake Optimize table design and indexing for end-user ease of use as well as workload performance.
Work with various input file formats including delimited text files, log files, Parquet files, JSON files, XML files, Excel files, and others.
Develop automated ETL procedures to load data from various sources into our application's data warehouse.
Ensure pipeline structure is standardized across different customers, each may have their own unique input data format.
Configure monitoring systems to detect failure and performance degradation of ETL pipelines.
Work with the DevOps team to design CI/CD pipelines to conduct ETL upgrades.
Deploy and leverage cloud infrastructure and services to assist in ETL pipeline definition and automation.
Understand data modeling (Dimensional and relational) concepts like Star-Schema Modeling, Schema Modeling, Fact, and Dimension tables.
Have strong knowledge of both SQL and No SQL databases.
Collaborate with business partners, operations, senior management, etc.
on day-to-day operational support.
Work with high volumes of data with stringent performance requirements Use programming languages like Python to clean raw data before processing (e.
g.
, removing newline characters/delimiters within fields) Define data quality and validation checks to preemptively detect potential issues.
Ensure ETL pipelines are HIPAA-compliant, run with minimal permissions, and securely manage any passwords and secrets used for authentication.
Document ETL pipeline logic, structure, and field lineage for review by both technical and non-technical audiences Expertise in processing data from Azure/AWS data storage using Databricks and Snowflake.
Key Technical Skills Set:
Data Modeling:
Understanding of data modeling concepts, including Star-Schema Modeling, Schema Modeling, Fact and Dimension tables.
Database Knowledge:
Strong knowledge of both SQL and NoSQL databases.
ETL Development:
Developing automated ETL procedures to load data from various sources into a data warehouse.
Ensuring ETL pipelines are HIPAA-compliant and securely manage authentication credentials.
File Formats:
Working with various input file formats, including delimited text files, log files, Parquet files, JSON files, XML files, Excel files, etc.
Performance Optimization:
Optimizing table design and indexing for end-user ease of use and workload performance.
Monitoring and CI/CD:
Configuring monitoring systems to detect pipeline failures and performance degradation.
Collaborating with the DevOps team to design CI/CD pipelines for ETL upgrades.
Data Quality and Validation:
Defining data quality and validation checks to preemptively detect potential issues in data.
Programming Languages:
Proficiency in programming languages like Python (for data cleaning and preprocessing), R, Java, and Scala.
Version Control:
Experience with version control systems like Git for managing code and configurations.
Problem-Solving:
Excellent problem-solving skills, including troubleshooting and resolving issues in data pipelines.
Documentation:
Documenting ETL pipeline logic, structure, and field lineage for technical and non-technical audiences.
Communication and Collaboration:
Strong communication and collaboration skills to work with stakeholders from different backgrounds and levels of expertise.
Data Engineer Requirements Bachelor's degree in computer science or a related field 5
years of experience in designing, developing, documenting, and integrating applications using Big Data platforms like Snowflake and Databricks Extensive experience working on both Azure and AWS, ideally using native ETL tooling (e.
g.
, Azure Data Factory) Strong experience in cleaning, pipelining, and analyzing large data sets.
Adept in programming languages like R, Python, Java, and Scala Experience with git for version control Excellent problem-solving skills and ability to work independently and as part of a team.
Strong communication and collaboration skills, with the ability to work with stakeholders from different backgrounds and levels of expertise.
Company DescriptionAlivia Analytics is helping customers Achieve Healthcare Payment Integrity, Finally.
By turning mountains of data into actionable answers, Alivia Analytics does the heavy lifting - delivering the accuracy, confidence, and speed our customers need to solve their healthcare payment integrity challenges.
Through the Alivia Analytics Healthcare Payment Integrity Suite TM we help private and public healthcare payers achieve payment integrity globally.
In the US alone, up to 10% of every dollar spent is attributed to Fraud, Waste, or Abuse which amounts to up to 370 Billion dollars lost annually.
If your ambition is to grow your responsibilities and career while building world-class analytic SaaS systems and fixing a huge problem for social good, please come and join us.
Company DescriptionAlivia Analytics is helping customers Achieve Healthcare Payment Integrity, Finally.
By turning mountains of data into actionable answers, Alivia Analytics does the heavy lifting - delivering the accuracy, confidence, and speed our customers need to solve their healthcare payment integrity challenges.
Through the Alivia Analytics Healthcare Payment Integrity Suite TM we help private and public healthcare payers achieve payment integrity globally.
In the US alone, up to 10% of every dollar spent is attributed to Fraud, Waste, or Abuse which amounts to up to 370 Billion dollars lost annually.
If your ambition is to grow your responsibilities and career while building world-class analytic SaaS systems and fixing a huge problem for social good, please come and join us.
Recommended Skills Apache Hadoop Apache Hive Big Data Communication Data Cleansing Data Modeling Estimated Salary: $20 to $28 per hour based on qualifications.

Don't Be a Victim of Fraud

  • Electronic Scams
  • Home-based jobs
  • Fake Rentals
  • Bad Buyers
  • Non-Existent Merchandise
  • Secondhand Items
  • More...

Don't Be Fooled

The fraudster will send a check to the victim who has accepted a job. The check can be for multiple reasons such as signing bonus, supplies, etc. The victim will be instructed to deposit the check and use the money for any of these reasons and then instructed to send the remaining funds to the fraudster. The check will bounce and the victim is left responsible.