Our story
We’re a fast growing, Series A stage health AI company building a clinical operating system for MSK practices. We are revenue generating and are looking to carefully expand the team to help us scale to 1,000,000 patients and beyond!
The team is headquartered in NYC with a satellite office in Vancouver, BC. Much of our engineering team is remote - distributed throughout the US and Canada.
We live in Notion docs, Slack channels and Github with weekly meetings over Zoom. Meetings may feature honorary advisors, MedTech execs, and, if we’re lucky, Will’s cat or our Chief Pup Officer.
As Flagler continues to grow, there is a unique opportunity to build the foundations of data and infrastructure to help the product and company reach our full potential. This is where you come in — to design and build reliable, trusted, and timely analytics that accelerate the decision-making process of key product and business functions. You will have a strong impact on the roadmap and growth trajectory of our company.
Key Responsibilities
Databricks Platform Expertise
• Develop, manage, and optimize data pipelines on the Databricks platform.
• Debug and troubleshoot Spark applications to ensure reliability and performance.
• Implement best practices for Spark compute and optimize workloads.
• Python Development:
• Write clean, efficient, and reusable Python code using object-oriented programming principles.
• Design and build APIs to support data integration and application needs.
• Develop scripts and tools to automate data processing and workflows.
MongoDB Management
• Integrate, query, and manage data within MongoDB.
• Ensure efficient storage and retrieval processes tailored to application requirements.
• Optimize MongoDB performance for large-scale data handling.
• Collaboration and Problem Solving:
• Work closely with data scientists, analysts, and other stakeholders to understand data needs and deliver solutions.
• Proactively identify and address technical challenges related to data processing and system design.
Required Qualifications
• Proven experience working with Databricks and Spark compute.
• Proficient in Python, including object-oriented programming and API development.
• Familiarity with NoSQL (MongoDB preferred), including querying, data modeling, and optimization.
• Strong problem-solving skills and ability to debug and optimize data processing tasks.
• Experience with large-scale data processing and distributed systems.
Preferred Qualifications
• Strong understanding of data architecture, ETL processes, and data warehousing concepts.
• Knowledge of other big data technologies like Delta Lake, Hadoop, or Kafka.
• Experience with cloud platforms (e.g., AWS, Azure, or GCP).
• Familiarity with CI/CD pipelines and version control systems like Git.
Hiring Process
Due to the high volume of applications, we only reach out to candidates selected for interviews. We do not have online leetcode assessments as an initial filter, so we only reach out to very few candidates for initial introduction.