About the Role: Our client is dedicated to fortifying the defenses of enterprises worldwide against a myriad of email and collaboration application cyber threats. The mission entails developing a suite of cutting-edge products to empower clients in visualizing, controlling, and combating cyber security attacks effectively.
This role presents a unique opportunity to make significant contributions to our client's mission of delivering unparalleled detection capabilities globally. The primary focus is on equipping the detection team with advanced tools and high-quality data to facilitate easy assessment and analysis of misclassifications. The strategic vision involves developing crucial detection tools, enhancing misclassification comprehension, and improving communication with clients.
Responsibilities: Enhance the efficiency and quality of detection systems development.Implement pipelines to process and respond to customer feedback promptly.Provide detection teams with precise, high-quality data for evaluating and training machine learning models.Automate processes to streamline customer responses.Design, develop, modify, and test systems to improve data quality and comprehension.Collaborate with Technical Program Managers, Product Managers, Data Engineers, Data Scientists, and operational and engineering teams to implement and iterate on product development.Exercise sound judgment in selecting methods and techniques for problem-solving.Write code with emphasis on testability, readability, handling edge cases, and error management.Prepare and review technical design documents. Requirements: Proficiency in Python, Databricks, Airflow, Django, and Postgres.Demonstrated experience in building data pipelines using PySpark.Familiarity with large-scale solutions or environments involving complex integrations, demanding latency requirements, or significant throughput challenges.Proficient in performance debugging and benchmarking to ensure efficient application operation.Production level experience with technologies such as PySpark, Data platform and Data coordination, Hadoop, Hive, and data processing frameworks.Ability to translate business requirements into detailed software requirements and effectively communicate system design to technical and non-technical stakeholders.Experience in identifying, analyzing, and resolving complex technical issues, demonstrating a methodical approach to troubleshooting and problem-solving.Proven track record of effective collaboration with cross-functional teams and diverse stakeholders. Note: Certain roles may be eligible for bonuses, restricted stock units (RSUs), and benefits. Compensation packages are tailored to individual candidates based on skills, experience, qualifications, and other job-related factors. Benefits are a significant part of the total compensation package. For more details, refer to our Compensation and Equity Philosophy on our Benefits & Perks page.
Base salary range: $175,800—$206,800 USD