Listing Description
Job Description
Mandiant Threat Intel Product & Engineering is setting a new standard and changing how organizations can identify and confirm relevant cyber security threats. We develop solutions that enable our customers to know more about their adversaries than anyone else by providing access to a wide array of data and indicators as active threats emerge. We do this by leveraging optimized breach intelligence data, cultivated from the frontlines of cyber security—this includes more than 300 intelligence analysts and researchers in 26 countries and over 200,000 hours responding to breaches, in the last year alone.
Data science infrastructure is a key component that enables data science teams to easily discover, analyze, and consume data from various source and formats, build models and model training pipelines. This role will support data science teams, build data pipelines and data catalogs, and deploy and manage tools like JupyterLab, MLFLow in k8s. The main focus will be to improve current workflows, provide support for adoption of new or improved tooling and simplify integration and discovery for various datasets across the company.
What You Will Do:
- Deploy and maintain tools for data science teams
- Build data processing, data migration and E(T)L pipelines
- Build a dataset catalog and data discovery capabilities
- Improve existing tools in the data science workflow
- Produce and maintain technical documentation.
Qualifications
Requirements:
- Strong experience in cloud infrastructure engineering
- Strong experience in data science team support
- Experience with tools like JupyterLab, Airflow, MLFLow, etc
- Experience with Apache Spark or Apache Flink
- Experience with cloud services and cloud platform landscape (AWS, GCP)
- Experience with one or more programming languages (Java, Python, Go)
- Ability to articulate, verbally and in writing, abstract concepts as well as detailed technical information.
- Excellent interpersonal and communication skills and the ability to work both independently and with a team.
Additional Qualifications:
- Experience with Trino, BigQuery, Dataproc is a plus
- Experience with Hadoop, Hive is a plus
- Familiarity or experience working with or deploying on Kubernetes is a plus.
- Familiarity or experience working with container build and deploy tools (Drone, Argo, Harbor) is a plus.
Listing Details
- Citizenship: Not Provided
- Incentives: Not Provided
- Education: Not Provided
- Travel: Not Provided
- Telework: Full Telecommute