- Company Name
- IQuest Solutions Corporation
- Job Title
- Data Scientist
- Job Description
-
**Job title:** Data Scientist
**Role Summary:** Design, develop, and deploy machine learning models across batch, streaming, and low‑latency environments. Own end‑to‑end ML pipelines, monitor model performance, and collaborate with cross‑functional stakeholders to drive data‑driven decision making.
**Expactations:** Deliver production‑ready models with robust monitoring, governance, and CI/CD. Mentor junior data scientists and maintain an ML platform stack aligned with enterprise standards.
**Key Responsibilities:**
- Build and tune predictive/prescriptive models (regression, classification, optimization) using Python, PySpark, XGBoost, LightGBM, PyTorch/TensorFlow.
- Design and implement ML pipelines (step functions, Airflow, Argo) and feature stores (Feast, Snowpark).
- Create automated workflows for model training, evaluation, deployment (batch, streaming, low‑latency endpoints).
- Implement model monitoring (drift, Shapley, performance metrics) and data validation with dashboards (Databricks, Palantir).
- Enforce governance: lineage, versioning, approvals, rollback, auditability, cost controls.
- Collaborate with stakeholders to translate business problems into data science solutions; maintain clear documentation and communication.
- Coach and mentor teammates on packaging, testing, and performance best practices.
**Required Skills:**
- 5+ years Python (pandas, PySpark, scikit‑learn); experience with PyTorch/TensorFlow.
- 5+ years ML tooling: SageMaker, Kubeflow, MLflow/Feast, Vertex, Databricks ML.
- 5+ years orchestration: Airflow, Step Functions, EventBridge/SQS/Kinesis.
- 3+ years cloud (AWS/Azure/GCP): ECR/ECS, Lambda, API Gateway, S3, Glue/Athena/EMR, RDS/Aurora, DynamoDB, CloudWatch, IAM, VPC, WAF.
- Snowflake: warehouses, schemas, Snowpark, UDF, RBAC.
- CI/CD: CodeBuild/CodePipeline or GitHub Actions/GitLab; blue‑green, canary, shadow deployments.
- Feature pipelines: batch/stream, parquet/iceberg, schema management, performance tuning.
- Testing & monitoring: unit/integration tests, contract tests, data drift, performance monitoring.
- Incident response: SLOs, dashboards, runbooks, cross‑layer debugging.
- Strong communication, collaborative mindset, automation bias.
**Required Education & Certifications:**
- Bachelor’s degree in Computer Science, Information Technology, Data Science, or related field.
- Machine learning certifications on AWS, GCP, or equivalent.
**Additional Qualifications:** Preferred experience in retail or manufacturing domains.