Norton_AI/MLEngineer

LS Digital Group

Job Summary

The AI/ML Engineer will design, build, and deploy supervised, unsupervised, and deep learning models, owning end-to-end NLP workflows from pre-processing to fine-tuning. This role requires hands-on experience with Sentence Transformers for semantic embeddings and integrating FAISS or Qdrant for scalable vector similarity search. The engineer will architect models for NER, relation extraction, intent classification, and LLM-based generation, integrating them with production-grade APIs. Expertise in Linux/Windows ML environments, MLOps best practices, and model optimization is essential, along with deploying and managing models on cloud platforms.

Must Have

  • Design, build, and deploy supervised, unsupervised, and deep learning models.
  • Own end-to-end NLP workflows: pre-processing, training, evaluation, fine-tuning.
  • Hands-on experience in building and deploying models using Sentence Transformers for semantic embeddings and integrating FAISS or Qdrant for scalable vector similarity search and retrieval.
  • Strong understanding of handling limited data scenarios using data augmentation strategies and Monte Carlo simulation techniques to improve model robustness and accuracy.
  • Architect models for NER, relation extraction, intent classification, and LLM-based generation like Mistral, LLama etc.
  • Integrate ML models with production-grade APIs and Flask/FastAPI endpoints.
  • Expertise in Linux/Windows-based ML environments and managing dependencies and virtual environments (venv, conda, pip).
  • Work with MLOps best practices: model versioning, reproducibility, monitoring, and rollback strategies.
  • Optimize models for speed/memory, leveraging techniques like quantization, ONNX, or TorchScript.
  • Deploy and manage machine learning models on cloud platforms using services like ML pipelines, object storage, serverless functions, and container orchestration tools.

Good to Have

  • Understanding of Azure Pipelines, Docker, Kubernetes (AKS) if required.
  • Understanding of GPU/CPU resource allocation, model parallelization, and scaling.

Job Description

Role and Responsibility Details:

  • Design, build, and deploy supervised, unsupervised, and deep learning models.
  • Own end-to-end NLP workflows: pre-processing, training, evaluation, fine-tuning.
  • Hands-on experience in building and deploying models using Sentence Transformers for semantic embeddings and integrating FAISS or Qdrant for scalable vector similarity search and retrieval.
  • Strong understanding of handling limited data scenarios using data augmentation strategies and Monte Carlo simulation techniques to improve model robustness and accuracy.
  • Architect models for NER, relation extraction, intent classification, and LLM-based generation like Mistral, LLama etc.
  • Integrate ML models with production-grade APIs and Flask/FastAPI endpoints.
  • Expertise in Linux/Windows-based ML environments and managing dependencies and virtual environments (venv, conda, pip).
  • Work with MLOps best practices: model versioning, reproducibility, monitoring, and rollback strategies.
  • Optimize models for speed/memory, leveraging techniques like quantization, ONNX, or TorchScript.
  • Deploy and manage machine learning models on cloud platforms using services like ML pipelines, object storage, serverless functions, and container orchestration tools.
  • Understanding of Azure Pipelines, Docker, Kubernetes (AKS) if required.
  • Understanding of GPU/CPU resource allocation, model parallelization, and scaling.

Job Description:

  • Design, build, and deploy supervised, unsupervised, and deep learning models.
  • Own end-to-end NLP workflows: pre-processing, training, evaluation, fine-tuning.
  • Hands-on experience in building and deploying models using Sentence Transformers for semantic embeddings and integrating FAISS or Qdrant for scalable vector similarity search and retrieval.
  • Strong understanding of handling limited data scenarios using data augmentation strategies and Monte Carlo simulation techniques to improve model robustness and accuracy.
  • Architect models for NER, relation extraction, intent classification, and LLM-based generation like Mistral, LLama etc.
  • Integrate ML models with production-grade APIs and Flask/FastAPI endpoints.
  • Expertise in Linux/Windows-based ML environments and managing dependencies and virtual environments (venv, conda, pip).
  • Work with MLOps best practices: model versioning, reproducibility, monitoring, and rollback strategies.
  • Optimize models for speed/memory, leveraging techniques like quantization, ONNX, or TorchScript.
  • Deploy and manage machine learning models on cloud platforms using services like ML pipelines, object storage, serverless functions, and container orchestration tools.
  • Understanding of Azure Pipelines, Docker, Kubernetes (AKS) if required.
  • Understanding of GPU/CPU resource allocation, model parallelization, and scaling.

Qualifications:

MTech, B.Tech, B.E

11 Skills Required For This Role

Resource Allocation Game Texts Resource Planning Linux Azure Fastapi Deep Learning Docker Flask Kubernetes Machine Learning

Similar Jobs