Epicareer Might not Working Properly
Learn More

Data Scientist with LLM exp @ Onsite (Only w2)

Salary undisclosed

Checking job availability...

Original
Simplified

Data Scientist with LLM exp

Location: Charlotte, NC

Duration: 12 Months+

This role is restricted to W2 candidates, with no sponsorship offered.

Key Responsibilities:

  • Optimize and deploy AI models on GPU clusters for enhanced performance and scalability.
  • Leverage multi-GPU training and distributed computing frameworks such as TensorFlow Distributed, PyTorch Distributed, and Horovod to accelerate AI/ML workloads.
  • Configure and manage NVIDIA GPU and Google Cloud Platform (GCP) resources, including TPUs and GPU instances.
  • Develop APIs and design cloud-native architectures to support AI/ML applications.
  • Work with Generative AI frameworks such as LLaMA and Mistral to drive innovation in AI solutions.
  • Utilize FastAPI, Unicorn, and Swagger to create efficient and scalable API solutions.
  • Implement solutions using Python, Apache Spark (PySpark), Kubernetes, and Django to ensure robust and scalable applications.
  • Employ Apache Kafka for real-time data streaming and ensure seamless integration with distributed computing frameworks.

Must-Have Qualifications:

  • Proven experience in optimizing and deploying AI models on GPU clusters.
  • Strong expertise in multi-GPU training and distributed computing frameworks, including TensorFlow Distributed, PyTorch Distributed, and Horovod.
  • Proficiency in configuring and managing NVIDIA GPU resources and GCP instances.
  • Knowledge of API development and cloud-native architectures.
  • Experience with Generative AI frameworks such as LLaMA and Mistral.
  • Proficient in FastAPI, Unicorn, and Swagger for API development.
  • Skilled in Python, Apache Spark (PySpark), Kubernetes, and Django.
  • Experience with Apache Kafka for real-time data streaming and integration with distributed systems.

Data Scientist with LLM exp

Location: Charlotte, NC

Duration: 12 Months+

This role is restricted to W2 candidates, with no sponsorship offered.

Key Responsibilities:

  • Optimize and deploy AI models on GPU clusters for enhanced performance and scalability.
  • Leverage multi-GPU training and distributed computing frameworks such as TensorFlow Distributed, PyTorch Distributed, and Horovod to accelerate AI/ML workloads.
  • Configure and manage NVIDIA GPU and Google Cloud Platform (GCP) resources, including TPUs and GPU instances.
  • Develop APIs and design cloud-native architectures to support AI/ML applications.
  • Work with Generative AI frameworks such as LLaMA and Mistral to drive innovation in AI solutions.
  • Utilize FastAPI, Unicorn, and Swagger to create efficient and scalable API solutions.
  • Implement solutions using Python, Apache Spark (PySpark), Kubernetes, and Django to ensure robust and scalable applications.
  • Employ Apache Kafka for real-time data streaming and ensure seamless integration with distributed computing frameworks.

Must-Have Qualifications:

  • Proven experience in optimizing and deploying AI models on GPU clusters.
  • Strong expertise in multi-GPU training and distributed computing frameworks, including TensorFlow Distributed, PyTorch Distributed, and Horovod.
  • Proficiency in configuring and managing NVIDIA GPU resources and GCP instances.
  • Knowledge of API development and cloud-native architectures.
  • Experience with Generative AI frameworks such as LLaMA and Mistral.
  • Proficient in FastAPI, Unicorn, and Swagger for API development.
  • Skilled in Python, Apache Spark (PySpark), Kubernetes, and Django.
  • Experience with Apache Kafka for real-time data streaming and integration with distributed systems.