AI Research Engineer (Model Serving & Inference)

29 Days Old

Join Tether and Shape the Future of Digital Finance

At Tether, we're pioneering a global financial revolution with innovative blockchain solutions that enable seamless digital token transactions worldwide. Our products include the trusted stablecoin USDT, energy-efficient Bitcoin mining solutions, advanced data sharing apps like KEET, and educational initiatives to democratize digital knowledge.

Why join us? Our remote, global team is passionate about fintech innovation. We seek individuals with excellent English communication skills eager to contribute to cutting-edge projects in a fast-growing industry.

About the job:

As part of our AI model team, you will innovate in model serving and inference architectures for advanced AI systems. Your focus will be on optimizing deployment strategies to ensure high responsiveness, efficiency, and scalability across various applications and hardware environments.

Responsibilities:

  1. Design and deploy high-performance, resource-efficient model serving architectures adaptable to diverse environments.
  2. Establish and track performance metrics like latency, throughput, and memory usage.
  3. Develop and monitor inference tests, analyze results, and validate performance improvements.
  4. Prepare realistic datasets and scenarios to evaluate model performance in low-resource settings.
  5. Identify bottlenecks and optimize serving pipelines for scalability and reliability.
  6. Collaborate with teams to integrate optimized frameworks into production, ensuring continuous improvement.

Qualifications:

#J-18808-Ljbffr
Location:
London, England, United Kingdom
Salary:
£150,000 - £200,000
Category:
Engineering

We found some similar jobs based on your search