logo
logo
https://groupby-public-image.s3.ap-northeast-2.amazonaws.com/startups/1236/1767856035/무제.001.png
(주)프렌들리에이아이Inference engineered for speed, scale, cost-efficiency, and reliability
서울 강남구B2B ‧ AI
Solution Architect - AI Inference Specialist
빠른 응답
포지션인공지능 · 머신러닝
경력 구분경력 3년 이상
특이사항CTO 있음, 사수 있음, 부분 원격근무 가능
스킬
Docker
Helm
Kubernetes
Terraform
AWS-EKS
주요업무

About the job

FriendliAI is seeking a Forward Deployed Engineer (FDE) to assist enterprises in deploying, scaling, and operating generative and agentic AI workloads on FriendliAI infrastructure. You will work directly with customers to solve and implement production-grade applications using our products, such as Serverless Endpoints, Dedicated Endpoints, or Container.

Friendli Container is our service that allows customers to download our inference engine as Docker images and deploy it in their chosen environment, such as private clouds or on-premises. Our Friendli Container can be adopted directly to AWS EKS clusters using our EKS add-on product.

You will work directly on our customers’ projects, collaborating with their engineering teams to solve AI inference challenges like scaling, orchestration, and monitoring. This is a hands-on, customer-embedded role. If you have worked in DevOps, platform engineering, or SRE for AI applications, this is your ideal position.

 

Key Responsibilities

  • Design and implement large-scale deployment architectures for LLM and multimodal inference
  • Deploy and manage containerized workloads across Kubernetes clusters
  • Diagnose production issues, such as performance bottlenecks, and implement temporary fixes as needed
  • Collaborate with customers’ DevOps teams to integrate FriendliAI’s infrastructure into their CI/CD workflows
  • Develop scripts, Helm charts, and Terraform modules that simplify repeated deployments
  • Contribute field insights to shape our platform reliability, observability, and scaling strategies
  • Lead workshops, technical sessions, or webinars to help customers master infrastructure best practices
자격요건

Qualifications

  • 3+ years of experience in cloud infrastructure, DevOps, or reliability engineering
  • Bachelor’s or Master's degree in Computer Science, Computer Engineering, Electrical Engineering, or equivalent
  • Proficiency with Kubernetes, Docker, Terraform, and Helm
  • Strong foundation in distributed systems, networking, and performance tuning
  • Experience with GPU-based computing and generative AI model serving workloads
  • Strong technical background in backend systems or AI tooling
  • Experience operating workloads on AWS, GCP, or OCI
  • Excellent problem-solving and debugging skills in real-world environments
우대사항

Preferred Experience

  • Experience deploying large models (LLMs, diffusion models) on GPUs or clusters
  • Familiarity with inference frameworks (Triton, vLLM, TensorRT, DeepSpeed-Inference)
  • Familiarity with observability stacks (Prometheus, Grafana, Loki, ELK, OTEL)
  • Understanding of networking security and compliance frameworks (e.g., SOC 2)
  • Experience supporting on-prem or hybrid-cloud deployments
채용절차
  • 서류 전형 → 1 차 인터뷰 → 2 차 인터뷰 → 컬처핏 인터뷰 → 최종 합격


  • 근무 형태: 정규직
  • 근무 환경: 유연한 원격 근무 및 필요시 오피스 출근을 병행하는 하이브리드 근무


  • Aggressive Compensation: 업계 최고 수준의 기본급과 성과에 따른 상한선 없는(Uncapped) 파격적인 커미션 구조 제공
  • Stock Option: 회사의 성장과 함께 자산 증식이 가능한 스톡옵션 부여
  • 채용 절차는 상황에 따라 일부 변경될 수 있습니다
합류하면 얻을 수 있는 것
#트렌디한_기술#기술_교육#의견_제시에_수용적#직무에_구애받지_않는_의견_개진#인재_밀도#도전적인_문제#성과에_따른_보상#인생을_걸_만한_비전
근무지
서울 강남구 봉은사로 222 (서울, 강남구)
logo
(주)그룹바이HR | 서울특별시 영등포구 영등포로 150 C동 908호대표 : 임진하 / 박상민사업자등록 : 333-88-02226유료직업소개업 등록번호 : 제 2005-3180270-14-5-00019호직업정보제공사업 신고번호 : J1200020230023문의 : hello_world@groupby.biz | 070-8018-7502 (평일 오전 10시 ~ 오후 6시)
instagramkakaotalkyoutube