Verigram is an international AI company focused on digital identity and fraud prevention. Our solutions are used by banks, fintechs, and online platforms to onboard users remotely and prevent fraud.
And now we are looking for an ML Engineer who can take models from research → production → scale. This role is perfect if you love building LLM systems, computer vision pipelines and high-load AI services used by thousands of real users.
We build production-ready AI systems that process large-scale data every day:
- Face biometrics and liveness detection
- Document verification and KYC/AML
- Real-time video analytics
- LLM assistants and RAG systems
- High-load ML services used by thousands of users
What you will do:
LLM / Generative AI
- Fine-tune open-source LLMs (LoRA / PEFT)
- Build and scale production RAG systems
- Develop embedding and vector search pipelines
- Build AI assistants and internal AI tools
- Prompt engineering and evaluation pipelines
- Run A/B testing and LLM experiments
- Optimize inference (latency / throughput / GPU)
Data & ML Pipelines
- Build ETL/ELT pipelines for ML
- Data validation and dataset versioning
- Automate model training and retraining
- Experiment tracking and quality metrics
ML Infrastructure & Production
- Develop ML services using FastAPI
- Containerize and deploy models
- Optimize GPU inference and performance
- Build batch and real-time inference pipelines
Monitor models:
- data/model drift
- quality degradation
- latency and reliability
Engineering culture:
- Participate in architecture decisions
- Code reviews and documentation
- Contribute to ML engineering best practices
Tech stack:
Python • PyTorch • FastAPI • PostgreSQL • Redis • Kafka • Docker
LangChain • LangGraph • RAG • Vector Search • GPU inference
Requirements:
Must-have
- 2–4 years of experience in ML / DL / NLP
- Experience shipping ML models to production
- Strong Python skills (backend + ML)
- Hands-on experience with PyTorch
- Understanding of the full ML lifecycle
Nice to have:
- Experience with LLMs / RAG / embeddings
- Inference optimization experience
- Experience building high-load services
What we offer:
- Real production ML & LLM projects
- Access to GPU infrastructure
- Strong impact on AI product architecture
- Competitive salary
- Strong engineering culture, low bureaucracy
Join us to build real-world AI products!