Lasso is on a mission to secure the use of LLMs in the real world protecting data, privacy, and businesses from AI threats. From the first line of code to protecting real-world data, every decision matters. If you're ready to shape the future of AI security - we would love to hear from you!
We’re looking for a DevOps Engineer to join our growing team at Lasso Security. You’ll be part of a growing and passionate team, building secure, scalable, and intelligent infrastructure that powers cutting-edge AI-driven products. This is a hands-on opportunity for someone who loves learning, solving complex challenges, and automating everything - an autodidact eager to grow into a key DevOps role in a fast-moving GenAI security startup.
Responsibilities
- Design, implement, and maintain scalable AWS and Azure environments, including Kubernetes clusters and related services.
- Build and enhance CI/CD pipelines (GitHub Actions, Argo CD, etc.) to streamline deployments and ensure reliability.
- Automate processes with Bash and Python scripts.
- Implement and manage monitoring, alerting, and observability stacks using Prometheus, Grafana, Loki, and Alertmanager.
- Deploy and support AI/ML/LLM-related infrastructure and services
- Collaborate closely with Engineering, Product, QA, and Data Science teams to support rapid development.
- Drive a culture of security, compliance, and reliability through best practices and continuous improvement.
- Optimize cloud costs, scalability, and system performance across environments.
- Continuously evaluate and integrate new DevOps tools and practices to enhance developer experience and operational excellence.
Requirements
- 3+ years of experience as a DevOps or Infrastructure Engineer.
- Hands-on experience with AWS (EKS, EC2, S3, CloudFront, Cognito, etc.).
- Familiarity with Azure or enthusiasm to expand into multi-cloud environments.
- Strong knowledge of Docker and Kubernetes (Helm, scaling, and management).
- Experience with CI/CD tools (GitHub Actions, Jenkins, or other GitOps platforms).
- Proficiency with Terraform or other Infrastructure-as-Code tools.
- Solid understanding of networking, DNS, load balancers, and CDNs.
- Experience with monitoring and observability stacks (Prometheus, Grafana, Loki, etc.).
- Comfortable in Linux environments and scripting in Bash or Python.
- Excellent communication and teamwork skills.
- A self-driven, curious, and autodidactic mindset - someone who loves to figure things out.
Nice to Have:
- Background in software development (Node.js, Python, or similar).
- Familiarity with AI/ML pipelines or MLOps.
- Knowledge of Azure Database, AWS RDS, Couchbase, or managed databases.
- Experience integrating and maintaining third-party and open-source tools securely in production.
- Understanding of SRE principles and observability best practices (SLOs, SLIs, SLAs).