All active Prompt Engineer roles based in Lithuania.
Pick a job to read the details
Tap any role on the left — its description and apply link will open here.
Share this job
We are launching a strategic pool of AI specialists to accelerate the growth of our AI Practice and turn strong market demand into scalable, repeatable capabilities. This role goes beyond project delivery-it is about shaping how AI is built, shared, and scaled across the company.
As part of this group, you will play a key role in internal capability development, with a strong focus on designing and evolving AI Accelerators-reusable frameworks, templates, and reference architectures covering GenAI, LLM-based solutions, ML pipelines, and cloud-native AI systems. These accelerators will become the backbone of our AI offerings, enabling faster delivery, higher quality, and consistent technical standards across projects.
A second core dimension of the role is knowledge leadership. Strategic hires will actively contribute to internal knowledge-sharing initiatives: technical deep dives, hands-on workshops, demos, and best-practice sessions covering topics such as GenAI orchestration, RAG, model deployment, prompt and system design, observability, and cost-efficient AI architectures. You will help raise the overall AI maturity of the organization while mentoring engineers and influencing technical direction. In addition, members of this group will act as AI project jumpstarters, joining new client engagements for up to 50% allocation over 1-2 months. In this capacity, you will help define solution architectures, validate technical approaches, build initial prototypes, and set strong engineering foundations before teams scale. This gives you exposure to diverse domains and cutting-edge AI use cases while maximizing your impact across multiple initiatives.
We are looking for a highly skilled and innovative Machine Learning Engineer to develop and expand our Generative AI and Computer Vision capabilities. This role involves accelerating Technical Sales Support, Architecture and PoC development.
The global benefits package includes:
Ready to apply?
Apply to Coherent Solutions
Share this job
Kimchi is the AI platform inside CAST AI. We started by helping companies run LLMs on their own Kubernetes clusters - now we're building the execution layer where agents do real work.
Our Infrastructure today: multi-model inference (MiniMax, Kimi, GLM-5, Nemotron, DeepSeek) with intelligent routing, an OpenAI-compatible API, and deployment flexibility from our GPUs to your VPC. The inference layer is the foundation. What we're hiring for sits on top of it: coding agents, agent runtimes, orchestration systems, and the reliability engineering that makes them actually finish things.
Tech Stack: TypeScript, Go, Kubernetes, AWS/GCP/Azure, MCP, Prometheus/Grafana/Loki, GitLab CI, ArgoCD.
Why harness engineering matters here
OpenAI and Anthropic ship models. They also ship one harness each - the scaffolding that turns a raw model into something that can plan, execute, recover, and complete work. We ship a different kind of harness: one built for cost-conscious, long-horizon autonomy, running on inference infrastructure we control end-to-end.
A decent model with a great harness beats a great model with a bad harness. We've watched this play out. The gap between what today's models can do and what you see them doing is largely a harness gap - and that gap is where we operate.
What you'll build
The ratchet.
Every time our agent makes a mistake, we engineer a solution so it never makes that mistake again. That means hooks that enforce constraints the model "knows" but forgets: pre-commit lint checks, permission gates, context compaction before the window fills. Success is silent, failures are verbose.
Long-horizon execution.
Our harness is built around spec-driven autonomy: meta-prompting, fresh context per task, worktree-per-slice git strategy, automatic replanning, crash recovery, stuck detection. We're implementing Ralph loops - when the model tries to exit, we intercept and reinject the goal into a fresh context. The agent reads state from disk and continues. Multi-session, multi-day work, without context rot.
Planner/executor splits.
Planning with a reasoning model, executing with a fast one, evaluating with a third. Separating generation from evaluation beats self-verification because agents reliably skew positive when grading their own work.
The harness surface.
CLI, TUI, MCP integration, sandboxed execution, telemetry. Our AGENTS.md is short - every line traces to a specific thing that went wrong. TypeScript on the surface, Go where it matters.
Memory and context.
Moving agents off laptops, giving them state that survives across sessions, managing context so information lands where it's actionable. Compaction, tool-call offloading, progressive skill disclosure.
What makes this different (with receipts)
You've seen the pitch: "we route to the best model." Everyone says that. Here's what we actually have:
What success looks like (after 6 months):
This is a location-specific opportunity. We are currently accepting applications from candidates residing in the following European countries: Bulgaria, Croatia, Estonia, Greece, Hungary, Latvia, Lithuania, Poland, Romania, Slovakia, Slovenia, and Ukraine.
*As part of our standard hiring process, we would like to inform you that a background check may be conducted at the final stage of recruitment through our third-party provider, Checkr.
*Please note that Cast AI does not provide any form of visa sponsorship/work permit.
#LI-Remote
Ready to apply?
Apply to Cast AI
Cookies & analytics
This site uses cookies from third-party services to deliver its features and to analyze traffic.