Loading...
Loading...
Loading...
Loading...
Loading...
7 years in data and machine learning, with experience at financial sector companies working on LLM agent architecture, RAG pipelines, and production evaluation systems. PromptEval was born from a real need: teams running LLMs in production were breaking agents due to poorly written prompts, with no objective way to diagnose the problem. The idea was to formalize the industry's technical criteria for reviewing production prompts (clarity, specificity, structure, robustness) into an automated, auditable, and repeatable score for any team working with LLMs.
...AI Engineer at PromptEval