at Carnegie Mellon University Pittsburgh, PA
hard very hard hard LM-generated code is more reliable than LLM-generated instructions.** A prompt that says "never generate img tags" is a suggestion. A Python function that strips img tags and returns a validation error is a guarantee. The most impactful part of Sentinel isn't the detection or the eval — it's the moment where it stops trying to convince the LLM to behave and just writes code to enforce it. - **Observability for AI agents is fundament
Studied Computer Science at Carnegie Mellon University?
Write a Review