Curious how you handle the "style vector" update frequency? Does it continuously learn as I dictate more, or is it a static snapshot? For brand/personal voice consistency, continuous learning would be huge but might drift.
Also, +1 for the retro CRT aesthetic.
Have you found that forcing the LLM into a structured scoring framework reduces its tendency to hallucinate specs? Or does it just hallucinate the scores with more confidence?
Also, curious if you've tried different models for the "scoring" vs "reasoning" steps. We've found Claude is much better at adhering to complex constraints than GPT-4o for tasks like this.