I totally agree with the sentiments in the video that the current state of agent architectures feels not there yet and we need an interoperable composable standard.
In my own experiments I have had major failures where much of the text is fabricated by the LLM to the point where I just find it hard to trust even with great prompt engineering. What I have been very impressed with is it’s ability to take medium quality ocr from acrobat with poor formatting, lots of errors and punctuation problems and render 100% accurate and properly formatted output by simply asking it to correct the ocr output. This approach using traditional cheap ocr for grounding might be a really robust and cheap option.
My brain must be warped by using LLMs because I could not help but think that it would be an amazing prompt, complete with multiple examples of each principal. Perhaps good prompting is good writing?
For me, it's a good argument against conspiracy theories. There's no way there could be secretive organisations running for years and controlling the world without screwing up in stupid and obvious ways.
To the contrary does it not imply any well organized group should have an easier time of accomplishing their aims while the rest are mired in the chaos?
The risks in LLM powered systems seems like an opened pandora's box the more I look into mitigating it.