On OpenAI o1

Andrew Bolster

Senior R&D Manager (Data Science) at Black Duck Software and Treasurer @ Bsides Belfast and NI OpenGovernment Network

Is LLM Smarter than a 12 year old?

Had a few people ask about the o1 models; at work we’ve requested preview access from Microsoft to get them added to our internal LLM Gateway, so we’ll just wait and see, but there’s been some interesting discourse on it so far. 

My 2c is that this is OpenAI trying to take the chain-of-thought (aka ‘talking to yourself’) in house rather than people doing the intermediate steps themselves. (That means, instead of just running the token prediction, it’s a repeated conversation with itself, with OpenAI providing the ‘inner monologe’ and just magically popping out the answer). This is fine in principal, and is how we do multi-shot RAG among other things, but the two(three) critical parts of this for me are 

a) They charge you for the tokens you don’t control (inner monologue)

b) They actively hide how this inner monologue is constructed, updated, and maintained, and threaten your account if you poke around. 

[c) just so happens that Open AI is doing a funding round. Go figure]

All in all, I’ll stick to 3.5/4o and wrangle constructive context data to inform LLM usecases, in workflows that we know better than Sam et al. and make context-specific data/LLM agents that collaborate to be useful, rather than a one size fits all magical black box that you don’t own. But its fun to watch. 

Highlights of the discourse I’ve come across:

Incidentally, any parents of young children or people who know developmental psychology able to point me at a similar ‘chart’ for accuracy of times-tables in kids?

Anyway, while I was writing this; Google alerts went off; so it seems The S*n picked up on some of my comments around AI; The bad grammar isn’t mine, and apologies to my scouse bretheren for being featured in that paper (at least it’s the Ireland one…)

Published: September 22 2024

  • category:
blog comments powered by Disqus