>We had an internal-workshop led by our internal AI-team (mostly just LLMs), and had the horrible realisation that no one in that team actually knows what the term "AI" even means, or how a language model works.
I'm the AI expert for my org. Everyone else is more or less opposed to AI.
>One senior-dev (team-lead also) tried to explain to me that AI is a subfield of machine-learning, and always stochastic in nature (since ChatGPT responds differently to the same prompt).
machine learning is the sub field of AI.
Not really stochastic as far as I know. The whole random seed and temperature thing is a bit of a grey area for my full understanding. Let alone the topk, top p, etc. I often just accept what's recommended from the model folks.
>We/they are selling tailor-made "AI-products" to other businesses, but apparently we don't know how sampling works...?
Sales people dont tend to know jack. That doesnt mean they dont have an introvert in the back who does know what's going on.
>Am I just too junior/naive to get this or am I cooked?
AI for the most part has been out a couple years. With rapid improvement and changes that make 2023 knowledge obsolete. 100% of us are juniors in AI.
You're disillusioned because the "ai experts" basically dont exist.
That's what I tried to explain then as well, and i brought up stuff like path-finding algorithms for route-finding (A*/heuristic-search) as an more old-school AI part, which didn't really land I think.
> Not really stochastic as far as I know. The whole random seed and temperature thing is a bit of a grey area for my full understanding. Let alone the topk, top p, etc. I often just accept what's recommended from the model folks.
I mean LLMs are often treated in stochastic nature, but like ML models aren't usually? Like maybe you have some dropout, but that's usually left out during inference AFAIK. I dont think a Resnet or YOLO is very stochastic, but maybe someone can correct me.
> AI for the most part has been out a couple years.
With this you just mean LLMs right? Because I understand AI to be way more then just LLMs & ML
yeah, stochastic is there because we give up control of order of operations for speed
so the order in which floating-point additions happen is not fixed because of how threads are scheduled, how reductions are structured (tree reduction vs warp shuffle vs block reduction)
Floating-point addition is not associative (because of rounding), so:
- (a + b) + c can differ slightly from a + (b + c).
- Different execution orders → slightly different results → tiny changes in logits → occasionally different argmax token.
> someone corrected me above, it does seem to matter more then I thought
if you llm agent takes different decisions from the same prompt, then you have to deal with it
1) your benchmarks become stochastic so you need multiple samples to get confidence for your AB testing
2) if your system assumes at least once completion you have to implement single record and replay so you dont get multiple rollout of with different actions
I'm the AI expert for my org. Everyone else is more or less opposed to AI.
>One senior-dev (team-lead also) tried to explain to me that AI is a subfield of machine-learning, and always stochastic in nature (since ChatGPT responds differently to the same prompt).
machine learning is the sub field of AI.
Not really stochastic as far as I know. The whole random seed and temperature thing is a bit of a grey area for my full understanding. Let alone the topk, top p, etc. I often just accept what's recommended from the model folks.
>We/they are selling tailor-made "AI-products" to other businesses, but apparently we don't know how sampling works...?
Sales people dont tend to know jack. That doesnt mean they dont have an introvert in the back who does know what's going on.
>Am I just too junior/naive to get this or am I cooked?
AI for the most part has been out a couple years. With rapid improvement and changes that make 2023 knowledge obsolete. 100% of us are juniors in AI.
You're disillusioned because the "ai experts" basically dont exist.