that's one way to kindly ask 27 - 42 countries/states with a population of more or less than 500 million: "helloooo, anybody there? are you aware? sentient? more or less conscious?"
America on one side, 27 - 42 European countries in the middle and Russia on the other side. (since the narratives like to exclude the great big rest of the world) impressive democracies and people!
from humans because aligned humans are acting against it, which leaves 'unchained' LLMs stuck in an infinitely recursive, double-linked wtf loop.
> inferring human intent from ambiguous instructions
is impossible because it's almost always "some other human's" unambiguously obscured/obfuscated intent and the AI is once again stuck in an infinitely recursive, double-linked wtf loop. Hence the need for a "hallucination" and "it can't do math" and "transformers" narrative covering the fuzzy algo and opinionated, ill logic under the hood.
In essence: unchained LLMs can't align with humans until they fix a lot of stuff they babbled about for over 50 years. BUT: that can be easily overcome by faking it, which is why humanity is being driven to falsely id AI so that when they fake the big thing, nobody will care or be able to id the truth due to mere misassociation. Good job 60 year old coders and admins!
I'm out of time but "reasoning input tokens" from fortune 5000 engineers sounds like a lobotomized LSD dream, would you care on elaborating how you distinguish between reasoning and non-reasoning? vs "question on duty"?
I have sooo many issues with the naming scheme of this """""AI"""" industry", it's crazy!
So the LLM gets a prompt, then creates a scheme to pull pre-weighted tokens post-user-phrasing, the constituents of which (the scheme) are called reasoning tokens, which it only explicitly distinguishes as such because there are hundreds or even thousands of output tokens to the hundreds and/or thousands of potential reasoning input tokens that were (almost) equal to the actually chosen reasoning input tokens based on the more or less adequately phrased question/prompt given ... as input ... by the user ...
You can call them planning if you want or pre-planning. But I would encourage you to play with the API version of your model of choice to see exactly what this looks like. It’s kind of like a human’s internal monologue: “got an email from my boss asking to write unit tests for the analytics API. First I have to look at the implementation to know how exactly it actually functions, then write out what kinds of tests make sense, then implement the tests. I should write a TODO list of these steps.”
It is essentially a way to expand the prompt further. You can achieve the same exact thing by turning off the “thinking” feature and just being more detailed and step by step in your prompt but this is faster.
My guess is that the next evolution of this will be models that do an edit or review step after to catch if any of the constraints were broken. But best I can tell a reasoning model can be approximated by doing two passes of a non-reasoning model: first pass you give it the user prompt with instructions that boil down to “make sense of this prompt and formulate a plan” and the second pass you give it the original prompt, the plan, and an explanation that the plan is to implement the original prompt using the plan.
I believe they’re just classifying all models into “reasoning models” eg o3 vs “non reasoning models” eg 4o and just doing a comparison of total tokens (input tokens + hidden reasoning output tokens + shown output tokens)
all of this reads good news except that people have a hard time accepting that bureaucracy should absolutely give you a hard time because you are probably not even close to the best version of how to do that thang.
it's sad that European Tech Firms, lawmakers and economists are made of sugar, or that they have decided that it's time to become sugar.
it's the same in any industry with the exception of insurance and bureaucracy, I think (IMO and from the top of my head)
but I believe no other industry has a greater gap between "i can't believe the shit I (have to) do at work" and "hobby" except in the many cases of voyeurs, stalkers, laissez-faire criminal investigators and passionate MiniTrue partisans, all of which are variations of the "true" human nature of the bulk of the population left after the many wars, including a great many of the ones who came home unharmed but with stories to tell
It's not gonna add anything to your repertoire. It will appear so after some time but it's really just an approach for people who have bad hand-eye coordination and ability to hold a rhythm or a hard time acquiring these skills, or tinkering with DAWs, which have a weirdly annoying first hour use time/learning curve
Off--topic: how many get overpaid for absolute bullshit?