With Gemini 3.1 Flash-Lite launching today, my mind goes towards thinking about things I wouldn't have thought to build because of expense.
However, when the most inexpensive models get better/cheaper it tends to sort of unlock ways of thinking about features we wouldn't have explored before the cheaper AI-driven tools are possible.
This is kind of abstract but is there a way you think about this sort of thing?
Top comments (7)
What @heckno said, and then expand it to any (all!?) hardware devices that I own.
The bigger question is: will it indeed be in the end/soon that SaaS will go away and everyone (almost everyone?) will build the tool that they need specifically tailored for them inhouse?
Oh yeah good call
I don’t think SaaS disappears.
Cheaper AI makes building more accessible, yes. But building something tailored in-house still requires architecture, long-term maintenance, security, evolution, and governance. AI can reduce implementation effort, but it doesn’t remove structural complexity.
In the short term, more teams might build internal tools. In the long term, the real constraint won’t be model cost — it will be software engineering capacity and organizational maturity.
SaaS solves that by externalizing complexity. That doesn’t go away just because tokens are cheaper.
I'm curious about stuff that can plug into my smart watch data where I can pipe in basic questions about the trends and get answers. It's not complicated stuff but I want to be able to do in high volume
I had a slightly different take on this recently and it was in the corporate setting, do people actually know or pay attention to tokens or other expense calculations on their AI use. Is it obfuscated?
I've noticed the people who consider expenses with models tend to be the same people who tinker and likely used a personal card at some point. Others openly mention they work for
insert big company hereand have essentially unlimited spending on certain tools and models.Lower costs definitely broaden the perspective. But it’s also worth asking whether there’s a real need in the first place — or whether we’re creating one just to take advantage of cheaper AI.
There is many factors going into this:
It's something I keep in mind as new models are being published and accessible for us to use. Of course, the main thing about AI is the environment factor that goes into it because you have to build a lot of Data centers to compute your model. I saw a video where it explains that OpenAI just simply "Made it bigger" and the LLM model just became smarter. It's pretty much the whole reason why there is so many data centers and RAM that is needed just to simply use AI. I might be wrong, but let me know!
Thanks for sharing @ben!