I don't think the idea of it is terrible, I can see it's use cases, but I think it being implemented in MySQL directly is extremely silly. Imagine having to have all the instances of your database server running on some beefy hardware to handle the prompts, just do the actual processing on a separate machine with the appropriate hardware then shove it into the database with another query when it's done.
It's nice to be able to do something like this without having to use an ORM. Especially if you need a version of the data that's limited to a certain character size.
Like having a replica on the edge that serves the 100 character summaries then only loads the full 1000+ character record when a user interacts with it.
A summary of the review is also more useful than just the first 100 characters of a review.
If the model that's used for that is moderately light I could see it actually using less energy over time for high volume applications if it is able to lower overall bandwidth.
This is all assuming that the model is only run one time or on record update though and not every time the query is run...
Another contender for top tech innovation
That at least has some conceivable benefit? Having a chatbot return memory allocation sizes is comically insecure and dumb
Oh yea, they're definitely not on the same level, it's just some bazinga brained image I had laying around and I thought it was somewhat fitting.
I might need reeducation because I think that image is probably the closest thing to an appropriate usecase for LLMs I've seen ever.
I don't think the idea of it is terrible, I can see it's use cases, but I think it being implemented in MySQL directly is extremely silly. Imagine having to have all the instances of your database server running on some beefy hardware to handle the prompts, just do the actual processing on a separate machine with the appropriate hardware then shove it into the database with another query when it's done.
Probably useful in data science
It's nice to be able to do something like this without having to use an ORM. Especially if you need a version of the data that's limited to a certain character size.
Like having a replica on the edge that serves the 100 character summaries then only loads the full 1000+ character record when a user interacts with it.
A summary of the review is also more useful than just the first 100 characters of a review.
If the model that's used for that is moderately light I could see it actually using less energy over time for high volume applications if it is able to lower overall bandwidth.
This is all assuming that the model is only run one time or on record update though and not every time the query is run...