Neat idea. Regarding the performance, I think you could get a lot better performance by training a small classifier model, essentially an embedding model, and using the LLM as the distillation source. This would both be much smaller, addressing your desire for it to run in browser, while also being much more performant, addressing your quantization need. Using the full LLM is a bit overkill and you can extract the core of what you're looking for out of it with something a little custom.
dheera 10 hours ago [-]
I just wish the stupid chatbots on customer service websites would show "Get human" as an option sooner. 99% of time there is a mistake and the bots don't have authority to correct it.
indexerror 8 hours ago [-]
Thats why I go for the nuclear option (the one that requires higest attention, like account cancellation) when presented with a list of options by a bot. More likely to get a person that way.
F7F7F7 16 hours ago [-]
This is great. But the vast majority of non-gaming non-terminal-loving-developers human beings want to avoid the keyboard at all costs.
They are doing it via trackpads, mousepads, touch screens, etc. Which are all inputs that transcend language or the ability to find meaningful words.
lukasb 15 hours ago [-]
ChatGPT has 400 million weekly users
indexerror 8 hours ago [-]
People are coming to ChatGPT with an intent to type. Its much harder to introduce text-based interface to otherwise click-based interfaces because of the friction. Thats why chatbots often fail to gain adoption, because they are a byproduct of the product team being lazy and not thinking through the UX.
lukasb 7 hours ago [-]
I 100% agree that grafting a chatbox on an existing app that wasn't meant for it sucks. (I'm working on an AI app where I set a "no chatbox" constraint, see my profile.)
However, I think a lot of people are willing to type if it will do something useful they can't otherwise accomplish. And: typing gives you user intent, very valuable.
frohrer 11 hours ago [-]
This is cool. Thanks for posting.
I could see this adaptive UI / reactive intent UX become valuable for a couple use cases:
1. Tutorial modes for complicated UX that gradually introduces the user to new features as they "progress" through the website.
2. Warning users that their inputs could lead to unforeseen consequences. Think "you forgot to attach the file you talked about in the email" warning that Gmail already does.
JamesBurdge 10 hours ago [-]
This feels similar to the experience you get when asking ChatGPT a programming question. The UI shifts to better display code and provides buttons to copy content.
ab-dm 15 hours ago [-]
This is a very cool idea, I think this could potentially be more powerful by (at least initially) limiting it to small sections of the UI (Like you've done in your example)
You could take it a step further and, instead of adding/removing buttons, simply updating the copy of (a) button(s) based on the question, which I imagine you could manage with a very small model.
nisalperi 11 hours ago [-]
Cool demo and a neat write-up! I've been experimenting with similar patterns to differentiate between search queries/question answering and presenting specific outputs to the user.
Thanks for taking your time to write this up
keyserj 17 hours ago [-]
Cool idea. FYI the GitHub link at the bottom leads to "page not found". Maybe the repo is not public?
eadz 13 hours ago [-]
Thank you, you're right. Now public.
artificialprint 17 hours ago [-]
Interesting, but also true that intent can be much more accurately passed with words, than inferred with guessing? Visit apple.com, reorganize this list in alphabet order, find me a ...
You get it
smokel 16 hours ago [-]
This looks like a neat idea, but I'm not too positive about it.
This makes using computers even harder to explain to people who do not spend their entire day keeping up with the latest developments. They cannot form a mental image or reuse any memory of what will happen next, because it is all context dependent.
On the other end of the spectrum, for power users, dynamically adapting user interfaces can also be quite annoying. One can't type ahead, or use shortcut keys, because one doesn't know what the context will be. Having to wait any positive amount of time for feedback is limiting.
Then again, there are probably tons of places where this is useful. I'm just a bit disappointed that we (as a society) haven't gotten the basics covered: programming still requires text files that can be sent to a matrix printer, and the latency of most applications is increasing instead of decreasing as computers become faster.
eadz 13 hours ago [-]
I think there is a good chance that this design is more suitable for power uses.
My hunch is that typing 5 words could be faster than going through 3 menus and screens. I don't see this as a replacement for normal UI, but as an optional shortcut - if you're happy typing.
The problem is it's hard to come up with better examples than your toy examples of weather and maps. Goodness there are so many travel planning demos. Who actually wants the context switch of a UI popping up mid-typed-sentence? Is a date picker really more convenient than typing "next school break"? Visualizations are interesting -- but that changes the framing from soliciting input to densifying information presentation. Datagrids and charts'll be valuable.
Anyway, it's a space that's still starving for great ideas. Good luck!
eadz 13 hours ago [-]
Thank you for the references. I agree there is a lot of room for innovation in this space.
Main differentiator here is as-you-type UI customization, and also in this case the UI isn't generative in the sense the options are hard coded - the LLM just chooses between them.
I think we'll see a mix of all of the above in the future as we take full advantage of LLMs.
n49o7 16 hours ago [-]
What the Windows search box wanted to be.
teaearlgraycold 17 hours ago [-]
This seems like a very expensive way to do basic NLP
Rendered at 14:45:47 GMT+0000 (UTC) with Wasmer Edge.
They are doing it via trackpads, mousepads, touch screens, etc. Which are all inputs that transcend language or the ability to find meaningful words.
However, I think a lot of people are willing to type if it will do something useful they can't otherwise accomplish. And: typing gives you user intent, very valuable.
You could take it a step further and, instead of adding/removing buttons, simply updating the copy of (a) button(s) based on the question, which I imagine you could manage with a very small model.
Thanks for taking your time to write this up
You get it
This makes using computers even harder to explain to people who do not spend their entire day keeping up with the latest developments. They cannot form a mental image or reuse any memory of what will happen next, because it is all context dependent.
On the other end of the spectrum, for power users, dynamically adapting user interfaces can also be quite annoying. One can't type ahead, or use shortcut keys, because one doesn't know what the context will be. Having to wait any positive amount of time for feedback is limiting.
Then again, there are probably tons of places where this is useful. I'm just a bit disappointed that we (as a society) haven't gotten the basics covered: programming still requires text files that can be sent to a matrix printer, and the latency of most applications is increasing instead of decreasing as computers become faster.
My hunch is that typing 5 words could be faster than going through 3 menus and screens. I don't see this as a replacement for normal UI, but as an optional shortcut - if you're happy typing.
The problem is it's hard to come up with better examples than your toy examples of weather and maps. Goodness there are so many travel planning demos. Who actually wants the context switch of a UI popping up mid-typed-sentence? Is a date picker really more convenient than typing "next school break"? Visualizations are interesting -- but that changes the framing from soliciting input to densifying information presentation. Datagrids and charts'll be valuable.
Anyway, it's a space that's still starving for great ideas. Good luck!
Main differentiator here is as-you-type UI customization, and also in this case the UI isn't generative in the sense the options are hard coded - the LLM just chooses between them.
I think we'll see a mix of all of the above in the future as we take full advantage of LLMs.