jackschedel / koalaclient Goto Github PK
View Code? Open in Web Editor NEWThe best LLM API Playground Interface (for me)
Home Page: https://client.koaladev.io/
License: Creative Commons Zero v1.0 Universal
The best LLM API Playground Interface (for me)
Home Page: https://client.koaladev.io/
License: Creative Commons Zero v1.0 Universal
Hey @jackschedel !
Thanks for the awesome repo!
Azure Openai models are named as 'gpt-35-turbo-1106' instead of 'gpt-3.5-turbo-1106'. Could you add a mapping for azure endpoints by checking the endpoint URL? Azure endpoints are in this format: 'https://DEPLOYMENT-NAME.openai.azure.com/'.
Something like this would work: if 'openai.azure.com' in endpoint_url, change 'gpt-3.5-turbo-1106' to 'gpt-35-turbo-1106'.
Bests
after upgrade to latest, have to delete all config and restart the software. The old config files have to be exported and imported
imported error
may need to come with full rework of chat dragging behavior (preventDefault and have custom animation and behavior)
i.e. I could specify that T-binding is always changed to keybinding, toad to code etc
would start as electron only, implementing local temp files on website sounds like a pain
If I try to move my cursor to a previous word by using Ctrl+Arrow key, it changes documents/chats?? Please don't hijack basic keyboard shortcuts.
If the timing is just right, the "Stop Generating" button can cause some wacky behavior - such as two responses being written at once to the same message, or the response stream getting appended to the last user message (rather than a new assistant message).
Usually this happens when I quickly edit and re-send a message (in under a second).
I think the reproduction trigger is pressing the "Stop Generating" button after the request is sent, but before a response stream has been received.
A simple fix might be to not show the button until the response has been received, although it might be better to also abort the request with the button.
Has anyone experienced this, and am I understanding correctly?
There are a bunch of models on https://openrouter.ai/docs#models, some of which are free. Is it possible to support them?
Do you have any plans on integrating Claude API not just OpenAI in your fork?
Thank you.
A couple small things to be ironed out with the max content setting calculations.
If I'm understanding this correctly, the API considers max_tokens
to be the context and the response combined. So there shouldn't be a need to subtract the max context setting from the max token setting.
Currently we are calculating the parameter like this:
max_tokens = (maxTokenSetting - maxContextSetting)
We should simply do this before each request:
maxTokenSetting
directly as the parameter in the requestmaxContextSetting
Or, to calculate on the fly similar to this, we'd have to introduce a third setting, like I described in another issue:
max_tokens = (maxContextSettings - desiredMaxResponseSetting)
(with the maxTokenSetting still serving to cap this at a maximum)
max_token
value that is too large to be allowed:
(response is an error because the maximum value formax_token
on gpt4 is ~8000).
Summary:
Allow users to input custom model names in the model selection dropdown or within the settings when choosing which AI model to interact with.
Problem:
Certain proxies for OpenAI services have been experiencing issues where the /models
endpoint might be malfunctioning or is not populating the model selection list as expected. This results in users being unable to select from the full range of available models, potentially limiting their access to specific or preferred AI models that are otherwise operational.
Proposed Solution:
Implement a feature within the chat interface that enables a user to add custom model names directly to the model selection dropdown. This could take the form of an input box within the dropdown itself, allowing for direct typing of the model name. Alternatively, a small section in the settings could be dedicated to managing custom model names where users could add or remove model names that they frequently use or wish to test.
Benefits:
i.e. neovim extension, chrome extension
I've had an idea for a while now:
Similar to the 'Model Swap' button you've added, I also frequently adjust temperature, top_p, and _ presence penalty_ from my defaults l, according to my needs at the time.
For example:
Perhaps:
happens when selecting and copying text using ctrl-a or dragging below the final line
Sometimes the max_token parameter for the title generation is negative:
The response is something like:
{
"error": {
"message": "-228 is less than the minimum of 1 - 'max_tokens'",
"type": "invalid_request_error",
"param": null,
"code": null
}
}
And other time it is way too long, and I get titles that are an entire paragraph.
Just notice that the "Max-Tokens" API setting isn't getting included in the chat request body, even though it appears to be a user configuration setting, and is likely being stored.
This would be best implemented as an optional setting, that a user could opt into. If enabled, the only models to choose from in the chat config would be:
gpt-3.5-turbo
gpt-4
The token length models could then be automatically selected only when required, determined by these factors:
gpt-3.5-turbo-16k
model if you are well under the 4000 token limit for gpt-3.5-turbo-16k
) likely electron only because system clipboard access on web is weird.
SHIFT
+ ENTER
behavior when editing existing chat messages.I am frequently editing existing System/User/Assistant chat messages. It is a useful technique to:
ENTER
key sends the messageSHIFT
+ ENTER
creates a new line without sending the message.
ENTER
key creates a new lineSHIFT
+ ENTER
will send the messageBecause
SHIFT
+ENTER
explicitly sends the HTTP request, it seems like this might've been deliberate by the original author? But I can't see any reason as to why!I've even been using a UserScript to accomplish this for a few weeks now, so I feel familiar with each UX.
But if there is a good reason for the inconsistency, it could be implemented as something that the user could toggle on or off, in the configuration settings.
(discord has this keybinding)
maybe down arrow to edit most recent Assistant Chat?
maybe want to add a setting toggle to this, and/or have different behavior for the left menu button and the mobile bar button
Steps to Reproduce:
gpt-4-0125-preview is a new api added in Openai.
we can add it.
Thanks!!
will likely want line numbers in conjunction with this
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.