For instance, imagine I'm writing a small TUI that allows you to browse jsonl files, and want to create a feature to enable natural language parsing. Is there an emerging standard for an implementation agnostic, "Translate this natural query to jq {natlang-query}: response here: "?
If we don't have this yet, what would it take to get this built and broadly available?
This is of course cross-platform and works with both models accessible through an API and local ones.
I'm afraid this might not solve your problem though, as this is not an out of the box solution, it requires the user to either provide their own API key or to install Ollama and wire it up on their own.
https://github.com/day50-dev/llcat
It might help things get closer..
It's under 2 days old and it's already really fundamentally changing how I do things.
Also for edge running look into the LFM 2.5 class of models: https://huggingface.co/LiquidAI/LFM2.5-1.2B-Instruct
`llm` gives my tool a standard bin to call to invoke completions, and configuring and managing it is the user's responsibility.
If more tools started expecting something like this, it could become a defacto standard. Then maybe the OS would begin to provide it.
In a project I’m working on I simply present some data and a prompt, the user can then pipe this into a LLM CLI such as Claude Code.
[0] https://modelcontextprotocol.io/specification/2025-06-18/cli...
Example: https://github.com/anthropics/claude-code/issues/1785
Maybe you can trailblaze and tell users your application will support the $LLM or $LLM_AUTOCOMPLETE variables (convene the committee for naming for better names).
For consistent LLM behaviour, you can use ollama api with your model of choice to generate. https://docs.ollama.com/api/generate
Chrome has a built-in Gemini Nano too. But there isn't an official way to use it outside chrome yet.
When I asked AI it said no such inbuilt model exists (possibly a knowledge date cutoff issue.)
It’s not exactly the intended use case, but it could be coerced to do that.
I’ve seen something else like that, though: voice transcription software that have access to the context the text is in, and can interact with it and modify it.
Like how some people use super whisper modes [0] to do some actions with their voice in any app.
It works because you can say "rewrite this text, and answer the questions it asks", and the dictation app first transcribes this to text, extract the whole text from the focused app, send both to an AI Model, get an answer back and paste the output.
[0]: https://superwhisper.com/docs/common-issues/context
We are also going to want the opposite. A way for an LLM to request tool calls so that it can drive an arbitrary application. MCP exists, but it expects you to preregister all your MCP servers. I am not sure how well preregistering would work at the scale of every application on your PC.
You could fork this, and shell out to an LLM before finally pasting the response.
I'm focused on building it for the macOS ecosystem
Maybe it can be repurposed for natural language in a specific implementation