Replies: 1 comment
-
Nope it's not implemented yet. Although there is a foundation for that in rust's part of this plugin. I mean that all the thinking part is stored separately in cache from the rest answer. But it still has to be pulled through the whole chain and I don't I'd have time to do that in this quarter unfortunately. UPD: there's literally no way to split it up in the broadcasting part in an easy reliable way without writing your own parser on the python's side, so it could be quite hacky to do that well |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
For reasoning models, now it by default shows the thinking process of the LLM with the tag, is there a way to hide it with a setting, or perhaps even collapse it if ST supports it?
Beta Was this translation helpful? Give feedback.
All reactions