

Quote from: the-Pi-guy on Mar 14, 2026, 08:45 PMLM Studio has so many options you can mess with.Is there a way to run two models in parallel so that the next token can be sampled using both their outputs? Like averaged or multiplied or whatever?
You can even set how the context window is managed (truncate middle, rolling).
Quote from: Legend on Mar 14, 2026, 05:52 PMDifferent sampling method/temperature?Most of the apps let you change the temperature, top K, top P.
Quote from: Legend on Mar 02, 2026, 09:40 PMAny word on game updates?
Quote from: the-Pi-guy on Mar 14, 2026, 05:35 PMFor some reason, running local models feels goofier than it should be. And I'm not sure why. Like for some reason, prompt adherence is worse when I'm using different Android app.Different sampling method/temperature?
Not sure if they're formatting the requests differently, or passing different default values for the model.
Page created in 0.034 seconds with 11 queries.