Installation
Download and install LM Studio directly from lmstudio.ai.Downloading Models
- Open LM Studio and click the Search tab (🔍)
- Search for “LiquidAI” or “LFM2”
- Select a model and quantization level (
Q4_K_Mrecommended) - Click Download
Using the Chat Interface
- Go to the Chat tab (💬)
- Select your model from the dropdown
- Adjust parameters (
temperature,max_tokens,top_p) in the sidebar - Start chatting
Generation Parameters
Control text generation behavior using the GUI sidebar or API parameters. Key parameters:temperature(float, default 1.0): Controls randomness (0.0 = deterministic, higher = more random). Typical range: 0.1-2.0top_p(float, default 1.0): Nucleus sampling - limits to tokens with cumulative probability ≤ top_p. Typical range: 0.1-1.0top_k(int, default 40): Limits to top-k most probable tokens. Typical range: 1-100max_tokens(int): Maximum number of tokens to generaterepetition_penalty(float, default 1.0): Penalty for repeating tokens (>1.0 = discourage repetition). Typical range: 1.0-1.5stop(strorlist[str]): Strings that terminate generation when encountered
Running the Server
Start an OpenAI-compatible server for programmatic access:- Go to the Developer tab (⚙️)
- Select your model
- Click Start Server (runs at
http://localhost:1234)
Streaming Responses
Curl request example
Curl request example
Vision Models
Search for “LiquidAI LFM2-VL” to download vision models. In the Chat tab:- Drag and drop images into the chat
- Click the image icon to upload
- Provide image URLs
Using the API
Using the API
Tips
- GPU Acceleration: Automatically detects and uses available GPUs
- Model Management: Delete models from the My Models section
- Performance: Adjust GPU layers in server settings for speed/memory balance
- Quantization: Q4 is faster, Q6/Q8 have better quality