-
Notifications
You must be signed in to change notification settings - Fork 43
lmstudio_en
LM Studio is a multi-platform (macOS, Windows, Linux) chat application for large language models, which supports GGUF models (llama.cpp compatible). The following briefly introduces basic usage. Please check their official website for more information.
![image](https://private-user-images.githubusercontent.com/16095339/300233814-43be5862-c01e-42bc-bd05-5d29428add39.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzk1NjEwNjgsIm5iZiI6MTczOTU2MDc2OCwicGF0aCI6Ii8xNjA5NTMzOS8zMDAyMzM4MTQtNDNiZTU4NjItYzAxZS00MmJjLWJkMDUtNWQyOTQyOGFkZDM5LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMTQlMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjE0VDE5MTkyOFomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTBiYWExMjRiY2FiMWVhN2EyODRmOTAzNWI1MWRmNDA1OWEzOWU2ODgyMTFmYTIzNzJlMmRmYzI2MzA1ZGJjN2MmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.YqidTFYBlw60cDPxQoKKpN7EdM41Wc7EqHN5xWq09K0)
Create a new folder models
, and organize the model file as follows (you can also use ln
command here). For example, hfl/chinese-mixtral-instruct
model can be organized as follows, where the GGUF model file should be placed.
- hfl
- chinese-mixtral-instruct
- ggml-model-q4_0.gguf
Click the last button. In local models folder, click change
, navigate the models
folder which was created in the previous step. After this step, all supported models will be listed. Please select Mistral Instruct as the chat template.
![image](https://private-user-images.githubusercontent.com/16095339/300234207-eb318cfd-dd08-4d34-8cf2-a1703e85db1b.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzk1NjEwNjgsIm5iZiI6MTczOTU2MDc2OCwicGF0aCI6Ii8xNjA5NTMzOS8zMDAyMzQyMDctZWIzMThjZmQtZGQwOC00ZDM0LThjZjItYTE3MDNlODVkYjFiLnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMTQlMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjE0VDE5MTkyOFomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTNhMDgwYWRmNTY0NmUyZDkyNzkzYzdkMTI5ZjE3NjU4OGUyMGNmMGM1OGFlMTUzZmE5NTIwMWE0NTBmYzI3NzImWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.MXZb610yLpcYBPLfHKG8eHR4_ELpEqjO8iOB4oF38g0)
Note: Currently, Mixtral model is recognized as LLaMA, which does not affect the model inference.
Click the third button for chatting. You can choose the model on top of the interface. Right side bar shows system prompt, enabling GPU, context size, etc. Left side bar shows history conversations.
![image](https://private-user-images.githubusercontent.com/16095339/300234387-bf78014f-1cdb-4363-b8bb-877f5c59eacb.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzk1NjEwNjgsIm5iZiI6MTczOTU2MDc2OCwicGF0aCI6Ii8xNjA5NTMzOS8zMDAyMzQzODctYmY3ODAxNGYtMWNkYi00MzYzLWI4YmItODc3ZjVjNTllYWNiLnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMTQlMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjE0VDE5MTkyOFomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTE1MmY4MWM5MDk5ZmQwZDY5ODRjNTA3YTA3YWJiNDVjN2JiNTlhY2M1YThmZDU0NjhmNDdiN2E0MzVlMmYyMjImWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.jpNPY44WZl2nmjBArW95xtMyTuayN0sTcp8JEMUIzRI)
- Model Reconstruction
- Model Quantization, Inference and Deployment
- System Performance
- Training Scripts
- FAQ