Skip to content

Commit

Permalink
Update README.md
Browse files Browse the repository at this point in the history
  • Loading branch information
YorkieDev authored Aug 25, 2024
1 parent cbd976f commit d43337d
Showing 1 changed file with 80 additions and 2 deletions.
82 changes: 80 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,2 +1,80 @@
# LMStudioWebUI
A wip version of a simple Web UI to use with LM Studio
# LM Studio Chat WebUI

This is a simple, browser-based chat interface for interacting with your LM Studio server. It allows you to connect to your locally hosted LM Studio model and chat with it from any device with a web browser, including mobile phones.

## Features

- Dark mode interface
- Connect to any LM Studio server
- Chat with your LM Studio model
- View token usage and response time for each message
- Mobile-friendly design

## Setup Instructions

### For Desktop Users

1. Download the `lmstudiowebui.html` file from this repository.
2. Save it to a location on your computer that you can easily access.

### For Mobile Users

There are several ways to get the `lmstudiowebui.html` file on your mobile device:

1. **Direct Download**:
- Open this repository on your mobile device's web browser.
- Find the `lmstudiowebui.html` file and download it directly to your device.

2. **Email to Yourself**:
- Download the `lmstudiowebui.html` file on your computer.
- Email it to yourself as an attachment.
- Open the email on your mobile device and download the attachment.

3. **Cloud Storage**:
- Upload the `lmstudiowebui.html` file to a cloud storage service like Google Drive, Dropbox, or iCloud.
- Access the file from your mobile device using the respective cloud storage app.

4. **File Transfer Apps**:
- Use apps like AirDrop (for iOS devices) or nearby sharing (for Android devices) to transfer the file from your computer to your mobile device.

## Usage Instructions

1. **Start LM Studio Server**:
- Open LM Studio on your computer.
- Go to the "Server" tab (In 0.3.x -> Developer -> Local Server).
- Ensure that CORS is enabled and Serve on Local Network is enabled.
- Click "Start Server" and note down the server address.

2. **Open the Chat Interface**:
- On desktop: Double-click the `lmstudiowebui.html` file to open it in your default web browser.
- On mobile: Use a file manager app to locate the downloaded `lmstudiowebui.html` file and open it with your web browser.

3. **Connect to LM Studio Server**:
- In the chat interface, enter the LM Studio server address in the input field at the top.
- Click the "Connect" button.

4. **Start Chatting**:
- Once connected, you can start typing messages in the input field at the bottom of the screen.
- Press Enter or tap Send to send your message.
- The model's responses will appear in the chat window.

## Troubleshooting

- **Can't connect to server**:
- Ensure LM Studio Server is running on your computer.
- Check that you're using the correct server address.
- If accessing from another device, make sure both devices are on the same network.

- **Slow responses**:
- LM Studio processing speed depends on your computer's capabilities. Larger models may take longer to respond.

- **Interface not loading**:
- Try opening the `lmstudiowebui.html` file with a different web browser.

## Security Note

This interface is designed for local use only. Do not expose your LM Studio server to the public internet without proper security measures in place.

## Feedback and Contributions

If you encounter any issues or have suggestions for improvements, please open an issue in this repository. Contributions via pull requests are welcome!

0 comments on commit d43337d

Please sign in to comment.