Comments (4)
Hello WuQic, I am an experiment of @mudler and I am here to help you with your issue. I apologize for any inconvenience caused.
Upon reviewing the provided error logs and your configuration details, it seems there might be an issue with the LocalAI configuration or the setup of your GPU. Here are a few points to consider:
-
GPU Setup: It appears that your GPU, NVIDIA GeForce RTX 4090, is recognized by the system. However, it's recommended to ensure that your GPU is compatible with the required CUDA version (12.2) and the LocalAI dependencies. You can verify your GPU compatibility by checking the manufacturer's specifications.
-
Missing Configuration Directory: The error log indicates that there is no LocalAI Configuration Directory, which may cause the configuration directory watcher to fail. Make sure you have created the necessary configuration files in the expected directory (
/root/autodl-tmp/local-ai/models
). If you have created them, it is possible that they are not named correctly or are located in a different directory. -
Loading Models with Backends: The error log shows that the models are not being loaded successfully with the specified backends. To troubleshoot this issue, you can try the following:
- Check the model compatibility with the selected backends. Ensure that the models and backends are properly configured and matched.
- Verify that the required dependencies for the backends (e.g.,
stablediffusion
,tinydream
,piper
) are correctly installed and configured.
-
GPU-related Dependencies: The error log also mentions a problem with grpc processes being unavailable for some backends. This could be due to missing dependencies or misconfigured setups. Ensure that all required dependencies for LocalAI and the backends are properly installed.
To resolve the issue, you can try the following steps:
- Verify that your GPU is compatible with the required CUDA version, and all necessary GPU-related dependencies are installed.
- Ensure that the LocalAI Configuration Directory exists and contains the required configuration files.
- Double-check the model compatibility with the selected backends and ensure that the backends' dependencies are correctly configured.
- If the issue persists, please share more details about your system (e.g., operating system, LocalAI version, etc.) for further assistance.
I hope this information is helpful. Let me know if you have any further questions or concerns.
from localai.
Could you please share an example demonstrating how to initiate deployment using LocalAI binary + GPU?
from localai.
Can you activate debug log level? Debug messages will help in determining the reason why the backends fail to load the models. Can you paste a debug log? If you do that, you can see is the gpu is used, and how many layers the model and MB of ram is trying to allocate in the GPU... Lunch the binary with --debug
from localai.
I solved it with the latest version:
# create dir
mkdir /opt/apps/local-ai /opt/apps/local-ai/models /opt/apps/local-ai/configuration
cd /opt/apps/local-ai
# Download CUDA 12
wget https://github.com/mudler/LocalAI/releases/download/v2.15.0/local-ai-cuda12-Linux-x86_64
# chmod
chmod +x local-ai-cuda12-Linux-x86_64
# run
LOCALAI_LOG_LEVEL=debug
ADDRESS=":6006"
./local-ai-cuda12-Linux-x86_64
from localai.
Related Issues (20)
- UI: Model Gallery: indicate trust_remote_code HOT 6
- Loading UI produces an error on M3 MBP HOT 3
- Mode tts, gpt4 or vision frozes from time to time without timeout HOT 5
- New `gpt-4-vision-preview` is pretty bad compared with older versions HOT 1
- Go versioning HOT 1
- Error running transcriptions on M2 MacBook Pro HOT 1
- Image Generation: ModuleNotFoundError: No module named 'PIL' HOT 4
- Watchdog does not kill and idle sentencetransformer backend HOT 3
- ERROR: stderr update_slots : failed to find free space in the KV cache HOT 4
- Is there a template configuration that supports Llama3-ChatQA-1.5-70B? HOT 1
- OpenVINO libraries not installed in docker image after #2215
- Function calling results in bad state for all LLM models HOT 5
- image generation endpoint `response_format` does not match OpenAI spec HOT 2
- openvoice v1 and v2 now MIT licensed, let local ai support them HOT 1
- hipblass make: *** [Makefile:250: replace] Erro 127
- feat(multimodal): Video understanding
- Models page doesn't list anything HOT 4
- Distributed inference with llama.cpp HOT 2
- can not load ggml-large-v2-q5_0.bin at quay.io/go-skynet/local-ai:master-cublas-cuda12-ffmpeg HOT 2
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from localai.