Comments (7)
I am running gpt-pilot with Llama-3-70B-Instruct.Q5_K_M It does not seem to suffer from malformed JSON issues, although it has other problems.
I changed the source to have 3000 timeout for reading instead of 300 and changed the code to use Llama tokenizer instead of tiktoken from OpenAI, Llama tokenizer uses tiktoken internally, so they are pretty close.
from gpt-pilot.
Thanks for sharing that. I guess I'd need a memory upgrade then. on 32gb. will 64gb do?
I already changed api timeout to 30 min because I still got "api timeout" error w/ 10min connecting to LM Studio (llm on cpu).
Curious how you changed to Llama tokenizer. Do you mind sharing changes?
Thanks
from gpt-pilot.
Thanks for sharing that. I guess I'd need a memory upgrade then. on 32gb. will 64gb do? I already changed api timeout to 30 min because I still got "api timeout" error w/ 10min connecting to LM Studio (llm on cpu). Curious how you changed to Llama tokenizer. Do you mind sharing changes? Thanks
import re
import requests
import os
import sys
import time
import json
import tiktoken
from prompt_toolkit.styles import Style
from jsonschema import validate, ValidationError
from utils.style import color_red, color_yellow
from typing import List
from const.llm import MAX_GPT_MODEL_TOKENS, API_CONNECT_TIMEOUT, API_READ_TIMEOUT
# alexo: Slow LLM, override
API_READ_TIMEOUT=3000
from const.messages import AFFIRMATIVE_ANSWERS
from logger.logger import logger, logging
from helpers.exceptions import TokenLimitError, ApiKeyNotDefinedError, ApiError
from utils.utils import fix_json, get_prompt
from utils.function_calling import add_function_calls_to_request, FunctionCallSet, FunctionType
from utils.questionary import styled_text
from .telemetry import telemetry
#tokenizer = tiktoken.get_encoding("cl100k_base")
# alexo: Should Llama-3 tokenizer be used?
from transformers import AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-70B", revision="refs/pr/6")
--------------------------------------------------- just a few top lines from the file.
this file is in utils folder, llm_connection.py
from gpt-pilot.
Yikes.. getting an error
Cannot access gated repo for url https://huggingface.co/meta-llama/Meta-Llama-3-8B/resolve/main/config.json.
Access to model meta-llama/Meta-Llama-3-8B is restricted. You must be authenticated to access it.
from gpt-pilot.
Yikes.. getting an error
Cannot access gated repo for url https://huggingface.co/meta-llama/Meta-Llama-3-8B/resolve/main/config.json. Access to model meta-llama/Meta-Llama-3-8B is restricted. You must be authenticated to access it.
Probably. It is free though.
from gpt-pilot.
Are you storing your token somewhere? how is it pulling?
from gpt-pilot.
Not sure what you mean by storing token. If you navigate to Hugging Face and try to access Llama models it will ask you to go through a quick process. You can pretty much invent the info. At some point in time, I think, I also set something up with ssh keys and HF account.
from gpt-pilot.
Related Issues (20)
- [BUG]:A lot of times it adds lines to the beginning and the end of a source file, making it unuseable without manual editing. HOT 2
- [Howto]: Mitigate the "INFO: Received incomplete JSON response from LLM. Asking for the rest..." Loop? HOT 2
- [Please i'm stucked at the Tech Lead Agent try to create a development ]: HOT 5
- [Bug]: Using the new GPT-4o as the LLM model makes Pythagora sometimes spew out and endless loop which stops the next agent from being able to do anything HOT 1
- [Bug]: When attempting to use 'skip_until_dev_step' or just 'step' → No change. 😢 [Bug]: HOT 2
- I can't create account, it report read ECONNRESET
- [Howto]: RAG function
- [Bug]: Pilot agent tries to create a feature that I discarded in the middle of the implementation
- [Enhancement]: Search Engine usage
- zApper
- [Howto]: prevent prompts via CLI
- [Howto]: How can debug GPT-Pilot with VS Code extension? HOT 1
- [Howto]: Where can I find the vs code ext. source code ?Does it Open Source?
- [Bug]: Azure Open AI Doesnt work inside config.json - Version 0.2.1 HOT 2
- [Bug]: Configuring Pythagora with Together.ai API
- [Bug]: "streamlit run app.py" gpt-pilot cannot recognize successful..
- Django app not able to access virtual environment [Bug]:
- AI
- [Howto]: 为什么在执行任务的途中不能实时根据情况对话呢?及时纠正呢?比如在架构生成的时候,我中途想指定具体的框架。
- Stuck at "Waiting" When Try to Create New App
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from gpt-pilot.