Comments (6)
@EmmaWebGH I just became aware of the prompting issues this afternoon and they have been solved in 0.8.7. I tested with both the "free-trial" and "together" providers.
If you're seeing template-related problems while using Ollama though, this problem might be on the Ollama side, as we rely on them to format messages. I haven't had the chance to test without having a GPU large enough for CodeLlama-70b, so let me know and I could reach out to them
from continue.
Great, I will update when I can!
You really are on top of everything. Thankyou for such a useful extension!
from continue.
Ok, sounds like this is about a typical response for CodeLlama-70b : ) which means that this should be resolved. Let me know if anything else comes up!
from continue.
I've met the same issue. I started codellama 70b instruct gguf on my mac m1 studio by llama.cpp server like this:
../llama.cpp/server -m ./codellama-70b-instruct.Q5_K_S.gguf -np 2 -c 4096 --host 0.0.0.0 --port 8080
And configured the model in config.json
{
"title": "codellama-70b",
"model": "codellama-70b",
"completionOptions": {},
"contextLength": 4096,
"provider": "llama.cpp",
"apiBase": "http://dev.myserver.com:8080"
},
On the VS Code Continue plugin, it keeps output lots of code and messages without stopping.
I wonder how could set the proper prompt template, and the stop token.
For huggingface chat-ui, the following config works:
{
"name": "codellama-70b-llamacpp",
"chatPromptTemplate" : "<s>{{#if @root.preprompt}}Source: system\n\n {{@root.preprompt}} <step> {{/if}}{{#each messages}}{{#ifUser}}Source: user\n\n {{content}} <step> {{/ifUser}}{{#ifAssistant}}Source: assistant\n\n {{content}} <step> {{/ifAssistant}}{{/each}}Source: assistant\nDestination: user\n\n ",
"parameters": {
"temperature": 0.5,
"top_p": 0.95,
"repetition_penalty": 1.2,
"top_k": 50,
"truncate": 3072,
"max_new_tokens": 2048,
"stop" : ["<step>", "Source: assistant"]
},
"endpoints": [{
"type": "openai",
"baseURL": "http://dev.myserver.com:8080/v1"
}]
}
from continue.
@davideuler Everything in your config looks right and seems to indicate that the prompt should be correctly set. here is the code where we format the prompt for codellama-70b. You could double check that the correct formatting is being sent by going to the "Output" tab in the bottom bar of VS Code (next to terminal) and then selecting "Continue - ..." in the dropdown on the right. It shows all raw prompt/completions
If this looks correct, then perhaps there might be a bad interaction with the server (e.g. it also formats the prompt, leading to it happening twice)
from continue.
@davideuler Everything in your config looks right and seems to indicate that the prompt should be correctly set. here is the code where we format the prompt for codellama-70b. You could double check that the correct formatting is being sent by going to the "Output" tab in the bottom bar of VS Code (next to terminal) and then selecting "Continue - ..." in the dropdown on the right. It shows all raw prompt/completions
If this looks correct, then perhaps there might be a bad interaction with the server (e.g. it also formats the prompt, leading to it happening twice)
Sestinj, Thanks, I've checked the output in VS Code. The request which was sent to llama.cpp is ok.
And when I am with the latest version of continue plugin, it shows me response related to the code and lots of apologizing message like "I apologize, but as a responsible AI language model".
from continue.
Related Issues (20)
- tag v0.8.31-vscode cannot be built as activate.ts has a duplicate definitions for resolveVsCodeExtension HOT 2
- Add an Open WebUI Provider HOT 3
- Extension host terminated unexpectedly
- Jetbrains plugin not working with Ollama, but VSCode plugin is working HOT 3
- Unable to reference other files
- Additional fields for dev_data/autocomplete.jsonl
- [BUG] OLLAMA Crash request HOT 4
- Json files using the @ RAG routine are being evaluated prior to being presented to the model.
- Functionality to exclude TabAutoComplete on certain extensions HOT 1
- Hang when opening tab in large codebase HOT 2
- [IntelliJ Plugin] Unauthorised when using Mistral Codestral (despite proper config which work for VSCode) HOT 5
- [CON-242] Autodetect language in code blocks
- [CON-243] systemMessage not effective for claude 3
- Incorrect autocomplete context in jupyter with IntelliSense
- Can continue support using free trial 10 times a day for free? I think this is a good solution instead of permanently limiting it to 25 times now HOT 1
- Add support provider dify.ai
- Stream as Parameter in Tab Autocomplete HOT 3
- Sample configs for diff deployments scenarios
- [Jetbrains] Error vectordb::voyage-code-2 index - failed to load native library (JetBrains Plugin)
- Delete key triggering strange behaviour from CLion tab completion suggestions
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from continue.