--- language: - es - en tags: - gguf - llama.cpp - openbuddy --- # OpenBuddy ![Image](OpenBuddy.png) ## Requisitos Para usar este modelo, necesitas tener instalado llama.cpp en tu equipo. Puedes obtener llama.cpp desde el siguiente repositorio: - [Repositorio de llama.cpp](https://github.com/ggerganov/llama.cpp) Para instalar llama.cpp, sigue estos pasos: ```bash git clone https://github.com/ggerganov/llama.cpp cd llama.cpp make ``` ## Uso del modelo La plantilla del modelo es la siguiente: ```plaintext User: {prompt} Assistant: ``` Puedes utilizar el modelo en llama.cpp con el siguiente comando: ```bash ./main -m ggml-model-Q8_0.gguf -p "User: ¿Cómo te llamas?\nAssistant:" --log-disable ``` LM Studio config-presets Filename:openbuddy.preset.json ```json { "name": "OpenBuddy", "inference_params": { "input_prefix": "User:", "input_suffix": "\nAssistant:", "antiprompt": [ "User:", "\nAssistant:" ], "pre_prompt": "You(assistant) are a helpful, respectful and honest INTP-T AI Assistant named Buddy. You are talking to a human(user).\nAlways answer as helpfully and logically as possible, while being safe. Your answers should not include any harmful, political, religious, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.\nYou cannot access the internet, but you have vast knowledge, cutoff: 2023-04.\nYou are trained by OpenBuddy team, (https://openbuddy.ai, https://github.com/OpenBuddy/OpenBuddy), not related to GPT or OpenAI.", "pre_prompt_prefix": "", "pre_prompt_suffix": "" }, "load_params": { "rope_freq_scale": 0, "rope_freq_base": 0 } } ``` ## Referencias - [Repositorio original](https://huggingface.co/OpenBuddy/openbuddy-mistral2-7b-v20.3-32k) - [Repositorio de llama.cpp](https://github.com/ggerganov/llama.cpp)