|
--- |
|
license: llama2 |
|
language: |
|
- en |
|
library_name: transformers |
|
pipeline_tag: text-generation |
|
model-index: |
|
- name: CodeMate-v0.1 |
|
results: |
|
- task: |
|
type: text-generation |
|
dataset: |
|
name: HumanEval |
|
type: openai_humaneval |
|
metrics: |
|
- name: pass@1 |
|
type: pass@1 |
|
value: 74.9% |
|
verified: false |
|
tags: |
|
- CodeMate |
|
- Code |
|
- CodeLLaMa |
|
--- |
|
|
|
|
|
# **CodeMate-v0.1** |
|
|
|
CodeMate-v0.1 is an intelligent programming assistant developed by [CodeMate](https://codemate.ai). |
|
This model aims to assist users in generating high-quality code solutions for programming problems. |
|
Please note that this model is currently in version 0.1. |
|
|
|
## Model Details |
|
|
|
- **Training Data:** Exclusively fine-tuned on a proprietary dataset of 1.8 billion tokens of high-quality programming problems and solutions. |
|
- The dataset was generated manually and is internal to CodeMate. |
|
|
|
- **Training Techniques:** The model was fine-tuned using Flash Attention 2, trained over 15 hours on 40 A100-80GB GPUs. |
|
- A sequence length of 8096 tokens was used during training. |
|
|
|
- **Multilingual Support:** CodeMate-v0.1 is proficient in multiple programming languages, including Python, C/C++, TypeScript, Java, and more. |
|
|
|
## How to Get Started with the Model |
|
|
|
Make sure to install Transformers from the main git branch: |
|
|
|
```bash |
|
pip install git+https://github.com/huggingface/transformers.git |
|
``` |
|
|
|
## How to Prompt the Model |
|
|
|
This model accepts prompts in the Alpaca/Vicuna instruction format. For example: |
|
|
|
```markdown |
|
### System Prompt |
|
You are an intelligent programming assistant. |
|
|
|
### User Message |
|
Implement a linked list in C++ |
|
|
|
### Assistant |
|
... |
|
``` |
|
|
|
## Load the Model: |
|
|
|
To load the model, utilize the following Python script: |
|
|
|
```python |
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
|
# Initialize the model |
|
model_path = "codemateai/CodeMate-v0.1" |
|
model = AutoModelForCausalLM.from_pretrained(model_path, device_map="auto") |
|
tokenizer = AutoTokenizer.from_pretrained(model_path) |
|
|
|
# ... generate response ... |
|
|
|
``` |
|
|
|
## Bias, Risks, and Limitations |
|
|
|
This model has undergone very limited testing. CodeMate recommends additional safety testing before any real-world deployments. |
|
|
|
For more information and updates, visit the [CodeMate website](https://codemate.ai). |