Llama3 CoEdit

This is a Llama3 8B based model trained using torchtune on the grammarly/coedit dataset.

Training details

The exact training script (lora_finetune_distributed) and config (8B_lora.yaml) are both included in this repository.

Training command: tune run --nproc_per_node 8 lora_finetune_distributed --config 8B_lora.yaml

Yes I used 8 GPUs :)

In order to add the dataset, I added the following lines to the config:

dataset:
  _component_: torchtune.datasets.instruct_dataset
  source: grammarly/coedit
  template: GrammarErrorCorrectionTemplate
  column_map: {"sentence": "src", "output": "tgt"}
  train_on_input: False
  split: train

Loss curve

image/png

Evaluation results

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Dataset used to train jrc/llama3-8b-coedit