instruction finetune README improvement (#897)
Signed-off-by: rbrugaro <rita.brugarolas.brufau@intel.com>
This commit is contained in:
@@ -1,6 +1,6 @@
|
||||
# Instruction Tuning
|
||||
|
||||
Instruction tuning is the process of further training LLMs on a dataset consisting of (instruction, output) pairs in a supervised fashion, which bridges the gap between the next-word prediction objective of LLMs and the users' objective of having LLMs adhere to human instructions.
|
||||
Instruction tuning is the process of further training LLMs on a dataset consisting of (instruction, output) pairs in a supervised fashion, which bridges the gap between the next-word prediction objective of LLMs and the users' objective of having LLMs adhere to human instructions. This implementation deploys a Ray cluster for the task.
|
||||
|
||||
## Deploy Instruction Tuning Service
|
||||
|
||||
@@ -38,6 +38,8 @@ curl http://${your_ip}:8015/v1/fine_tuning/jobs \
|
||||
}'
|
||||
```
|
||||
|
||||
The outputs of the finetune job (adapter_model.safetensors, adapter_config,json... ) are stored in `/home/user/comps/finetuning/output` and other execution logs are stored in `/home/user/ray_results`
|
||||
|
||||
### 3. Manage fine-tuning job
|
||||
|
||||
Below commands show how to list finetuning jobs, retrieve a finetuning job, cancel a finetuning job and list checkpoints of a finetuning job.
|
||||
|
||||
Reference in New Issue
Block a user