-
Notifications
You must be signed in to change notification settings - Fork 30
49 lines (42 loc) · 1.33 KB
/
workflow_finetune_gpu.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
name: Finetune-GPU
on:
workflow_call:
inputs:
runner_container_image:
type: string
default: '10.1.2.13:5000/llmray-build'
http_proxy:
type: string
default: 'http://proxy-prc.intel.com:912'
https_proxy:
type: string
default: 'http://proxy-prc.intel.com:912'
permissions: # added using https://github.com/step-security/secure-repo
contents: read
jobs:
finetune-gpu:
name: finetune-gpu
strategy:
matrix:
model: [ NousResearch/Llama-2-7b-chat-hf ]
runs-on: self-hosted
defaults:
run:
shell: bash
container:
image: ${{ inputs.runner_container_image }}
env:
http_proxy: ${{ inputs.http_proxy }}
https_proxy: ${{ inputs.https_proxy }}
SHELL: bash -eo pipefail
volumes:
- /var/run/docker.sock:/var/run/docker.sock
steps:
- name: Checkout
uses: actions/checkout@692973e3d937129bcbf40652eb9f2f61becf3332 # v4.1.7
- name: Running task on Intel GPU
run: |
rm /home/ci/borealis-runner/llm-on-ray.tar.gz -f
tar zcf /home/ci/borealis-runner/llm-on-ray.tar.gz -C /home/ci/actions-runner/_work/llm-on-ray .
cd /home/ci/borealis-runner/
python3 finetune_on_pvc.py --need_create_conda_env true --base_models "${{ matrix.model }}"