|
33 | 33 | needs: code-quality |
34 | 34 | timeout-minutes: 30 |
35 | 35 |
|
| 36 | + strategy: |
| 37 | + fail-fast: false |
| 38 | + matrix: |
| 39 | + backend: |
| 40 | + - name: opencl |
| 41 | + - name: ptx |
| 42 | + |
36 | 43 | steps: |
37 | 44 | - name: Checkout GPULlama3 |
38 | 45 | uses: actions/checkout@v4 |
|
54 | 61 | source venv/bin/activate |
55 | 62 | echo "=== Building TornadoVM ===" |
56 | 63 | |
57 | | - make |
| 64 | + make BACKEND=${{ matrix.backend.name }} |
58 | 65 | |
59 | 66 | echo "=== Searching for TornadoVM SDK directory ===" |
60 | 67 | SDK_DIR=$(find dist -type d -maxdepth 3 -path "*/tornadovm-*-opencl" | head -n 1) |
@@ -88,69 +95,69 @@ jobs: |
88 | 95 | run: | |
89 | 96 | cd ${{ github.workspace }} |
90 | 97 | export PATH="$TORNADO_SDK/bin:$JAVA_HOME/bin:$PATH" |
91 | | - ./llama-tornado --gpu --opencl \ |
| 98 | + ./llama-tornado --gpu --${{ matrix.backend.name }} \ |
92 | 99 | --model $MODELS_DIR/Llama-3.2-1B-Instruct-F16.gguf \ |
93 | 100 | --prompt "Say hello" |
94 | 101 | - name: FP16 - Run Qwen3-4B-f16.gguf |
95 | 102 | run: | |
96 | 103 | cd ${{ github.workspace }} |
97 | 104 | export PATH="$TORNADO_SDK/bin:$JAVA_HOME/bin:$PATH" |
98 | | - ./llama-tornado --gpu --opencl \ |
| 105 | + ./llama-tornado --gpu --${{ matrix.backend.name }} \ |
99 | 106 | --model $MODELS_DIR/Qwen3-4B-f16.gguf \ |
100 | 107 | --prompt "Say hello" |
101 | 108 | - name: FP16 - Run Mistral-7B-Instruct-v0.3.fp16.gguf |
102 | 109 | run: | |
103 | 110 | cd ${{ github.workspace }} |
104 | 111 | export PATH="$TORNADO_SDK/bin:$JAVA_HOME/bin:$PATH" |
105 | | - ./llama-tornado --gpu --opencl \ |
| 112 | + ./llama-tornado --gpu --${{ matrix.backend.name }} \ |
106 | 113 | --model $MODELS_DIR/Mistral-7B-Instruct-v0.3.fp16.gguf \ |
107 | 114 | --prompt "Say hello" |
108 | 115 | - name: FP16 - Run Qwen2.5-1.5b-instruct-fp16.gguf |
109 | 116 | run: | |
110 | 117 | cd ${{ github.workspace }} |
111 | 118 | export PATH="$TORNADO_SDK/bin:$JAVA_HOME/bin:$PATH" |
112 | | - ./llama-tornado --gpu --opencl \ |
| 119 | + ./llama-tornado --gpu --${{ matrix.backend.name }} \ |
113 | 120 | --model $MODELS_DIR/qwen2.5-1.5b-instruct-fp16.gguf \ |
114 | 121 | --prompt "Say hello" |
115 | 122 | - name: FP16 - Run Phi-3-mini-4k-instruct-fp16.gguf |
116 | 123 | run: | |
117 | 124 | cd ${{ github.workspace }} |
118 | 125 | export PATH="$TORNADO_SDK/bin:$JAVA_HOME/bin:$PATH" |
119 | | - ./llama-tornado --gpu --opencl \ |
| 126 | + ./llama-tornado --gpu --${{ matrix.backend.name }} \ |
120 | 127 | --model /$MODELS_DIR/Phi-3-mini-4k-instruct-fp16.gguf \ |
121 | 128 | --prompt "Say hello" |
122 | 129 | - name: Q8 - Run Llama-3.2-1B-Instruct-Q8_0.gguf |
123 | 130 | run: | |
124 | 131 | cd ${{ github.workspace }} |
125 | 132 | export PATH="$TORNADO_SDK/bin:$JAVA_HOME/bin:$PATH" |
126 | | - ./llama-tornado --gpu --opencl \ |
| 133 | + ./llama-tornado --gpu --${{ matrix.backend.name }} \ |
127 | 134 | --model $MODELS_DIR/Llama-3.2-1B-Instruct-Q8_0.gguf \ |
128 | 135 | --prompt "Say hello" |
129 | 136 | - name: Q8 - Run Qwen3-0.6B-Q8_0.gguf |
130 | 137 | run: | |
131 | 138 | cd ${{ github.workspace }} |
132 | 139 | export PATH="$TORNADO_SDK/bin:$JAVA_HOME/bin:$PATH" |
133 | | - ./llama-tornado --gpu --opencl \ |
| 140 | + ./llama-tornado --gpu --${{ matrix.backend.name }} \ |
134 | 141 | --model $MODELS_DIR/Qwen3-0.6B-Q8_0.gguf \ |
135 | 142 | --prompt "Say hello" |
136 | 143 | - name: Q8 - Run Phi-3-mini-4k-instruct-Q8_0.gguf |
137 | 144 | run: | |
138 | 145 | cd ${{ github.workspace }} |
139 | 146 | export PATH="$TORNADO_SDK/bin:$JAVA_HOME/bin:$PATH" |
140 | | - ./llama-tornado --gpu --opencl \ |
| 147 | + ./llama-tornado --gpu --${{ matrix.backend.name }} \ |
141 | 148 | --model $MODELS_DIR/Phi-3-mini-4k-instruct-Q8_0.gguf \ |
142 | 149 | --prompt "Say hello" |
143 | 150 | - name: Q8 - Run Qwen2.5-1.5b-instruct-q8_0.gguf |
144 | 151 | run: | |
145 | 152 | cd ${{ github.workspace }} |
146 | 153 | export PATH="$TORNADO_SDK/bin:$JAVA_HOME/bin:$PATH" |
147 | | - ./llama-tornado --gpu --opencl \ |
| 154 | + ./llama-tornado --gpu --${{ matrix.backend.name }} \ |
148 | 155 | --model $MODELS_DIR/qwen2.5-1.5b-instruct-q8_0.gguf \ |
149 | 156 | --prompt "Say hello" |
150 | 157 | - name: Q8 - Mistral-7B-Instruct-v0.3.Q8_0.gguf |
151 | 158 | run: | |
152 | 159 | cd ${{ github.workspace }} |
153 | 160 | export PATH="$TORNADO_SDK/bin:$JAVA_HOME/bin:$PATH" |
154 | | - ./llama-tornado --gpu --opencl \ |
| 161 | + ./llama-tornado --gpu --${{ matrix.backend.name }} \ |
155 | 162 | --model $MODELS_DIR/Mistral-7B-Instruct-v0.3.Q8_0.gguf \ |
156 | 163 | --prompt "Say hello" |
0 commit comments