Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
161 commits
Select commit Hold shift + click to select a range
e8fdd78
[docstring] Fix docstring for `LlamaConfig` (#26685)
pavaris-pm Oct 10, 2023
975003e
fix a typo in flax T5 attention - attention_mask variable is misnamed…
giganttheo Oct 10, 2023
3eceaa3
Fix source_prefix default value (#26654)
jheitmann Oct 10, 2023
fc63914
[JAX] Replace uses of `jnp.array` in types with `jnp.ndarray`. (#26703)
Oct 10, 2023
1e3c9dd
Make Whisper Encoder's sinusoidal PE non-trainable by default (#26032)
gau-nernst Oct 11, 2023
dcc49d8
In assisted decoding, pass model_kwargs to model's forward call (fix …
sinking-point Oct 11, 2023
9f40639
Update docs to explain disabling callbacks using report_to (#26155)
bgub Oct 11, 2023
5334796
`Copied from` for test files (#26713)
ydshieh Oct 11, 2023
da69de1
[Assistant Generation] Improve Encoder Decoder (#26701)
patrickvonplaten Oct 11, 2023
cc44ca8
[docstring] `SwinModel` docstring fix (#26679)
shivanandmn Oct 11, 2023
69873d5
fix the model card issue as `use_cuda_amp` is no more available (#26731)
pacman100 Oct 11, 2023
6ecb2ab
Fix stale bot for locked issues (#26711)
LysandreJik Oct 11, 2023
1d6a847
Fix checkpoint path in `no_trainer` scripts (#26733)
muellerzr Oct 11, 2023
b219ae6
Update docker files to use `torch==2.1.0` (#26735)
ydshieh Oct 11, 2023
e58cbed
Revert #20715 (#26734)
ydshieh Oct 11, 2023
aaccf18
[docstring] Fix docstring for `LlamaTokenizer` and `LlamaTokenizerFas…
minhoryang Oct 11, 2023
797a1ba
[docstring] Fix docstring for `CodeLlamaTokenizer` (#26709)
Bojun-Feng Oct 11, 2023
9b7668c
add japanese documentation (#26138)
rajveer43 Oct 11, 2023
e1cec43
Translated the accelerate.md file of the documentation to Chinese (#2…
liteli1987gmail Oct 11, 2023
3bc6550
Fix doctest for `Blip2ForConditionalGeneration` (#26737)
ydshieh Oct 12, 2023
40ea9ab
Add many missing spaces in adjacent strings (#26751)
tomaarsen Oct 12, 2023
ab0ddc9
Warnings controlled by logger level (#26527)
LysandreJik Oct 12, 2023
72256bc
Fix `PersimmonIntegrationTest` OOM (#26750)
ydshieh Oct 12, 2023
db5e0c3
Fix `MistralIntegrationTest` OOM (#26754)
ydshieh Oct 12, 2023
57632bf
Fix backward compatibility of Conversation (#26741)
wdhorton Oct 12, 2023
0ebee8b
[docs] LLM prompting guide (#26274)
MKhalusova Oct 12, 2023
eb734e5
[docstring] Fix `UniSpeech`, `UniSpeechSat`, `Wav2Vec2ForCTC` (#26664)
gizemt Oct 12, 2023
b4199c2
[docstring] Update `GPT2` and `Whisper` (#26642)
McDonnellJoseph Oct 12, 2023
33df09e
[docstring] Fix docstring for 'BertGenerationConfig' (#26661)
AdwaitSalankar Oct 12, 2023
a243cdc
Fix `PerceiverModelIntegrationTest::test_inference_masked_lm` (#26760)
ydshieh Oct 12, 2023
883ed4b
chore: fix typos (#26756)
afuetterer Oct 12, 2023
3e93dd2
Skip `TrainerIntegrationFSDP::test_basic_run_with_cpu_offload` if `to…
ydshieh Oct 12, 2023
7790943
🌐 [i18n-KO] Translated `big_models.md` to Korean (#26245)
wonhyeongseo Oct 12, 2023
21da3b2
Update expect outputs of `IdeficsProcessorTest.test_tokenizer_padding…
ydshieh Oct 13, 2023
d085662
[docstring] Fix docstring for `RwkvConfig` (#26782)
Bojun-Feng Oct 13, 2023
288bf5c
Fix num. of minimal calls to the Hub with peft for pipeline (#26385)
ydshieh Oct 13, 2023
5bfda28
[docstring] fix docstring `DPRConfig` (#26674)
AVAniketh0905 Oct 13, 2023
6df9179
[`core`] Fix fa-2 import (#26785)
younesbelkada Oct 13, 2023
c9785d9
Disable default system prompt for LLaMA (#26765)
Rocketknight1 Oct 13, 2023
bdb391e
Fix Falcon generation test (#26770)
Rocketknight1 Oct 13, 2023
762af3e
Add OWLv2, bis (#26668)
NielsRogge Oct 13, 2023
8e05ad3
Fixed KeyError for Mistral (#26682)
MatteoRaso Oct 13, 2023
7cc6f82
[`Flava`] Fix flava doc (#26789)
younesbelkada Oct 13, 2023
d6e5b02
Add CLIP resources (#26534)
eenzeenee Oct 13, 2023
21dc585
translation brazilian portuguese (#26769)
alvarorichard Oct 13, 2023
0dd58d9
Fixed typos (#26810)
Zhreyu Oct 16, 2023
0e52af4
[docstring] Fix docstring for `CanineConfig` (#26771)
Sparty Oct 16, 2023
69a26c7
Add Japanese translation (#26799)
shinshin86 Oct 16, 2023
5c081e2
[docstring] Fix docstring for `CodeLlamaTokenizerFast` (#26666)
Bojun-Feng Oct 16, 2023
5d997f2
Image-to-Image Task Guide (#26595)
merveenoyan Oct 16, 2023
a5f5568
Make fsdp ram efficient loading optional (#26631)
pacman100 Oct 16, 2023
b91cff5
fix resume_from_checkpoint bug (#26739)
Jintao-Huang Oct 16, 2023
570b3f9
[OWL-ViT, OWLv2] Add resources (#26822)
NielsRogge Oct 16, 2023
805d5d2
Add LLM doc (#26058)
patrickvonplaten Oct 16, 2023
3ef7134
Llama tokenizer: remove space in template comment (#26788)
pcuenca Oct 16, 2023
12cc123
Better way to run AMD CI with different flavors (#26634)
ydshieh Oct 16, 2023
5c6b83c
[docstring] Fix bert generation tokenizer (#26820)
przemL Oct 16, 2023
14b04b4
Conversation pipeline fixes (#26795)
Rocketknight1 Oct 16, 2023
fd6a0ad
🚨🚨🚨 [`Quantization`] Store the original dtype in the config as a priv…
younesbelkada Oct 16, 2023
b8f1cde
Fix Mistral OOM again (#26847)
ydshieh Oct 16, 2023
b3961f7
Chore: Typo fixed in multiple files of docs/source/en/model_doc (#26833)
Oct 17, 2023
85e9d64
fix: when window_size is passes as array (#26800)
dotneet Oct 17, 2023
0b8604d
Update logits_process.py docstrings to clarify penalty and reward cas…
larekrow Oct 17, 2023
4b423e6
🚨🚨 Generate: change order of ops in beam sample to avoid nans (#26843)
gante Oct 17, 2023
41c42f8
[`FA2`] Fix flash attention 2 fine-tuning with Falcon (#26852)
younesbelkada Oct 17, 2023
db611aa
🚨 🚨 Raise error when no speaker embeddings in speecht5._generate_spe…
ylacombe Oct 17, 2023
51042ae
[docstring] Fix docstring for LukeConfig (#26858)
louietouie Oct 17, 2023
46092f7
Fixed a typo in mistral.md (#26879)
DTennant Oct 17, 2023
b002353
Translating `en/internal` folder docs to Japanese 🇯🇵 (#26747)
rajveer43 Oct 17, 2023
ef42cb6
Fix TensorFlow pakage check (#26842)
jayfurmanek Oct 17, 2023
e893b1e
Generate: improve docstrings for custom stopping criteria (#26863)
gante Oct 18, 2023
6d644d6
Bump urllib3 from 1.26.17 to 1.26.18 in /examples/research_projects/v…
dependabot[bot] Oct 18, 2023
bece55d
Bump urllib3 from 1.26.17 to 1.26.18 in /examples/research_projects/d…
dependabot[bot] Oct 18, 2023
280c757
Knowledge distillation for vision guide (#25619)
merveenoyan Oct 18, 2023
34678db
Fix Seq2seqTrainer decoder attention mask (#26841)
Rocketknight1 Oct 18, 2023
ef7e936
[`Tokenizer`] Fix slow and fast serialization (#26570)
ArthurZucker Oct 18, 2023
de55ead
Emergency PR to skip conversational tests to fix CI (#26906)
Rocketknight1 Oct 18, 2023
d933818
Add default template warning (#26637)
Rocketknight1 Oct 18, 2023
eec5a3a
Refactor code part in documentation translated to japanese (#26900)
rajveer43 Oct 18, 2023
732d2a8
[i18n-ZH] Translated fast_tokenizers.md to Chinese (#26910)
yyLeaves Oct 18, 2023
5a73316
[`FA-2`] Final fix for FA2 dtype (#26846)
younesbelkada Oct 18, 2023
caa0ff0
Add fuyu model (#26911)
molbap Oct 18, 2023
574a538
[`FA-2`] Revert suggestion that broke FA2 fine-tuning with quantized …
younesbelkada Oct 18, 2023
816c223
[docstring] Fix docstring for `ChineseCLIP` (#26880)
Sparty Oct 19, 2023
734dd96
[Docs] Make sure important decode and generate method are nicely disp…
patrickvonplaten Oct 19, 2023
bdbcd5d
Fix and re-enable ConversationalPipeline tests (#26907)
Rocketknight1 Oct 19, 2023
ad08137
[docstring] Fix docstrings for `CodeGen` (#26821)
daniilgaltsev Oct 19, 2023
73dc23f
Fix license (#26931)
MedAymenF Oct 19, 2023
cbd278f
Pin Keras for now (#26904)
Rocketknight1 Oct 19, 2023
bc4bbd9
[`FA-2` / `Mistral`] Supprot fa-2 + right padding + forward (#26912)
younesbelkada Oct 19, 2023
ae4fb84
Generate: update basic llm tutorial (#26937)
gante Oct 19, 2023
08a2edf
Corrected modalities description in README_ru.md (#26913)
letohx Oct 19, 2023
929134b
[docstring] Fix docstring for speech-to-text config (#26883)
R055A Oct 20, 2023
9b19766
fix set_transform link docs (#26856)
diegulio Oct 20, 2023
c030fc8
Fix Fuyu image scaling bug (#26918)
pcuenca Oct 20, 2023
224794b
Update README_hd.md (#26872)
biswabaibhab007 Oct 20, 2023
093848d
Added Telugu [te] translations (#26828)
hakunamatata1997 Oct 20, 2023
f71c9cc
fix logit-to-multi-hot conversion in example (#26936)
ranchlai Oct 23, 2023
7003294
Limit to inferior fsspec version (#27010)
LysandreJik Oct 23, 2023
4542566
python falcon doc-string example typo (#26995)
SoyGema Oct 23, 2023
ef978d0
skip two tests (#27013)
ArthurZucker Oct 23, 2023
d33d313
Nits in Llama2 docstring (#26996)
osanseviero Oct 23, 2023
50d0cf4
Change default `max_shard_size` to smaller value (#26942)
younesbelkada Oct 23, 2023
cb45f71
Add Seamless M4T model (#25693)
ylacombe Oct 23, 2023
244a53e
[`NLLB-MoE`] Fix NLLB MoE 4bit inference (#27012)
younesbelkada Oct 23, 2023
f9f27b0
[`SeamlessM4T`] fix copies with NLLB MoE int8 (#27018)
ArthurZucker Oct 23, 2023
c0b5ad9
small typos found (#26988)
rafaelpadilla Oct 23, 2023
f7354a3
Remove token_type_ids from default TF GPT-2 signature (#26962)
Rocketknight1 Oct 23, 2023
f09a081
Translate `pipeline_tutorial.md` to chinese (#26954)
jiaqiw09 Oct 23, 2023
33f98cf
Remove ambiguous `padding_mask` and instead use a 2D->4D Attn Mask Ma…
patrickvonplaten Oct 23, 2023
19ae050
🌐 [i18n-ZH] Translate multilingual into Chinese (#26935)
yyLeaves Oct 23, 2023
b0d1d7f
translate `preprocessing.md` to Chinese (#26955)
jiaqiw09 Oct 23, 2023
f370beb
Bugfix device map detr model (#26849)
pedrogengo Oct 23, 2023
25c022d
Fix little typo (#27028)
mertyyanik Oct 23, 2023
32f799d
🌐 [i18n-ZH] Translate create_a_model.md into Chinese (#27026)
yyLeaves Oct 23, 2023
ede051f
Fix key dtype in GPTJ and CodeGen (#26836)
fxmarty Oct 24, 2023
cc7803c
Register ModelOutput as supported torch pytree nodes (#26618)
XuehaiPan Oct 24, 2023
fc142bd
Add `default_to_square_for_size` to `CLIPImageProcessor` (#26965)
ydshieh Oct 24, 2023
576e282
Add descriptive docstring to WhisperTimeStampLogitsProcessor (#25642)
jprivera44 Oct 24, 2023
e2d6d5c
Normalize only if needed (#26049)
mjamroz Oct 24, 2023
7bde5d6
[`TFxxxxForSequenceClassifciation`] Fix the eager mode after #25085 (…
ArthurZucker Oct 24, 2023
cb0c680
Safe import of rgb_to_id from FE modules (#27037)
amyeroberts Oct 24, 2023
b18e314
add info on TRL docs (#27024)
lvwerra Oct 24, 2023
41496b9
Add fuyu device map (#26949)
SunMarc Oct 24, 2023
9da4517
Device agnostic testing (#25870)
vvvm23 Oct 24, 2023
13ef14e
Fix config silent copy in from_pretrained (#27043)
patrickvonplaten Oct 24, 2023
9333bf0
[docs] Performance docs refactor p.2 (#26791)
MKhalusova Oct 24, 2023
a0fd344
Add a default decoder_attention_mask for EncoderDecoderModel during t…
hackyon Oct 24, 2023
6cbc136
Fix RoPE config validation for FalconConfig + various config typos (#…
tomaarsen Oct 24, 2023
9286f0a
Skip-test (#27062)
ArthurZucker Oct 25, 2023
06e782d
[`core`] Refactor of `gradient_checkpointing` (#27020)
younesbelkada Oct 25, 2023
0baa924
Fix TypicalLogitsWarper tensor OOB indexing edge case (#26579)
njhill Oct 25, 2023
a64f8c1
[docstring] fix incorrect llama docstring: encoder -> decoder (#27071)
ztjhz Oct 25, 2023
ba073ea
[DOCS] minor fixes in README.md (#27048)
Akash190104 Oct 25, 2023
c34c50c
[`docs`] Add `MaskGenerationPipeline` in docs (#27063)
younesbelkada Oct 25, 2023
ba5144f
🌐 [i18n-ZH] Translate custom_models.md into Chinese (#27065)
yyLeaves Oct 25, 2023
a2f55a6
Hindi translation of pipeline_tutorial.md (#26837)
AaryaBalwadkar Oct 25, 2023
df2eebf
Handle unsharded Llama2 model types in conversion script (#27069)
coreyhu Oct 26, 2023
9c5240a
Bump werkzeug from 2.2.3 to 3.0.1 in /examples/research_projects/deci…
dependabot[bot] Oct 26, 2023
3c26924
Bump urllib3 from 1.26.17 to 1.26.18 in /examples/research_projects/l…
dependabot[bot] Oct 26, 2023
9041240
Bring back `set_epoch` for Accelerate-based dataloaders (#26850)
muellerzr Oct 26, 2023
efba1a1
Bump`flash_attn` version to `2.1` (#27079)
younesbelkada Oct 26, 2023
fe2877c
Remove unneeded prints in modeling_gpt_neox.py (#27080)
younesbelkada Oct 26, 2023
15cd096
Create SECURITY.md
ArthurZucker Oct 26, 2023
4864d08
Add-support for commit description (#26704)
ArthurZucker Oct 26, 2023
d7cb5e1
[Llama FA2] Re-add _expand_attention_mask and clean a couple things (…
patrickvonplaten Oct 26, 2023
8214d6e
add exllamav2 arg (#26437)
SunMarc Oct 26, 2023
1892592
Correct docstrings and a typo in comments (#27047)
lewis-yeung Oct 26, 2023
34a6406
Save TB logs as part of push_to_hub (#27022)
muellerzr Oct 26, 2023
6f31601
Added huggingface emoji instead of the markdown format (#27091)
shettyvarshaa Oct 26, 2023
aa4198a
[`T5Tokenizer`] Fix fast and extra tokens (#27085)
ArthurZucker Oct 27, 2023
90ee9ce
Revert "add exllamav2 arg" (#27102)
ArthurZucker Oct 27, 2023
e2bffcf
Add early stopping for Bark generation via logits processor (#26675)
isaac-chung Oct 27, 2023
66b088f
Provide alternative when warning on use_auth_token (#27105)
Wauplin Oct 27, 2023
5be1fb6
Fix no split modules underlying modules (#27090)
SunMarc Oct 27, 2023
ffff9e7
[`core`/ `gradient_checkpointing`] Refactor GC - part 2 (#27073)
younesbelkada Oct 27, 2023
29c74f5
fix detr device map (#27089)
SunMarc Oct 27, 2023
ac58937
[Attention Mask] Refactor all encoder-decoder attention mask (#27086)
patrickvonplaten Oct 27, 2023
96f9e78
Added Telugu [te] translation for README.md in main (#27077)
hakunamatata1997 Oct 27, 2023
ef23b68
translate transformers_agents.md to Chinese (#27046)
jiaqiw09 Oct 27, 2023
9e87618
Fix docstring and type hint for resize (#27104)
daniilgaltsev Oct 27, 2023
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
2 changes: 2 additions & 0 deletions .circleci/create_circleci_config.py
Original file line number Diff line number Diff line change
Expand Up @@ -127,6 +127,8 @@ def to_dict(self):
},
]
steps.extend([{"run": l} for l in self.install_steps])
steps.extend([{"run": 'pip install "fsspec>=2023.5.0,<2023.10.0"'}])
steps.extend([{"run": "pip install pytest-subtests"}])
steps.append(
{
"save_cache": {
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/build_documentation.yml
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ jobs:
commit_sha: ${{ github.sha }}
package: transformers
notebook_folder: transformers_doc
languages: de en es fr it ko pt zh
languages: de en es fr hi it ko pt zh ja te
secrets:
token: ${{ secrets.HUGGINGFACE_PUSH }}
hf_token: ${{ secrets.HF_DOC_BUILD_PUSH }}
2 changes: 1 addition & 1 deletion .github/workflows/build_pr_documentation.yml
Original file line number Diff line number Diff line change
Expand Up @@ -14,4 +14,4 @@ jobs:
commit_sha: ${{ github.event.pull_request.head.sha }}
pr_number: ${{ github.event.number }}
package: transformers
languages: de en es fr it ko pt zh
languages: de en es fr hi it ko pt zh ja te
25 changes: 25 additions & 0 deletions .github/workflows/self-push-amd-mi210-caller.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,25 @@
name: Self-hosted runner (AMD mi210 CI caller)

on:
workflow_run:
workflows: ["Self-hosted runner (push-caller)"]
branches: ["main"]
types: [completed]
push:
branches:
- run_amd_push_ci_caller*
paths:
- "src/**"
- "tests/**"
- ".github/**"
- "templates/**"
- "utils/**"

jobs:
run_amd_ci:
name: AMD mi210
if: (cancelled() != true) && ((github.event_name != 'schedule') || ((github.event_name == 'push') && startsWith(github.ref_name, 'run_amd_push_ci_caller')))
uses: ./.github/workflows/self-push-amd.yml
with:
gpu_flavor: mi210
secrets: inherit
25 changes: 25 additions & 0 deletions .github/workflows/self-push-amd-mi250-caller.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,25 @@
name: Self-hosted runner (AMD mi250 CI caller)

on:
workflow_run:
workflows: ["Self-hosted runner (push-caller)"]
branches: ["main"]
types: [completed]
push:
branches:
- run_amd_push_ci_caller*
paths:
- "src/**"
- "tests/**"
- ".github/**"
- "templates/**"
- "utils/**"

jobs:
run_amd_ci:
name: AMD mi250
if: (cancelled() != true) && ((github.event_name != 'schedule') || ((github.event_name == 'push') && startsWith(github.ref_name, 'run_amd_push_ci_caller')))
uses: ./.github/workflows/self-push-amd.yml
with:
gpu_flavor: mi250
secrets: inherit
31 changes: 9 additions & 22 deletions .github/workflows/self-push-amd.yml
Original file line number Diff line number Diff line change
@@ -1,21 +1,11 @@
name: Self-hosted runner AMD GPU (push)

on:
workflow_run:
workflows: ["Self-hosted runner (push-caller)"]
branches: ["main"]
types: [completed]
push:
branches:
- ci_*
- ci-*
paths:
- "src/**"
- "tests/**"
- ".github/**"
- "templates/**"
- "utils/**"
repository_dispatch:
workflow_call:
inputs:
gpu_flavor:
required: true
type: string

env:
HF_HOME: /mnt/cache
Expand Down Expand Up @@ -45,8 +35,7 @@ jobs:
strategy:
matrix:
machine_type: [single-gpu, multi-gpu]
gpu_flavor: [mi210]
runs-on: [self-hosted, docker-gpu, amd-gpu, '${{ matrix.machine_type }}', '${{ matrix.gpu_flavor }}']
runs-on: [self-hosted, docker-gpu, amd-gpu, '${{ matrix.machine_type }}', '${{ inputs.gpu_flavor }}']
container:
image: huggingface/transformers-pytorch-amd-gpu-push-ci # <--- We test only for PyTorch for now
options: --device /dev/kfd --device /dev/dri --env HIP_VISIBLE_DEVICES --env ROCR_VISIBLE_DEVICES --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/
Expand All @@ -65,8 +54,7 @@ jobs:
strategy:
matrix:
machine_type: [single-gpu, multi-gpu]
gpu_flavor: [mi210]
runs-on: [self-hosted, docker-gpu, amd-gpu, '${{ matrix.machine_type }}', '${{ matrix.gpu_flavor }}']
runs-on: [self-hosted, docker-gpu, amd-gpu, '${{ matrix.machine_type }}', '${{ inputs.gpu_flavor }}']
container:
image: huggingface/transformers-pytorch-amd-gpu-push-ci # <--- We test only for PyTorch for now
options: --device /dev/kfd --device /dev/dri --env HIP_VISIBLE_DEVICES --env ROCR_VISIBLE_DEVICES --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/
Expand Down Expand Up @@ -164,8 +152,7 @@ jobs:
matrix:
folders: ${{ fromJson(needs.setup_gpu.outputs.matrix) }}
machine_type: [single-gpu, multi-gpu]
gpu_flavor: [mi210]
runs-on: [self-hosted, docker-gpu, amd-gpu, '${{ matrix.machine_type }}', '${{ matrix.gpu_flavor }}']
runs-on: [self-hosted, docker-gpu, amd-gpu, '${{ matrix.machine_type }}', '${{ inputs.gpu_flavor }}']
container:
image: huggingface/transformers-pytorch-amd-gpu-push-ci # <--- We test only for PyTorch for now
options: --device /dev/kfd --device /dev/dri --env HIP_VISIBLE_DEVICES --env ROCR_VISIBLE_DEVICES --shm-size "16gb" --ipc host -v /mnt/cache/.cache/huggingface:/mnt/cache/
Expand Down Expand Up @@ -321,7 +308,7 @@ jobs:
CI_SLACK_CHANNEL_DUMMY_TESTS: ${{ secrets.CI_SLACK_CHANNEL_DUMMY_TESTS }}
CI_SLACK_REPORT_CHANNEL_ID: ${{ secrets.CI_SLACK_CHANNEL_ID_AMD }}
ACCESS_REPO_INFO_TOKEN: ${{ secrets.ACCESS_REPO_INFO_TOKEN }}
CI_EVENT: push
CI_EVENT: Push CI (AMD) - ${{ inputs.gpu_flavor }}
CI_TITLE_PUSH: ${{ github.event.head_commit.message }}
CI_TITLE_WORKFLOW_RUN: ${{ github.event.workflow_run.head_commit.message }}
CI_SHA: ${{ env.CI_SHA }}
Expand Down
Loading