Skip to content

remove dependency of flash_attn when use_flash_attn is set to false #132

remove dependency of flash_attn when use_flash_attn is set to false

remove dependency of flash_attn when use_flash_attn is set to false #132

Workflow file for this run

name: e2e-tests
on:
pull_request:
branches:
- "develop"
paths-ignore:
- "doc/**"
- "**.md"
env:
SLURM_PARTITION: llm_s
jobs:
training_8GPU:
runs-on: [t_cluster]
timeout-minutes: 10
steps:
- name: mask env
run: |
echo "::add-mask::${{env.WORKSPACE_PREFIX}}"
echo "::add-mask::$path_prefix"
- uses: actions/checkout@v3
- name: training_8GPU
run: |
source $evo_env
srun -p ${SLURM_PARTITION} --kill-on-bad-exit=1 --job-name=${GITHUB_RUN_ID}-${GITHUB_JOB} -n8 --ntasks-per-node=8 --cpus-per-task=4 --gpus-per-task=1 pytest -s -v --color=yes -m "training_8GPU" ./tests/test_training/test_loss.py
training_8GPU_ISP:
runs-on: [t_cluster]
timeout-minutes: 10
steps:
- name: mask env
run: |
echo "::add-mask::${{env.WORKSPACE_PREFIX}}"
echo "::add-mask::$path_prefix"
- uses: actions/checkout@v3
- name: training_8GPU_ISP
run: |
source $evo_env
conda activate /mnt/petrelfs/share_data/huangting.p/envs/llm-torch2.1-flash2
conda activate /mnt/petrelfs/share_data/huangting.p/envs/llm-torch2.1-flash2
srun -p ${SLURM_PARTITION} --kill-on-bad-exit=1 --job-name=${GITHUB_RUN_ID}-${GITHUB_JOB} -n8 --ntasks-per-node=8 --cpus-per-task=4 --gpus-per-task=1 pytest -s -v --color=yes -m "training_8GPU_ISP" ./tests/test_training/test_loss.py