Skip to content

Instantly share code, notes, and snippets.

@lmBored
Created March 2, 2026 14:25
Show Gist options
  • Select an option

  • Save lmBored/65cfb39ef096682d2cd2e1fc523573a4 to your computer and use it in GitHub Desktop.

Select an option

Save lmBored/65cfb39ef096682d2cd2e1fc523573a4 to your computer and use it in GitHub Desktop.
#!/bin/bash
#SBATCH --job-name=comrad
#SBATCH --partition=mcs.gpu.q
#SBATCH --gres=gpu:1
#SBATCH --cpus-per-task=16
#SBATCH --mem=32G
#SBATCH --time=6:12:00
#SBATCH --output=/home/20231193/ViZDoom/train_dir/armory_siege_QMIX_20260302_081241_j%j/slurm_%j.out
#SBATCH --error=/home/20231193/ViZDoom/train_dir/armory_siege_QMIX_20260302_081241_j%j/slurm_%j.err
#SBATCH --mail-type=BEGIN,END,FAIL
#SBATCH --mail-user=mhjjj2113@gmail.com
# ======================================================================
set -e
COMRAD_RUN_NAME="armory_siege_QMIX_20260302_081241_j${SLURM_JOB_ID}"
export COMRAD_RUN_NAME
echo "$SLURM_JOB_ID" > "/home/20231193/ViZDoom/train_dir/${COMRAD_RUN_NAME}/slurm_job_id"
echo "Start: $(date)"
echo "Node: $SLURM_NODELIST"
echo "Partition: $SLURM_JOB_PARTITION"
echo "GPUs: $SLURM_GPUS_ON_NODE"
echo "CPUs: $SLURM_CPUS_PER_TASK"
echo "Memory: 32G"
# ======================================================================
module purge
module load GCC/12.3.0
module load cuda12.2/toolkit/12.2.1
module load uv
module load CMake/3.26.3-GCCcore-12.3.0
module load Boost/1.83.0-GCC-12.3.0
module load SDL2/2.28.2-GCCcore-12.3.0
export CMAKE_PREFIX_PATH=$HOME/.local:$CMAKE_PREFIX_PATH
export PKG_CONFIG_PATH=$HOME/.local/lib/pkgconfig:$HOME/.local/lib64/pkgconfig:$PKG_CONFIG_PATH
export LD_LIBRARY_PATH=$HOME/.local/lib:$HOME/.local/lib64:$LD_LIBRARY_PATH
cd /home/20231193/ViZDoom
# python 3.11
source /home/20231193/ViZDoom/.venv311/bin/activate
uv sync --active -p /home/20231193/ViZDoom/.venv311
echo "python: "
python --version
# source ~/miniconda3/etc/profile.d/conda.sh
# conda activate vizdoom
python -m comrad.train --env=armory_siege --algo=QMIX --mixer=qmix --train_for_env_steps=100000 --num_workers=4 --num_envs_per_worker=4 --policy_workers_per_policy=2 --batch_size=2048 --env_frameskip=4 --wide_aspect_ratio=False --with_wandb=True --wandb_dir=. --wandb_project=marl_vizdoom --use_rnn=True --rnn_type=gru --rnn_size=256 --rollout=32 --gamma=0.99 --learning_starts=50000 --qmix_sequence_batch_size=64 --replay_buffer_size=500000 --epsilon_decay_steps=20000000 --epsilon_end=0.005 --learning_rate=0.0001 --num_agents=2 --dqn_max_updates_per_batch=4 --target_update_tau=0.005 --use_huber_loss=True --q_value_clamp=100 --train_frequency=8 --batched_sampling=True --actor_critic_share_weights=True --per=False --experiment=$COMRAD_RUN_NAME
EXIT_CODE=$?
echo "Done: $(date)"
exit $EXIT_CODE
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment