# | |
# Infers log-likelihoods from an HMM and writes results to a CSV file. | |
# | |
#SBATCH --cluster=<clustername> | |
#SBATCH --partition=<partitionname> | |
#SBATCH --account=<accountname> | |
#SBATCH --job-name=hmminf | |
#SBATCH --output=logs/hmminf.out | |
#SBATCH --gres=gpu:0 # Number of GPU(s) per node. | |
#SBATCH --cpus-per-task=1 # CPU cores/threads | |
#SBATCH --mem=4000M # memory per node | |
#SBATCH --time=0-24:00 # Max time (DD-HH:MM) | |
#SBATCH --ntasks=1 # Only set to >1 if you want to use multi-threading | |
dataset=$1 | |
hmmpath=$2 | |
runid=$3 | |
dir="inference/${dataset}/hmm" | |
mkdir -p $dir | |
hmmsearch --tblout ${dir}/${runid}.tblout $hmmpath data/${dataset}/seqs.fasta | |
python scripts/tblout2csv.py ${dir}/${runid}.tblout ${dir}/${runid}.csv | |