add autoconfig
#34 opened 3 days ago
by
Sifal

CUDA: OutOfMemoryError
#33 opened about 1 month ago
by
8497prashant
Make compatible with recent versions of triton
#32 opened 3 months ago
by
tanhevg
Adding `safetensors` variant of this model
#31 opened 4 months ago
by
SFconvertbot

Example Notebook
#29 opened 6 months ago
by
GCabas
Encoding for long contexts
1
#28 opened 6 months ago
by
yoavkt
Fix AutoModel not loading model correctly due to config_class inconsistency
12
#26 opened 9 months ago
by
liamclarkza

Update bert_layers.py
#25 opened 9 months ago
by
ruffy369

A modified DNABERT2 that returns the attention too
#24 opened 10 months ago
by
jaandoui
Triton version
9
#23 opened 11 months ago
by
JiayiJennie
how do I output the attention scores from the last layer of the encoder?
1
#22 opened 11 months ago
by
jkb0722
Tokenization of more than 2 sequences
1
#21 opened 11 months ago
by
jaandoui
Impact of Padding on DNABERT Model Performance
#20 opened 12 months ago
by
poilkjhytg
Setting DNABERT-2 revision
#19 opened 12 months ago
by
SeanDoyle
TypeError: forward() got an unexpected keyword argument 'attention_mask'
#18 opened 12 months ago
by
jkb0722
Adding `safetensors` variant of this model
#17 opened 12 months ago
by
SFconvertbot

Assertion Error / Implementation Error
1
#16 opened 12 months ago
by
8497prashant
Extract attention from model
1
#15 opened 12 months ago
by
kaustabanv
error on inference using DNABERT2 - can you please share the environment you used for running it?
2
#12 opened about 1 year ago
by
NettaB
Adding `safetensors` variant of this model
#11 opened over 1 year ago
by
SFconvertbot

Is the datasets for foundational model pre-training publicly accessible?
#10 opened over 1 year ago
by
JayceCeleste

Expecting bi-modal disstribution of probabilities
#9 opened over 1 year ago
by
christianclough
Inference fails with output_all_encoded_layers=True.
1
#8 opened over 1 year ago
by
pg20sanger
Adding `safetensors` variant of this model
#7 opened over 1 year ago
by
SFconvertbot

The model class you are passing has a `config_class` attribute that is not consistent with the config class you passed
4
#6 opened over 1 year ago
by
saikiran7
About the output of tokenizer and the model
2
#4 opened over 1 year ago
by
RandyWang504
Despite multiple trials and examining the model configuration, it seems that the model hosted on Hugging Face (`huggingface.co`) cannot handle sequences that exceed a length of 512 tokens.
2
#3 opened over 1 year ago
by
hengchuangyin
Issue in Code, at 114:24: def _fwd_kernel(..)
4
#1 opened over 1 year ago
by
Hosna
