Björn Plüster
bjoernp
AI & ML interests
None yet
Organizations
bjoernp's activity
Can you share how you converted this?
7
#1 opened 3 months ago
by
bjoernp
Hf safetensors version
9
#3 opened 3 months ago
by
ehartford
use_flash_attention_2=True
3
#9 opened 4 months ago
by
TillFetzer
leo-mistral-hessianai-7b-chat for privateGPT
3
#8 opened 5 months ago
by
Dodo124
Update tokenizer_config.json
#1 opened 5 months ago
by
bjoernp
Problems with flash-attention2
1
#13 opened 6 months ago
by
omaer0
Loss function?
1
#10 opened 9 months ago
by
narvind2003
No multi GPU inference support?
8
#4 opened 9 months ago
by
dataautogpt3
Llama2 vs Mistral
1
#2 opened 9 months ago
by
lightningRalf
Add languages
#8 opened 9 months ago
by
lbourdois
Missing module/classes: from transformers.cache_utils import Cache, DynamicCache
1
#7 opened 9 months ago
by
panopstor
changed "tokenizer" typo to be the one we create.
#4 opened 9 months ago
by
dyngnosis
Which transformers version is being used here?
2
#6 opened 9 months ago
by
Promptengineering
Flash dependency (locks out non-NVIDIA GPUs)
3
#4 opened 9 months ago
by
Thalesian
Update modeling_moe_mistral.py
#5 opened 9 months ago
by
bjoernp
Trying to quantize. Running into the issue below. Any suggestions?
1
#5 opened 9 months ago
by
BigDeeper
small readme fix
#1 opened 10 months ago
by
jphme
Update modeling_moe_mistral.py
2
#1 opened 10 months ago
by
bjoernp
AWQ-Variante
4
#2 opened 10 months ago
by
SebastianBodza
Little Mistake :)
1
#1 opened 10 months ago
by
DRXD1000
Can you incorporate madlad400 training data ?
1
#11 opened 10 months ago
by
cmp-nct
Is this instruction following model?
1
#1 opened 10 months ago
by
rjmehta
fix vocab size
4
#4 opened 10 months ago
by
jphme
Inconsistency in effective batch size reporting
3
#1 opened 10 months ago
by
bjoernp
Update README.md
1
#2 opened 10 months ago
by
waler4ik28
Update READEME.md to include system prompt
1
#3 opened 11 months ago
by
aari1995
Quantise this model - missing file
1
#10 opened 11 months ago
by
cuh008
gguf version?
1
#2 opened 11 months ago
by
guido1893
Sentence Transformers
2
#8 opened 11 months ago
by
jdjayakaran
Ambiguity in Language detection
5
#7 opened 11 months ago
by
jdjayakaran
tokenizer.model missing?
2
#2 opened 12 months ago
by
darule
Quantised models by thebloke
#1 opened 12 months ago
by
choltha
Training code
2
#2 opened 11 months ago
by
robert-h
First sentence of the description wrong?
1
#1 opened 11 months ago
by
h3ndrik
Add a `chat template` to this repository
1
#6 opened 12 months ago
by
LLukas22
How to achieve better results with fine-tuning
1
#5 opened 12 months ago
by
jdjayakaran
Some weights of LlamaForCausalLM were not initialized from the model checkpoint
1
#3 opened 12 months ago
by
fcivardi
CUDA out of memory applying to a dataset of texts
3
#4 opened 12 months ago
by
fcivardi
how to prompt
2
#5 opened 12 months ago
by
g58892881
Falsche Ausgabe bei Abfrage von Landeshauptstädten:
1
#4 opened 12 months ago
by
darule
Flash attention NVCC requirements
3
#2 opened 12 months ago
by
jdjayakaran
missing tokenizer.model?
7
#2 opened 12 months ago
by
b0968
Can't get any reasonable output
3
#3 opened 12 months ago
by
Sebbecking
tokenizer.model missing?
1
#1 opened 12 months ago
by
b0968
Commercial use
5
#1 opened 12 months ago
by
BramVanroy
Is there a problem with year numbers?
3
#1 opened 12 months ago
by
stelterlab
Fixed typo in FP16 and 8bit examples
1
#4 opened over 1 year ago
by
bjoernp