Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

New posts in huggingface-transformers

Continual pre-training vs. Fine-tuning a language model with MLM

How to use fine-tuned model in huggingface for actual prediction after re-loading?

HuggingFace AutoModelForCasualLM "decoder-only architecture" warning, even after setting padding_side='left'

How to understand hidden_states of the returns in BertModel?

BERT - Is that needed to add new tokens to be trained in a domain specific environment?

Sentence embeddings from LLAMA 2 Huggingface opensource

Longformer get last_hidden_state

Strange results with huggingface transformer[marianmt] translation of larger text

Transformers model from Hugging-Face throws error that specific classes couldn t be loaded

How to load a fine-tuned peft/lora model based on llama with Huggingface transformers?

huggingface transformers: truncation strategy in encode_plus