Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

New posts in huggingface-transformers

RuntimeError: Failed to import transformers.pipelines because of the following error (look up to see its traceback): initialization failed

Semantic searching using Google flan-t5

Can mT5 model on Huggingface be used for machine translation?

Saving BERT Sentence Embedding

Loading saved NER back into HuggingFace pipeline?

Is the default `Trainer` class in HuggingFace transformers using PyTorch or TensorFlow under the hood?

Hugging Face Transformers trainer: per_device_train_batch_size vs auto_find_batch_size

pytorch summary fails with huggingface model II: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cpu

Config change for a pre-trained transformer model

cannot import name 'flash_attn_func' from 'flash_attn'

How to specify number of target classes for TFRobertaSequenceClassification?

How to return a dictionary of tensors from tf.py_function?

Are the pre-trained layers of the Huggingface BERT models frozen?

BERT for time series classification

Model not calculating loss during training returning ValueError (Huggingface/BERT)

Tensorflow BERT for token-classification - exclude pad-tokens from accuracy while training and testing

Getting "Unable to load weights from pytorch checkpoint file" when loading model from transformers

AttributeError: module 'huggingface_hub.constants' has no attribute 'HF_HUB_CACHE'

Force BERT transformer to use CUDA

BertModel or BertForPreTraining