About 696,000 results
Open links in new tab
  1. python - How to use DistilBERT Huggingface NLP model to perform ...

    Oct 24, 2021 · I am using DistilBERT to do sentiment analysis on my dataset. The dataset contains text and a label for each row which identifies whether the text is a positive or negative movie review (eg: …

  2. why take the first hidden state for sequence classification ...

    Feb 6, 2020 · In the last few layers of sequence classification by HuggingFace, they took the first hidden state of the sequence length of the transformer output to be used for classification. hidden_state =

  3. How to solve this problem with distilbert tokenizer?

    Nov 24, 2021 · How to solve this problem with distilbert tokenizer? Ask Question Asked 4 years ago Modified 2 years, 9 months ago

  4. How to Fine-tune HuggingFace BERT model for Text Classification

    Sep 2, 2021 · Base Model - TFDistilBertModel Hugging Face Transformers: Fine-tuning DistilBERT for Binary Classification Tasks TFDistilBertModel class to instantiate the base DistilBERT model without …

  5. No model was supplied, defaulted to - Stack Overflow

    Apr 23, 2023 · It tells you, that the pipeline is using distilbert-base-uncased-finetuned-sst-2-english because you haven't specified a model_id. In other words, it might not yield the best results for your …

  6. nlp - Fine-Tuning DistilBertForSequenceClassification: Is not learning ...

    Aug 2, 2020 · This Stack Overflow question discusses issues with fine-tuning DistilBertForSequenceClassification, including unchanged loss and weights not updating during training.

  7. How to get output_attentions of a pretrained Distilbert Model?

    Apr 19, 2022 · I got the solution to anyone looking: modified the code as ` dbert_model = TFDistilBertModel.from_pretrained ('distilbert-base-uncased',config, output_attentions=True)` This …

  8. How can I train distilBERT more efficiently on my large text ...

    Jul 25, 2023 · Model Training: How can I make DistilBERT handle this beast of a dataset more efficiently? Anyone got experience tweaking the training strategy, batch size, learning rate, etc.? …

  9. Download pre-trained BERT model locally - Stack Overflow

    After instantiating the SentenceTransformer via download, you can then save it to any path of your choosing with the 'save ()' method.

  10. cannot import name …

    cannot import name 'DISTILBERT_PRETRAINED_MODEL_ARCHIVE_MAP' from 'transformers.modeling_distilbert' Asked 5 years, 6 months ago Modified 5 years, 5 months ago …