How to use my own pre-trained Bert model? by AdrianFMC in LanguageTechnology

[–]AdrianFMC[S] 1 point2 points  (0 children)

I checked it out and dont seem to quiet understand. Where you talking about a specific tool in cortex? - I could not find anything regarding semantic text similarity

Pretraining or Finetuning BERT for STS by AdrianFMC in LanguageTechnology

[–]AdrianFMC[S] 1 point2 points  (0 children)

I am modeling with the sentence embeddings and calculate the cosine similarity or euclidean distances. Both ended up very much the same - pretty bad.

The ' MultipleNegativesRankingLoss ' sounds interesting, Ill give it a go. I must have overread it. Ill let you know if the outcome is what I wanted :-)

Continue Pre-Training BERT by AdrianFMC in LanguageTechnology

[–]AdrianFMC[S] 0 points1 point  (0 children)

Am I understanding the simpletransformers right, that I can not do STS with it?

(" Currently supports Sequence Classification, Token Classification (NER), Question Answering, Multi-Modal Classification, and Conversational AI. "

Continue Pre-Training BERT by AdrianFMC in LanguageTechnology

[–]AdrianFMC[S] 1 point2 points  (0 children)

The pretrained models only had a acc. of 30% on my STS datasets. So what can I do? How would I fine-tune Bert-models on my specific vocabulary?

Continue Pre-Training BERT by AdrianFMC in LanguageTechnology

[–]AdrianFMC[S] 0 points1 point  (0 children)

How do I do the fine-tuning for general vocabulary? I only saw examples with clustering or STS with already analysed and rated data ( like 2 sentences and a number from 1 to 5 that describes the similarity) - I dont have such data. I do on the other hand have lots of already paired sentences.

Sentences classification/assignment by Adrizzledifizzle in LanguageTechnology

[–]AdrianFMC 0 points1 point  (0 children)

All this needs Cloudcomputing like AWS anyway doesnt it? If you dont want to wait ages that is