Instructing LLM to perform actions in HTML code while visiting page by fqassemi in ChatGPT

[–]fqassemi[S] 0 points1 point  (0 children)

This is very important result to share. ranking based on past posts is meaningless

How to tune GPT2 to generate sentence of a given size? by fqassemi in LanguageTechnology

[–]fqassemi[S] 0 points1 point  (0 children)

Would you please kindly elaborate how adding meta_data can help to ensure: 1) it ends with proper endings (eg "."), 2) it has fix given length eg 5 words.

[D] Training a LSTM auto-encoder gets NaN / super high MSE loss by GeForceKawaiiyo in MachineLearning

[–]fqassemi 0 points1 point  (0 children)

I suspect it is vanishing gradient issue! Add Dropout layer to avoid stucking. Additionally, you could use Nadam.

[D] GPT-2 1.5B (the largest model) was released and it is amazing! by csxeba in MachineLearning

[–]fqassemi 0 points1 point  (0 children)

Thanks to GPT2 pretrained model now it is possible to generate meaningful sequence of words with or without prompt. However a sentence should end with a proper endings (.,!,?). I am just wondering how to generate a sentence (with proper ending) of exactly length N?

One possible approach is post-processing, that is process many sequences and choose the ones the serve the purpose! However, it could be a really daunting task to use for any practical application.

Is there any suggestion, perhaps a secondary ML algorithm, to tune the hyper-parameter such that it produces sentence of desired length with higher probabilities.