Comments on: An Analysis of Large Language Models in the HealthCare Domain /2023/04/14/omobolanle-oladeji/ Mon, 15 May 2023 14:45:53 +0000 hourly 1 https://wordpress.org/?v=6.9.4 By: Bolanle /2023/04/14/omobolanle-oladeji/#comment-30005 Fri, 21 Apr 2023 21:45:21 +0000 /?p=41454#comment-30005 Anjola: Thank you!

1. I picked DialoGPT, BERT, and T5 to train because they were open source models. They also represent each category of the three major types of models that most Transformer-based models fall into (encoder only, decoder only, encoder-decoder).

2. It was not difficult in terms of putting the code together, however it was very computationally heavy and expensive to train these models. They also took a lot of time (> 20 hours on average).

3. They could answer some complex questions in the medical realm, some times I was surprised by some of the answers. However, because my model is not trained on the most current diseases, they could not answer questions on CoVid-19 for example.

If you have any more questions, please do not hesitate to contact me by email or Instagram. Thank you for the thoughtful questions!

]]>
By: Bolanle /2023/04/14/omobolanle-oladeji/#comment-30003 Fri, 21 Apr 2023 21:34:56 +0000 /?p=41454#comment-30003 Thank you so much Saralee!

]]>
By: Anjolaoluwa Olubusi /2023/04/14/omobolanle-oladeji/#comment-29842 Fri, 21 Apr 2023 14:12:52 +0000 /?p=41454#comment-29842 Great IS. Got a few questions.
1.) Why did you pick DialoGPT, BERT and T5 to train?
2.) How difficult was it to train the LLMs?
3.) How complex of a question could the LLMs answer? Were there questions the models could not answer?

]]>
By: Saralee /2023/04/14/omobolanle-oladeji/#comment-29821 Fri, 21 Apr 2023 13:13:41 +0000 /?p=41454#comment-29821 This is super cool research. Congratulations!!!

]]>