One approach is to create a sentence embedding. That article's example of best results is from a 2017 paper, and things have moved on a lot since then. A Transformer-based model is going to be superior to a BiLSTM model in just about any NLP task. One relatively straightforward approach would be to run the sentences through a pre-built BERT model.
That typically gives you a 512 dimensional vector, which may be impractical to work with? You could use PCA to reduce that to a more manageable handful of dimensions.
Another approach would be to list up all the distinct words used, and with the help of a domain specialist narrow this down to N key words. Where N is on the order of 20 to 50? Then you could do add N columns to your data, storing the number of times each keyword is mentioned.
A more structured way to do this would to use an ontology for your domain. Here is one of the first google hits I got for papers on medical ontologies, just to give you an idea: https://pubmed.ncbi.nlm.nih.gov/31094361/
Yet another approach would be to do sentiment analysis on the sentences. That will give you a single number, typically -1.0 to +1.0.
If the sentences are full of facts like "Patient smokes, has bruising around eyes, ..." then the keyword approach is better. But if it is more like "Patient unlikely to make it through the night" vs. "Patient lost a lot of blood, but is over the worst now." then the sentiment approach could be more useful.