The existence of Stein's Example prima facie appears similar to seemingly unrelated regression (SUR) insofar as simultaneously estimating multiple parameters seems more effective than training them separately.
In the Wikipedia articles linked above there seems to be a somewhat different emphasis. The article on Stein's example is focused on a minimum example to show the seemingly paradoxical decrease in mean squared error. The article on SUR emphasizes the effectiveness of training multiple regression equations together that could have been trained separately. But neither article refers to the other.
Is there a precise relationship between these notions? For instance, might SUR be an example of Stein's Example/Paradox?