In terms of normalizing the probability mass function, it should be a relatively simple fix.
Define your current probability mass function $P(X=x)$. Suppose that this sum of all probabilities $\Sigma_{i=1}^n P(X=x_i)$ equals some value $c$, where $c$ in this case is greater than 1. Note that when $\Sigma_{i=1}^n P(X=x_i)=c\neq 1$, this is not a valid probability function.
Define a new probability function $P'(X=x_i)=\frac{P(X=x_i)}{c}$ for all values $x_i$. Because you are normalizing by the sum of probabilities $c$, the new probability function $P'(X=x_i)$ will now sum to 1 and, provided it satisfies all other Kolmogorov Axioms, will be a valid probability function.
(What I described here will actually be applicable for any value of $c\neq 0$, so $c$ need not be greater than 1 in order for this to work.)
As for the missing/lost values, that will take some additional treatment. If you are comfortable proceeding without that lost information and assuming that, for example, the proportion of 1s you see across the three datasets is an accurate portrayal of the true proportion of 1s you expect to see in the aggregated dataset, then this normalization technique is all you need to do. (Note that this should be true for the proportion of 1s, 2s, and so on.)
If you do not believe these values to be missing completely at random (and I strongly doubt that they are MCAR), then you may need to look into survival analysis or missing data imputation to estimate or impute these values before estimating the p.m.f. If you proceed with a p.m.f. ignoring the missingness or the censoring mechanism, then your p.m.f. will likely be an inaccurate representation of the true distribution of the variable of interest. (It follows that if your p.m.f. is an inaccurate representation of the variable of interest, then your c.d.f. will also be an inaccurate representation.)