OP asks about values below "limit of detection" in a quantitation (regression) context.
Limit of detection, however is a method performance characteristic that refers to qualitative rather than quantitative tasks. This is why my original answer refers to the lower limit of qantitation (quantification) instead.
See e.g. Currie, Pure&Appl. Chem., 67, 10, 1699-1723, 1995 on these terms.
Turned out in the discussion, that the limit in question is actually the lower limit (end) of the calibration range. I.e. OP is asking how to deal with samples that turn out to be outside the range they calibrated and validated their method for. This makes my original answer somewhat off-topic.
original answer:
Don't exclude cases solely because they are below LLOQ! (lower limit of quantitation)
- The LLOQ is not a magic hard threshold below which nothing can be said. It is rather a convention to mark the concentration where the relative error of the analyses falls below 10 %.
- Note that LLOQ is often computed assuming homescedasticity, i.e. the absolute error being independent of the concentration. That is, you don't even assume different absolute error for cases below or above LLOQ. From that point of view, LLOQ is essentially just a way to express the absoute uncertainty of the analytical method in a concentration unit. (Like fuel economy in l/100 km vs. miles/gallon)
- Even if analytical error is concentration dependent, two cases with true concentration almost the same but slightly below and above LLOQ have almost the same uncertainty.
- (Left) censoring data (which is the technical term for excluding cases below LLOQ) leads to all kinds of complications in consecutive data analyses (and you'd need to use particular statistical methods that can deal with such data).
- Say thank you to your clinical lab that they provide you with full data: I've met many people who have the opposite difficulty: getting a report that just says below LLOQ, and no possibility to recover any further information.
Bottomline: never censor your data unless you have really, really good reasons for doing so.
update about replacing below LLOQ results by $\frac{LLOQ}{\sqrt 2}$ and in view of the comments below:
From the comments below I think that we may be using terms in slightly different ways - possibly due to coming from different fields:
- linear range: concentration range where we get linear (can be relaxed to strictly monotonous) dependency of signal as function of concentration.
calibration range: concentration range covered by calibration and validation samples. Outside this [precisely: the range covered by validation], we don't really know how our method behaves.
=> I totally agree that something$^{TM}$ needs to be done if 1/8 of the samples are outside calibration range and those samples moreover are important to reach correct conclusions in the study.
LLOQ (aka LQ or LOQ) is a method performance characteristic. They can be inside or outside the calibration range. The most basic definition of LLOQ I'm aware of is specifying a relative error that must not be exceeded in the quantitation range. In my field, it is typically set to 10 % (but that can vary, and in fact should vary according to the requirements of the application. So the 10 % relative error to me is like the 5 % threshold for $p$ in significance testing, a default convention)
I've encountered various replacement constants for $\leq$ LLOQ concentrations, ranging from 0 over $\frac{LLOQ}{2}$ to LLOQ and random values in the concentration range below LLOQ ( $\frac{LLOQ}{\sqrt 2}$ is new in that collection - what's the idea behind the $\frac{1}{\sqrt 2}$?). These are normally desperate attempts to impute values for censored data where not even (gu)estimates are available.
With the above definition of LLOQ and validation data available for the whole concentration range encountered in your samples, replacing concentrations below LLOQ would amount to throwing away most of the little information you have for those samples, and that's typically not what you want. See my answer to this related question for illustrations.
Also, this wouldn't avoid the necessity for using statistical methods that work with censored data, so there really isn't much difference to excluding.
In your case, however, would it be possible to extend the validated concentration range somewhat further to the low side? Even if you don't extend the calibration range accordingly (which weuld IMHO be best).
You say that excluding or not those 12 % of your samples has a large influence on your findings. So the idea here would be to rescue the study by establishing linearity and analytical error for a sufficiently large fraction of your samples to get stable estimates at the higher level (study question).
While not being as good as having proper calibration from the beginning, the unexpected after all is something that has to be expected in research. With appropriate caution in the conclusions, this would IMHO be acceptable for early research (in contrast to later stages where better knowledge of expected concentration range would be available or work to establish this as analytical/clinical method).
There are situations where your calibration function starts with a plateau for low concentrations before reaching a suitable sensitivity in the linear range (e.g. a particular absolute amount of analyte is masked/lost due to adsorption, ...). In contrast to low concentrations outside calibration/validation range but still inside linear range, you basically cannot say anything for concentrations in that plateau.
I think of this as a "wet lab/chemical censoring". In that case => use statistical methods for censored data.
The higher level question here is whether your analytical method is fit for purpose.