@Tom's response is excellent, but I'd like to offer a version that's more heuristic and that introduces an additional concept.
Logistic regression
Imagine we have a number of binary questions. If we are interested in the probability of responding yes to any one of the questions, and if we're interested in the effect of some independent variables on that probability, we use logistic regression:
$P(y_i = 1) = \frac{1}{1 + exp(X\beta)} = logit^-1(X\beta)$
where i indexes the questions (i.e. the items), X is a vector of characteristics of the respondents, and $\beta$ is the effect of each of those characteristics in log odds terms.
IRT
Now, note that I said we had a number of binary questions. Those questions might all get at some kind of latent trait, e.g. verbal ability, level of depression, level of extraversion. Often, we are interested in the level of the latent trait itself.
For example, in the Graduate Record Exam, we're interested in characterizing the verbal and math ability of various applicants. We want some good measure of their score. We could obviously count how many questions someone got correct, but that does treat all questions as being worth the same amount - it doesn't explicitly account for the fact that questions might vary in difficulty. The solution is item response theory. Again, we're (for now) not interested in either X or $\beta$, but we're just interested in the person's verbal ability, which we'll call $\theta$. We use each person's pattern of responses to all the questions to estimate $\theta$:
$P(y_i = 1) = logit^-1[a_i(\theta_j - b_i)]$
where $a_i$ is discrimination of item i and $b_i$ is its difficulty.
So, that's one obvious distinction between regular logistic regression and IRT. In the former, we're interested in the effects of independent variables on one binary dependent variable. In the latter, we use a bunch of binary (or categorical) variables to predict some latent trait. The original post said that $\theta$ is our independent variable. I'd respectfully disagree, I think it's more like this is the dependent variable in IRT.
I used binary items and logistic regression for simplicity, but the approach generalizes to ordered items and ordered logistic regression.
Explanatory IRT
What if you were interested in the things that predict the latent trait, though, i.e. the Xs and $\beta$s previously mentioned?
As mentioned earlier, one model to estimate the latent trait is just count the number of correct answers, or add up all the values of your Likert (i.e. categorical) items. That has its flaws; you're assuming that each item (or each level of each item) is worth the same amount of the latent trait. This approach is common enough in many fields.
Perhaps you can see where I'm going with this: you can use IRT to predict the level of the latent trait, then conduct a regular linear regression. That would ignore the uncertainty in each person's latent trait, though.
A more principled approach would be to use explanatory IRT: you simultaneously estimate $\theta$ using an IRT model and you estimate the effect of your Xs on $\theta$ as if you were using linear regression. You can even extend this approach to include random effects to represent, for example, the fact that students are nested in schools.
More reading available on Phil Chalmers' excellent intro to his mirt
package. If you understand the nuts and bolts of IRT, I'd go to the Mixed Effects IRT section of these slides. Stata is also capable of fitting explanatory IRT models (albeit I believe it can't fit random effects explanatory IRT models as I described above).