Generalized Linear Models (GLMs) play a critical role in fields including Statistics, Data Science, Machine Learning, and other computational sciences.

In Part I of this Series, we provided a thorough mathematical overview (with proofs) of common GLMs both in Canonical and Non-Canonical forms. The next problem to tackle is, how do we actually fit data to GLM models?

When looking at GLMs from a historical context, there are three important data-fitting procedures which are closely connected:

- Newton-Raphson
- Fisher Scoring
- Iteratively Reweighted Least Squares (IRLS)

I have found the relationships and motivations of these techniques is often poorly understood, with the…

Generalized Linear Models (GLMs) play a critical role in fields including Statistics, Data Science, Machine Learning, and other computational sciences. This class of models are a generalization of ordinary linear regression for certain response variable types with error distribution models other than a normal distribution.

To refresh our memories on ordinary linear regression:

Causal Inference is a field that touches several domains and is of interest to a wide range of practitioners including Statisticians, Data Scientists, Machine Learning Scientists, and other Computational Researchers.

To date I have written several pieces on methods/topics in the Causal Inference space. These include:

- Efficient Sampling Frameworks in Causal Inference
- Doubly Robust Estimation Techniques in Causal Inference
- G-Estimation of Semi-Parametric Structural Nested Models
- Recovery of Causal Effects in the presence of M-Bias Structures
- The need for Marginal Structural Modeling in AB Tests / Randomized Trials for Informative Censoring Adjustment
- Valid Inferential Coverage with Multiple Comparisons

This piece concerns…

Causal Inference is a field of interest to a wide range of practitioners including Statisticians, Data Scientists, Machine Learning Scientists, and other Computational Researchers. With respect to the recovery of unbiased estimates of location parameters of Causal Effects in both randomized and non-randomized settings, I have written several pieces including:

- Efficient Sampling Frameworks in Causal Inference
- Doubly Robust Estimation Techniques in Causal Inference
- G-Estimation of Semi-Parametric Structural Nested Models
- Recovery of Causal Effects in the presence of M-Bias Structures
- The need for Marginal Structural Modeling in AB Tests / Randomized Trials for Informative Censoring Adjustment

In this piece, we shift…

Causal Inference is a field that touches several domains and is of interest to a wide range of practitioners including Statisticians, Data Scientists, Machine Learning Scientists, and other Computational Researchers. Recovery of unbiased estimates of Causal Effects is at times a tough task, particularly in non-randomized settings. I have written several technical pieces on leveraging G-Methods for necessary adjustment to recover causal inferences/contrasts of interests; these include pieces on **Efficient Sampling Designs in Causal Inference**, **Doubly Robust Estimation Techniques**, **G-Estimation of Structural Nested Models**, and **Marginal Structural Modeling for informative censoring adjustment in randomized A/B Tests**.

This piece is a…

Causal Inference is of interest to a wide range of practitioners including Statisticians, Data Scientists, Machine Learning Scientists, and other Computational Researchers. Recovery of unbiased estimates of Causal Effects is at times a tough task. In my previous pieces on **Doubly Robust Estimation** and **G-Estimation of Structural Nested Models**, we discussed leveraging G-Methods in non-randomized settings. We also discussed issues with **M-Bias and confounding identification in non-randomized settings**.

**Contrary to popular belief, recovery of valid estimates of Causal Effects can be equally as difficult (or sometimes even unidentifiable and impossible) with A/B Testing (aka randomized trials)**. Regarding the use of…

Causal Inference is a field that touches several domains and is of interest to a wide range of practitioners including Statisticians, Data Scientists, Machine Learning Scientists, and other Computational Researchers. Recovery of unbiased estimates of Causal Effects is at times a tough task, even in randomized settings. This task can be particularly challenging in non-randomized settings, requiring an array of often empirically untestable assumptions to hold that have both mathematical and philosophical implications.

In the interest of recovering unbiased estimates of Mean Causal Effects of an Intervention-Outcome relationship, there are several tools we can leverage. Under the assumptions of correct…

Causal Inference is a field with wide-ranging implications, from clinical trials and A/B testing to observational and natural experiments; it’s a field that touches nearly every domain and is of interest to many practitioners including Statisticians, Machine Learning Scientists, and Computational Researchers. Recovery of unbiased estimates of Causal Effects is at times a tough task, even in Randomized settings. This task can be particularly challenging in Non-Randomized settings, requiring an array of often empirically untestable assumptions to hold that have both mathematical and philosophical implications.

In the interest of recovering an unbiased estimate of a Mean Causal Effect of an…

When applying methods for Causal Inference, particularly in non-randomized settings, the family of Case-Control Designs are a powerful tool. They are an efficient sampling framework for recovering unbiased estimates of Causal Effects that would have been recovered from a full Cohort study, without actually conducting said Cohort study.

Like nearly all methods in Causal Inference, Case-Control designs were initially developed within medical and public-health related fields; much of the language and nomenclature in these design paradigms reflect this fact. …

For anyone pursuing study in Data Science, Statistics, or Machine Learning, stating that *“The Central Limit Theorem (CLT) is important to know”* is an understatement. Particularly from a Mathematical Statistics perspective, in most cases the CLT is what makes recovery of valid inferential coverage around parameter estimates a tractable and solvable problem.

There are several articles on the Medium platform regarding the CLT. I noticed however not a single article (as to my knowledge) that delved into the mathematics of the theorem, nor even properly specified the assumptions under which the CLT holds. **This is a tremendous disservice in my…**

Chief Data Scientist & Head of AI @ AIMatters | Harvard trained Statistician and Machine Learning Scientist | Interests in Statistical ML & Causal Inference