1

Fine tuning is a concept commonly used in deep learning. We may have a pre-trained model and then fine-tune it to our specific task.

Does that apply to simple models, such as logistic regression?

For example, let's say I have a dataset with attribute variables of an animal and I want to classify whether or not it is a mammal or not. The labels on that dataset are only "mammal"/"not mammal". I then train a logistic regression model for this task, which performs fairly good.

Now, let's say I just received some new data, which has the same variables, but only labels observations as "dog" or "not dog". Given this, could I fine-tune my previous model for this new task? My previous models already performs good, as in, it know how to identify a mammal, so maybe I could fit a new model initializing the coefficients with the previous model coefficient values.

What would you expect, performance-wise, for this approach?

Some assumptions:

  • The first dataset is way bigger than the second one.
  • Both datasets have the same variables, but different labels
  • The Logit model is specifically the sklearn implementation.
eduardokapp
  • 111
  • 2

0 Answers0