Well, instruction tuning is closely related to both.
For most commercial use, you want the thing to answer questions, but refuse to answer some. So you have an appropriate dataset that encourages it to be cooperative, not make up stuff, and not be super eager to go on rants about "the blacks" even though that's well-represented in its training data.
For most commercial use, you want the thing to answer questions, but refuse to answer some. So you have an appropriate dataset that encourages it to be cooperative, not make up stuff, and not be super eager to go on rants about "the blacks" even though that's well-represented in its training data.