Hacker News new | past | comments | ask | show | jobs | submit login

fine-tuning is not learning, it's controlling the response and you see it's absurd effects in countless examples where in the name of being politically correct the "weights" have been modified also for the past. (classic example of Gemini's German Nazi "representative" photo or even more: https://art-for-a-change.com/blog/2024/02/gemini-artificial-...)



The mechanism for fine-tuning and the original training is exactly the same (gradient descent on weights). The effects you describe are results of what exactly is used to fine tune on.


the mechanism is the same (that's why it impacts all weights) but the target of gradient descent is not the same. in finetuning they aren't saying "go down the mountain" anymore, but "go down toward this plateau" ofcourse this changes the gradient.

is it "learning" in a certain sense? sure, the same way like all indoctrinations are sold as "teaching". the model "learned" to be rapresentative but forgot what 1940 nazi soldier was like...

and fine tuning is not a scalable approach because it has human feeback in the loop. could they fix this error with more fine tuning? yes and they tried abut then the users simply asked "give me a picture a viking warriors" and the problems was again there, you can't fine-tune everything even if we assume the purpose is always noble.


I think all this ideological stuff is completely unrelated to the issue we started talking about. You can fine tune a large model on whatever data you want, and so you can also fine tune it on the most recent user inputs. You can do unsupervised fine-tuning btw, no need for human in the loop. It all depends on what you want to achieve.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: