Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
stainablesteel
10 months ago
|
parent
|
context
|
favorite
| on:
Alignment faking in large language models
pretty interesting that pointing it in the direction of its own self awareness by indicating that it's going to affect it's own training brings about all of these complications
Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: