189 points by aanet 2 days ago | 113 comments | View on ycombinator
Animats 1 day ago |
theptip 1 day ago |
I would say his core point does still apply; autonomous learning is not solved by ICL. But it seems a strawman to ignore the topic entirely and focus on training.
From what I see on the ground, some degree of autonomous learning is possible; Agents can already be set up to use meta-learning skills for skill authoring, introspection, rumination, etc - but these loops are not very effective currently.
I wonder if this is the myopic viewpoint of a scientist who doesn’t engage with the engineering of how these systems are actually used in the real world (ie “my work is done once Llama is released with X score on Y eval”) which results in a markedly different stance than the guys like Sutskever, Karpathy, Amodei who have built end-to-end systems and optimized for customer/business outcomes.
zhangchen 1 day ago |
aanet 2 days ago |
"he proposed framework integrates learning from observation (System A) and learning from active behavior (System B) while flexibly switching between these learning modes as a function of internally generated meta-control signals (System M). We discuss how this could be built by taking inspiration on how organisms adapt to real-world, dynamic environments across evolutionary and developmental timescales. "
utopiah 1 day ago |
TL;DR: depends where you defined the boundaries of your "system".
krinne 1 day ago |
beernet 2 days ago |
The proposed System M (Meta-control) is a nice theoretical fix, but the implementation is where the wheels usually come off. Integrating observation (A) and action (B) sounds great until the agent starts hallucinating its own feedback loops. Unless we can move away from this 'outsourced learning' where humans have to fix every domain mismatch, we're just building increasingly expensive parrots. I’m skeptical if 'bilevel optimization' is enough to bridge that gap or if we’re just adding another layer of complexity to a fundamentally limited transformer architecture.
logicchains 1 day ago |
est 1 day ago |
Imagine if AI learns all your source code and apply them to your competitor /facepalm
jdkee 1 day ago |
tranchms 1 day ago |
shevy-java 1 day ago |
However had, there will come a time when AI will really learn. My prediction is that it will come with a different hardware; you already see huge strides here with regards to synthetic biology. While this focuses more on biology still, you'll eventually see a bridging effort; cyborg novels paved the way. Once you have real hardware that can learn, you'll also have real intelligence in AI too.
Garlef 1 day ago |
That's why I think the term "system" as used in the paper is much better.
himata4113 1 day ago |
They're capable enough to put themselves in a loop and create improvement which often includes processing new learnings from bruteforcing. It's not in real-time, but that probably a good thing if anyone remembers microsofts twitter attempt.
BrianFHearn 1 day ago |
lock-locku 1 day ago |
webagent255 about 21 hours ago |
theLewisLu 1 day ago |
scotttaylor about 23 hours ago |
followin_io82 1 day ago |
seedpi 1 day ago |
Frannky 1 day ago |
lovebite4u_ai 1 day ago |
Today's locked-down pre-trained models at least have some consistency.
[1] https://www.bbc.com/news/technology-35890188