ℏεsam's picture
10 4

ℏεsam PRO

hesamation

AI & ML interests

post-training / reasonign models / RAG

Recent Activity

reacted to their post with ❤️ about 15 hours ago
The best researchers from Yale, Stanford, Google DeepMind, and Microsoft laid out all we know about Agents in a 264-page paper [book], Here are some of their key findings: They build a mapping of different agent components, such as perception, memory, and world modelling, to different regions of the human brain and compare them: - brain is much more energy-efficient - no genuine experience in agents - brain learns continuously, agent is static An agent is broken down to: - Perception: the agent's input mechanism. can be improved with multi-modality, feedback mechanisms (e.g., human corrections), etc. - Cognition: learning, reasoning, planning, memory. LLMs are key in this part. - Action: agent's output and tool use. Agentic memory is represented as: - Sensory memory or short-term holding of inputs which is not emphasized much in agents. - Short-term memory which is the LLM context window - Long-term memory which is the external storage such as RAG or knowledge graphs. The memory in agents can be improved and researched in terms of: - increasing the amount of stored information - how to retrieve the most relevant info - combining context-window memory with external memory - deciding what to forget or update in memory The agent must simulate or predict the future states of the environment for planning and decision-making. ai world models are much simpler than the humans' with their causal reasoning (cause-and-effect) or physical intuition. LLM world models are mostly implicit and embedded. EMOTIONS are a deep aspect of humans, helping them with social interactions, decision-making, or learning. Agents must understand emotions to better interact with us. But rather than encoding the feeling of emotions, they have a surface-level modelling of emotions. Perception is the process by which an agent receives and interprets raw data from its surroundings. READ PAPER: https://huggingface.co/papers/2504.01990
posted an update about 15 hours ago
The best researchers from Yale, Stanford, Google DeepMind, and Microsoft laid out all we know about Agents in a 264-page paper [book], Here are some of their key findings: They build a mapping of different agent components, such as perception, memory, and world modelling, to different regions of the human brain and compare them: - brain is much more energy-efficient - no genuine experience in agents - brain learns continuously, agent is static An agent is broken down to: - Perception: the agent's input mechanism. can be improved with multi-modality, feedback mechanisms (e.g., human corrections), etc. - Cognition: learning, reasoning, planning, memory. LLMs are key in this part. - Action: agent's output and tool use. Agentic memory is represented as: - Sensory memory or short-term holding of inputs which is not emphasized much in agents. - Short-term memory which is the LLM context window - Long-term memory which is the external storage such as RAG or knowledge graphs. The memory in agents can be improved and researched in terms of: - increasing the amount of stored information - how to retrieve the most relevant info - combining context-window memory with external memory - deciding what to forget or update in memory The agent must simulate or predict the future states of the environment for planning and decision-making. ai world models are much simpler than the humans' with their causal reasoning (cause-and-effect) or physical intuition. LLM world models are mostly implicit and embedded. EMOTIONS are a deep aspect of humans, helping them with social interactions, decision-making, or learning. Agents must understand emotions to better interact with us. But rather than encoding the feeling of emotions, they have a surface-level modelling of emotions. Perception is the process by which an agent receives and interprets raw data from its surroundings. READ PAPER: https://huggingface.co/papers/2504.01990
View all activity

Organizations

lora concepts library's profile picture AI Zero to Hero's profile picture The Waifu Research Department's profile picture Blog-explorers's profile picture Multi🤖Transformers's profile picture Team Tonic's profile picture C4AI Community's profile picture Hugging Face Discord Community's profile picture

Posts 3

view post
Post
718
The best researchers from Yale, Stanford, Google DeepMind, and Microsoft laid out all we know about Agents in a 264-page paper [book],

Here are some of their key findings:

They build a mapping of different agent components, such as perception, memory, and world modelling, to different regions of the human brain and compare them:

- brain is much more energy-efficient
- no genuine experience in agents
- brain learns continuously, agent is static

An agent is broken down to:
- Perception: the agent's input mechanism. can be improved with multi-modality, feedback mechanisms (e.g., human corrections), etc.
- Cognition: learning, reasoning, planning, memory. LLMs are key in this part.
- Action: agent's output and tool use.

Agentic memory is represented as:
- Sensory memory or short-term holding of inputs which is not emphasized much in agents.
- Short-term memory which is the LLM context window
- Long-term memory which is the external storage such as RAG or knowledge graphs.

The memory in agents can be improved and researched in terms of:
- increasing the amount of stored information
- how to retrieve the most relevant info
- combining context-window memory with external memory
- deciding what to forget or update in memory

The agent must simulate or predict the future states of the environment for planning and decision-making.

ai world models are much simpler than the humans' with their causal reasoning (cause-and-effect) or physical intuition.

LLM world models are mostly implicit and embedded.

EMOTIONS are a deep aspect of humans, helping them with social interactions, decision-making, or learning.

Agents must understand emotions to better interact with us.

But rather than encoding the feeling of emotions, they have a surface-level modelling of emotions.

Perception is the process by which an agent receives and interprets raw data from its surroundings.

READ PAPER: Advances and Challenges in Foundation Agents: From Brain-Inspired Intelligence to Evolutionary, Collaborative, and Safe Systems (2504.01990)
view post
Post
2556
What, How, Where, and How Well? This paper reviews test-time scaling methods and all you need to know about them:
> parallel, sequential, hybrid, internal scaling
> how to scale (SFT, RL, search, verification)
> metrics and evals of test-time scaling

🔗paper: What, How, Where, and How Well? A Survey on Test-Time Scaling in Large Language Models (2503.24235)

If you want to learn what inference-time compute scaling is @rasbt has a great blog post on that:
https://magazine.sebastianraschka.com/p/state-of-llm-reasoning-and-inference-scaling

models

None public yet