toad.social is one of the many independent Mastodon servers you can use to participate in the fediverse.
Mastodon server operated by David Troy, a tech pioneer and investigative journalist addressing threats to democracy. Thoughtful participation and discussion welcome.

Administered by:

Server stats:

269
active users

#reinforcementlearning

1 post1 participant0 posts today

[AGI discussion, DeepMind] Welcome to the Era of Experience
storage.googleapis.com/deepmin
old.reddit.com/r/MachineLearni

* threshold of new era in AI that promises unprecedented level of ability
* new generation of agents will acquire superhuman capabilities, learning predominantly f. experience
* paradigm shift, accompanied by algorithmic advancements in RL, will unlock new supra-human capabilities

#Google#DeepMind#AI

Happy birthday to Cognitive Design for Artificial Minds (lnkd.in/gZtzwDn3) that was released 4 years ago!

Since then its ideas have been presented and discussed widely in the research fields of AI/Cognitive Science/Robotics and - nowadays - both the possibilities and the limitations of: #LLMs, #GenerativeAI and #ReinforcementLearning (already envisioned and discussed in the book) have become a common topic of research interests in the AI community and beyond.
Similarly also the topic concerning the evaluation - in human-like and human-level terms - of the current AI systems has become a critical theme related to the problem Anthropomorphic interpretation of AI output (see e.g. lnkd.in/dVi9Qf_k ).
Book reviews have been published on ACM Computing Reviews (2021) lnkd.in/dWQpJdkV and on Argumenta (2023): lnkd.in/derH3VKN

I have been invited to present the content of the book in over 20 official scientific events in international conferences, Ph.D Schools in US, China, Japan, Finland, Germany, Sweden, France, Brazil, Poland, Austria and, of course, Italy.

A news I am happy to share is that Routledge/Taylor & Francis contacted me few weeks ago for a second edition! Stay tuned!

The #book is available in many webstores:
- Routledge: lnkd.in/dPrC26p
- Taylor & Francis: lnkd.in/dprVF2w
- Amazon: lnkd.in/dC8rEzPi

@academicchatter @cognition
#AI #minimalcognitivegrid #CognitiveAI #cognitivescience #cognitivesystems

The article provides good insights into industry leaders such as Waymo, DeepMind, and Amazon demonstrate the transformative power of Reinforcement Learning (RL).

Takeaways:
➡️ RL drives autonomy and innovation across industries, but challenges like interpretability remain pivotal.
➡️ Hybrid systems that blend RL and symbolic reasoning hint at breakthroughs in high-level decision-making.

computer.org/publications/tech

IEEE Computer Society · Reinforcement Learning in Agentic SystemsThis article explores the role of RL in agentic systems and showcase its transformative impact across industries.

My colleagues at TU Delft are seeking to hire a postdoc to work on Applied Planning and Scheduling under Uncertainty, with applications in modelling supply chain scenarios for offshore wind farm installation: careers.tudelft.nl/job/Delft-P

careers.tudelft.nlPostdoc in Applied Planning and Scheduling under UncertaintyPostdoc in Applied Planning and Scheduling under Uncertainty

How to formulate exploration-exploitation trade-off better than all the hacks on top of Bellman equation?

We can first of all simply estimate the advantage of exploration by Monte-Carlo in a swarm setting: Pitting fully exploitative agents against fully exploitative agents which have the benefit of recent exploration. This can be easily done by lagging policy models.

Of course the advantage of exploration needs to be divided by the cost of exploration, which is linear to the number of agents used in the swarm to explore at a particular state.

Note that the advantage of exploration depends on the state of the agent, so we might want to define an explorative critic to estimate this.

What's beautiful in this formulation is that we can incorporate autoregressive #WorldModels naturally, as the exploitative agents only learn from rewards, but the explorative agents choose their actions in a way which maximizes the improvement of the auto-regressive World Model.

It brings these two concepts together as sides of the same coin.

Exploitation is reward-guided action, exploration is auto-regressive state transition model improvement guided action.

Balancing the two is a swarm dynamic which encourages branching where exploration has an expected value in reward terms. This can be estimated by computing the advantage of exploitative agents utilizing recent exploration versus agents which do not, and returning this advantage to the points of divergence between the two.

DeepSeek R1: All you need to know 🐳

The article covers various aspects of the model, from its architecture to training methodologies and practical applications. The explanations are mostly clear and detailed, making complex concepts like Mixture of Experts (#MoE) and reinforcement learning easy to understand.

fireworks.ai/blog/deepseek-r1-

DeepSeek R1: All you need to know 🐳DeepSeek R1: All you need to know 🐳

New instance, new #introduction!

I'm a #DataScientist with a background in #ReinforcementLearning and #ElectricalEngineering. Well, that's what my resume says, but really I'm a #poet and a SF/F #writer. I love to play #DnD and other #TTRPGs.

I use they/them pronouns, and "Dr." not "Mr.", please and thank you.

I maintain a blog at www.seanpatrick.phd which includes a current list of publications, including my debut sonnet collection, "Love, Death, and Other Surprises."