Habr

Головоломка, кофе и охапка книг, или как я искал истоки термина «Deep Learning». Часть 2

Привет! Некоторое время назад я начал искать истоки термина «Deep Learning» . Тогда я изучал только зарубежные источники и обещал вернуться позже с обзором советской и российской литературы. Что ж, откладывать это больше нельзя. Посмотрим, на кого будут ссылаться отечественные авторы в том, что касается истории развития глубокого обучения. Без долгого вступления — берем в руку пальцы Ctrl/Cmd+F и начинаем раскопки!

habr.com/ru/companies/selectel

#selectel #ии #искусственный_интеллект #машинное_обучение #ml #dl #deep_learning #глубокое_обучение #познавательное

Головоломка, кофе и охапка книг, или как я искал истоки термина «Deep Learning». Часть 2

Привет! Некоторое время назад я начал искать истоки…

Хабр
Scott McCarty

Red Hat named to Fast Company’s annual list of the World’s Most Innovative Companies of 2025 buff.ly/yJk2dvw #AI #ML #DL #NN #oss #opensource

Pustam | पुस्तम | পুস্তম🇳🇵

Moore’s Law for AI agents: the length of tasks that AIs can do is doubling about every 7 months.

These results appear robust. The authors were able to retrodict back to GPT-2. They further ran experiments on SWE-bench Verified and found a similar trend.

Read more: metr.org/blog/2025-03-19-measu

#AIBoom #AI #AIAgents #AIAgent #ArtificialIntelligence #GPT2 #MooreLaw #Tasks #DL #ML #Pustam #Raut #AIRevolution

ma𝕏pool

Self-Improving Reasoners.

Both expert human problem solvers and successful language models employ four key cognitive behaviors

1. verification (systematic error-checking),

2. backtracking (abandoning failing approaches),

3. subgoal setting (decomposing problems into manageable steps), and

4. backward chaining (reasoning from desired outcomes to initial inputs).

Some language models naturally exhibits these reasoning behaviors and exhibit substantial gains, while others don't and quickly plateau.

The presence of reasoning behaviors, not the correctness
of answers is the critical factor. Models with incorrect solutions containing proper reasoning patterns achieve comparable performance to those trained on correct solutions.

It seems that the presence of cognitive behaviors enables self-improvement through RL.

Cognitive Behaviors that Enable Self-Improving Reasoners, or, Four Habits of Highly Effective STaRs
arxiv.org/abs/2503.01307

#reinforcementlearning #RL
#AI #DL #LLM

Cognitive Behaviors that Enable Self-Improving Reasoners, or, Four Habits of Highly Effective STaRs

Test-time inference has emerged as a powerful paradigm…

arXiv.org
Mar 06, 2025, 10:48 · · · 0 · 0
Scott McCarty

"We encourage the open source community, regulatory authorities and industry to continue to strive toward greater transparency and alignment with open source development principles when training and fine-tuning AI models" buff.ly/3Eyn85w #AI #ML #DL #NN #oss #opensource

Scott McCarty

Why is transparent, open data important to LLMs (Part 2)? buff.ly/3QhjQ9t #AI #ML #DL #NN #oss #opensource

HGPU group

Thesis: Towards autonomous resource management: Deep learning prediction of CPU-GPU load balancing

#OpenCL #LoadBalancing #DeepLearning #DL

hgpu.org/?p=29728

Scott McCarty

I agree with RedMonk/O'Grady, that's why I recommend - ramalama run ollama://deepseek-r1:7b - instead of using web or app versions of DeepSeek buff.ly/4gDvLsR #AI #ML #DL #NN #oss #opensource

Scott McCarty

“...If LLMs are just software, then containers are really convenient for LLMs....” buff.ly/42zNVZs#AI #ML #DL #NN #oss #opensource

Scott McCarty

It's interesting to see the business news evangelize what Red Hat has been saying! AI in 2025: Rush for use cases buff.ly/4hwrxnS
#AI #ML #DL #NN #oss #opensource