Technology

1139 readers
20 users here now

A tech news sub for communists

founded 2 years ago
MODERATORS
51
52
53
54
55
56
57
58
59
60
1
submitted 3 weeks ago* (last edited 3 weeks ago) by yogthos@lemmygrad.ml to c/technology@lemmygrad.ml
 
 

Traditional autoregressive language models generate text sequentially, one token at a time, leading to slower outputs with limited coherence and quality.

Diffusion models are an alternative approach. Instead of direct prediction, they iteratively refine noise, enabling faster generation, dynamic error correction, and greater control. This makes them particularly effective for editing tasks, including in math and code.

https://github.com/ML-GSAI/LLaDA

61
62
 
 

Here are the PRs that are created entirely by AI:

Devs are talking to the AI trying to get it to fix some mistakes. It's so cringe.

63
64
65
66
67
68
69
70
71
72
 
 

the goal is to have an agent that can:

  • Understand a complex problem description.
  • Generate initial algorithmic solutions.
  • Rigorously test its own code.
  • Learn from failures and successes.
  • Evolve increasingly sophisticated and efficient algorithms over time.

https://storage.googleapis.com/deepmind-media/DeepMind.com/Blog/alphaevolve-a-gemini-powered-coding-agent-for-designing-advanced-algorithms/AlphaEvolve.pdf

73
74
75
 
 

MiMo-7B, a series of reasoning-focused language models trained from scratch, demonstrating that small models can achieve exceptional mathematical and code reasoning capabilities, even outperforming larger 32B models. Key innovations include:

  • Pre-training optimizations: Enhanced data pipelines, multi-dimensional filtering, and a three-stage data mixture (25T tokens) with Multiple-Token Prediction for improved reasoning.
  • Post-training techniques: Curated 130K math/code problems with rule-based rewards, a difficulty-driven code reward for sparse tasks, and data re-sampling to stabilize RL training.
  • RL infrastructure: A Seamless Rollout Engine accelerates training/validation by 2.29×/1.96×, paired with robust inference support. MiMo-7B-RL matches OpenAI’s o1-mini on reasoning tasks, with all models (base, SFT, RL) open-sourced to advance the community’s development of powerful reasoning LLMs.

an in-depth discussion of mimo-7b https://www.youtube.com/watch?v=y6mSdLgJYQY

view more: ‹ prev next ›