How It Works

Wobble is powered by a reinforcement learning agent trained to solve Wordle-like puzzles using a simplified Q-learning algorithm.

The Learning Process

The agent plays Wordle-style games repeatedly, learning from trial and error. Each game is an episode, and after thousands of episodes the model improves its ability to pick strong guesses.

Learning is done using the Q-learning update rule, which adjusts the value of each state-action pair:

Q[s][a] ← Q[s][a] + α * (reward + γ * max(Q[s’]) - Q[s][a])

Training

The agent is trained by playing against a large set of possible words. Over time, it learns which strategies increase the chance of solving the puzzle within the allowed attempts. The results of training are stored in a Q-table, which the bot then uses to make informed decisions during play.

Outcome

After training, the agent is able to approach Wordle systematically — starting with informative guesses, narrowing down possibilities, and converging on the correct word more efficiently than random play.


Note:


Wordle is a trademark of The New York Times Company. This project is not affiliated with or endorsed by The New York Times Company.