Working Of Transformers

Chethan GChethan G
1 min read

#chaicode
In 2017, a research paper by Google, “Attention is All You Need“. This was created to enhance Google Translate’s capabilities.

Architecture of Transformers

This was the architecture that was proposed in the research paper. This is the brain behind all the Next-word predictors out there, i.e ChatGPT, Gemini2.5, Claude, etc.

  1. Step 1: Tokenisation is the process that converts the given input strings into ‘Tokens‘ ,which are assigned numbers to the given strings according to their (Vocab size: are size of model’s vocabulary).

  2. Step 2: Vector Embeddings creates relationships across those tokens inside a 3D vector space by holding their semantics or meaning.

  3. Step 3: Positional Encoding positions the Vector according to the position of the input strings

    Ex: “Chethan bunked college” and “College bunked chethan” are 2 different things.

  4. Step 4: Self Attention a crucial mechanism that enables the vectors to communicate with each other. and generates the probability of next word occurrences.

  5. Step 5: The decoding process are started.

Conclusion

In reality, GPT or Generative Pre-Trained Transformers is just a next-word generator, but with immense capabilities to change the future of work.

0
Subscribe to my newsletter

Read articles from Chethan G directly inside your inbox. Subscribe to the newsletter, and don't miss out.

Written by

Chethan G
Chethan G