Use https://github.com/marketplace/models/azure-openai/gpt-4-1/playground with Python wrapper.
Transformer blocks are
- Create tokens from input text and assign/create token_ids to each token.
- Assign vector embeddings for each token. (semantic meaning)
- Add positional value to each vector embedding.
- Add contextual meaning to each vector. This is done by an attention mechanism. Q, K, V matrix concept.
- Feed forward network / activation function on each context rich vector.
- Logits, convert the vectors into logits size and apply softmax to predict the next token. Each context-rich vector predicts the next token.