Matyas.
ServicesProjectsExperienceBlogContact
CSGet in touch
Back to Dictionary
ai

Transformer

The transformer is a neural network architecture introduced in the 2017 paper "Attention Is All You Need" that revolutionized natural language processing. Unlike recurrent networks, transformers process entire sequences in parallel using a self-attention mechanism, which allows them to capture long-range dependencies efficiently. Virtually all modern LLMs, including GPT and Claude, are built on the transformer architecture.

#ai

Related Terms

Embedding

An embedding is a dense numerical vector representation of data — such as text, images, or code — in a high-dimensional space where semantically similar items are positioned closer together. Embeddings are fundamental to semantic search, recommendation systems, and RAG pipelines. They are generated by specialized models and typically stored in vector databases for efficient similarity lookups.

Hallucination

In AI, hallucination refers to when a language model generates confident-sounding but factually incorrect or fabricated information. This occurs because LLMs predict statistically likely text rather than retrieving verified facts. Mitigation strategies include RAG, grounding responses in source documents, structured output validation, and using temperature settings to reduce creative deviation.

Reinforcement Learning

Reinforcement learning (RL) is a machine learning paradigm where an agent learns optimal behavior by interacting with an environment and receiving rewards or penalties. RLHF (Reinforcement Learning from Human Feedback) is a key technique used to align LLMs with human preferences, making their outputs more helpful and safe. RL is also behind breakthroughs in game-playing AI and robotics.

Neural Network

A neural network is a computational model inspired by the human brain, consisting of layers of interconnected nodes (neurons) that process data by adjusting weighted connections during training. Deep neural networks with many layers form the foundation of modern AI, powering everything from image recognition to language understanding. Common architectures include feedforward networks, convolutional networks (CNNs), and transformers.

Prompt Engineering

Prompt engineering is the practice of crafting and optimizing input instructions to guide AI models toward producing desired outputs. Techniques include few-shot examples, chain-of-thought reasoning, role assignment, and structured output formatting. Effective prompt engineering can dramatically improve the quality, accuracy, and consistency of LLM responses without modifying the underlying model.

Natural Language Processing

Natural Language Processing (NLP) is a branch of AI focused on enabling computers to understand, interpret, and generate human language. NLP powers applications like chatbots, translation services, sentiment analysis, and text summarization. Modern NLP has been transformed by transformer-based models, which achieve remarkable performance on tasks that previously required extensive hand-crafted rules.

All Words
Matyas.

Web apps, mobile apps, AI automation. I help businesses save time and money with tech that actually works.

Links

  • Services
  • Projects
  • Experience
  • Blog
  • Dictionary
  • Contact

Coming Soon

  • Case StudiesSoon
  • Resources

© 2026 Matyas Prochazka. All rights reserved.