../transformer

Transformer

Table of contents

From AutoEncoder to Transformer (?)

How does Transformer evolve today?

[2024 GPT on a Quantum Computer] Integrate quantum computing with LLM behind ChatGPT - Transformer architecture - , which introduced as Quantum Machine Learning.

[2024 Evolution Transformer: In-Context Evolutionary Optimization] Introduce Evolution Transformer, which performance improved from supervised optimization technique and updating invariance of the search distribution.

[2024 Empowering Data Mesh with Federated Learning] Introduce Federated learning into Data Mesh, a decentralized data architecture. Distributed and privacy-preserved.
Parallel computing, Distributed computing, and Concurrent computing
Parallel computing, Distributed computing, and Concurrent computing

[2024 PerOS: Personalized Self-Adapting Operating Systems in the Cloud] Proposed a personalized OS ingrained with LLM capabilities - PerOS.

‘Connection’

Encoder and decoder in a typical autoencoder.
Schema of a basic autoencoder. [Reference from Wiki.]
Schema of a basic autoencoder. Reference from Wiki.
Encoding concept is also adopted in Transformer.
Encoder-Decoder architecture in Transformer
Encoder-Decoder architecture in Transformer

Transformer is for NLP, while AutoEncoder is for image processing.

Recent research about AutoEncoder: See here in arXiv

Quick review of Transformer

<a rel="noopener" target="_blank" href="https://machinelearningmastery.com/what-are-large-language-models/">Reference: What are Large Language Models

Reference: What are Large Language Models

[1950, Shannon] Shannon, C. E. (1951). Prediction and entropy of printed English. The Bell System Technical Journal30(1), 50–64. https://doi.org/10.1002/j.1538-7305.1951.tb01366.x

[2017, Vaswani] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L., & Polosukhin, I. (2023). Attention Is All You Need.

/ML/