Blog

The Oldest AI Blog in the Czech Republic

We have been writing about artificial intelligence since 2017. 1000+ articles, thousands of pages of thoughts, experiments, and reflections. No sensationalism, no ads.

Tag Filter: BERT × cancel

Browse by Topics

Showing 7 of 7 articles

A New Champion on the Scene?Archive 2021
1 min

A New Champion on the Scene?

The Switch Transformer language model from Google is nearly six times larger than GPT-3! The Switch Transformer has 9 times more parameters, totalling 1.6 trillion. Google has optimised…

Read
A Reflection on the Past DecadeArchive 2021
1 min

A Reflection on the Past Decade

You won't find the year 2020 or GPT there, but overall it's quite well done :) https://towardsdatascience.com/the-decade-of-artificial…

Read
Outputs from AI Explored at the NeurIPS Conference. What Did It Conclude?Archive 2019
1 min

Outputs from AI Explored at the NeurIPS Conference. What Did It Conclude?

Recently, the Neural Information Processing Systems (NeurIPS) conference took place in Vancouver, Canada, gathering over 13,000 scientists from various fields….

Read
Another Course CompletedArchive 2019
1 min

Another Course Completed

I have just completed the course ‘Learn BERT – the most powerful NLP algorithm by Google’. It is an advanced course that introduces…

Read
The Latest Artificial Intelligence Model for Natural Language Processing – ALBERT!Archive 2019
2 min

The Latest Artificial Intelligence Model for Natural Language Processing – ALBERT!

Among the top artificial intelligence models for natural language processing (Bert, Robert, GPT-2, or Megatron), a new player ALBERT joined the ranks last week! ALBERT is brought to us by…

Read
Nvidia Announces It Has Trained the Largest Language Model in the World, GPT-2 8B!Archive 2019
1 min

Nvidia Announces It Has Trained the Largest Language Model in the World, GPT-2 8B!

The model uses 8.3 billion parameters and is 24 times larger than BERT and 5 times larger than the previously largest GPT-2 from OpenAI. Nvidia employed parallelism that…

Read
This Week, Google Unveiled Its Latest Technological Toy – Bidirectional Encoder Representations Transformers, or BERTArchive 2018
1 min

This Week, Google Unveiled Its Latest Technological Toy – Bidirectional Encoder Representations Transformers, or BERT

How does BERT differ from traditional NLP models like word2vec and GloVe? Word2vec and other models generate context-free word embeddings. Each word...

Read