Featured image of post Large Language Models: Comparing Gen2/Gen3 Models (Bloom, Gopher, OPT and More)

Large Language Models: Comparing Gen2/Gen3 Models (Bloom, Gopher, OPT and More)

Large Language Models are sophisticated neural networks that produce texts. Since their inception in 2018, they evolved dramatically and deliver texts that can rival humans. To better understand this evolution, this blog series investigates models to uncover how they advance. Specifically, insights from published papers about each model are explained, and conclusions from benchmark comparisons are drawn.

Featured image of post Large Language Models: Comparing Gen2/Gen3 Models (GPT-3, GPT-J, MT5 and More)

Large Language Models: Comparing Gen2/Gen3 Models (GPT-3, GPT-J, MT5 and More)

Large Language Models are sophisticated neural networks that produce texts. By creating one word at a time, given a context of other words, these models produce texts that can rival a humans output. The creation of LLMs began back in 2018 when the transformer neural network architecture was discovered. Since then, ever more complex transformer models in terms of parameter amount, and continues up to this data with ever more complex model architectures, consumed amount of texts, and parametric complexity.

Featured image of post Large Language Models: Comparing Gen 1 Models (GPT, BERT, T5 and More)

Large Language Models: Comparing Gen 1 Models (GPT, BERT, T5 and More)

The creation of Large Language Models (LLMs) began in 2018. Three factors emerged and were combined in LLMs: powerful computer and graphics processing units, huge amounts of structured and unstructured data that could be processed fast, and first-grade open-source project for the creation and training of neural networks.

Featured image of post Introduction to Large Language Models

Introduction to Large Language Models

Large Language Models (LLMs) are a ubiquities technology enabling humans to use their natural language for interacting with a computer in a broad range of tasks. LLMs can answer questions about history and real-world events, they can create step-by-step tasks plans, solve mathematical questions, and can reflect on any input text to create summaries or identify text characteristics. Using most recent LLMs like GPT4 is a fascinating and surprising event.

Featured image of post Arduino Nano BLE 33 Sense Microcontroller: Hardware and GPIO Functions

Arduino Nano BLE 33 Sense Microcontroller: Hardware and GPIO Functions

The Arduino Nano BLE 33 Sense Microcontroller is an Arduino-compatible board with a fleet of onboard sensors, including sound, light, temperature, and a microphone. It can be programmed with Arduino C and MicroPython to read and write data. This article introduces this unique board, details how to use the digital and analog pins and which functions the board supports.

Featured image of post Wikipedia Article Crawler & Clustering: Text Classification with Spacy

Wikipedia Article Crawler & Clustering: Text Classification with Spacy

Spacy is a powerful NLP library that performs many NLP tasks in its default configuration, including tokenization, stemming and part-of-speech tagging. These steps can be extended with a text classification task as well, in which training data in the form of preprocessed text and expected categories as dictionary objects are provided. Both multi-label and single-label classification is supported.