Abhay Shuklaabhayshukla.com·Dec 1, 2024Luong AttentionLuong Attention is proposed in the paper Effective Approaches to Attention-based Neural Machine Translation with two variants Global Attention Local Attention The proposed NMT model architecture consists of LSTM encoder to encode input sequences ...Attention Mechanismattention-mechanism
Abhay Shuklaabhayshukla.com·Nov 30, 2024Hard, Soft and Doubly Stochastic AttentionFor the image captioning task, the paper Show, Attend and Tell: Neural Image Caption Generation with Visual Attention proposes three different attention mechanisms, Stochastic Hard Attention Deterministic Soft Attention Doubly Stochastic Attention...Attention Mechanismattention-mechanism
Abhay Shuklaabhayshukla.com·Nov 29, 2024Bahdanau AttentionThe paper Neural Machine Translation by Jointly Learning to Align and Translate formally introduced the concept of attention for the first time. One of the illustrations (shown below) presented in the paper Learning Phrase Representations using RNN E...Attention Mechanismattention-mechanism
Roberto Andrade Fonsecarandradedev.hashnode.dev·Nov 19, 2024El Camino Hacia la Comprensión Infinita: Cómo los Nuevos LLMs Revolucionan la IALos modelos de lenguaje masivo (LLMs, por sus siglas en inglés) han transformado nuestra interacción con la tecnología, desde la redacción de correos electrónicos hasta la generación de contenido creativo. Estos sistemas, entrenados con vastas cantid...27 readsgenai
William Stetarcopin43.hashnode.dev·Sep 26, 2024Understanding Transformer Networks Through a Fluid Semiotic LensBy William Stetar Abstract: This article explores the concept of transformer-based language models operating under a "fluid semiotic regime." It delves into how these models process language dynamically, adjusting meanings based on context, and argu...Semiotics
Mehul Pardeshimehulpardeshi.hashnode.dev·Sep 1, 2024Attention Is All You NeedWelcome, everyone! Today's topic is going to be very interesting and insightful. Think of it as the BIG BANG of deep learning, particularly in the field of natural language processing (NLP) and sequence modeling. The Transformer architecture has had ...21 likes·31 readsattention-mechanism
pankaj chauhanpankajchauhanblogs.hashnode.dev·Aug 30, 2024From Bag of Words to Self-Attention: The Evolution of Understanding Text in Machine LearningIntroduction: The Journey of Text Understanding In the realm of natural language processing (NLP), one of the most intriguing challenges has always been how machines can effectively understand and interpret human language. This journey has seen the e...nlp
Rashid Ul Haqrashid-ul-haq.hashnode.dev·Apr 26, 2024Unveiling dEFEND Study: How Explainable Fake News Detection WorksComponents The architecture consists of four components. (1) News content encoder (word encoder and sentence encoder) The news content encoder consists of two parts. Word encoder: In an RNN, the memory tends to diminish as the sentence length incre...Book Bytesresearch paper
Priyanshu Mohantypriyanshu-mohanty.hashnode.dev·Jan 2, 2024Snippets of Learnings from My First Research PaperNearly a year ago, I was imbued with a Herculean dilemma - whether to focus my endeavors on solely delivering results of a project or go the extra mile and attempt to present at a technical conference. All of this bewilderment for my engineering prog...140 readsDeep Learning
Parv Pareekparvpareek.hashnode.dev·Aug 27, 2023Understanding Transformer Models - Part 1| The Self Attention MechanismIntroduction In a world where cutting-edge technologies like ChatGPT, Bard, and DALL-E are making waves, you might be wondering how these technologies work. The enginge propelling these innovations is the revolutionary (In Optimus Prime voice)Transfo...1 like·95 readstransformers