Bilgisayar ve Bilişim BilimleriİngilizceMedium

Stories by Research Graph on Medium

Stories by Research Graph on Medium
Stories by Research Graph on Medium
Ana SayfaRSS Besleme
language
Artificial-intelligenceModel-quantizationLarge-language-modelsBilgisayar ve Bilişim Bilimleriİngilizce
Yayınlandı

A novel compression technique ensuring comparable performance with 70% less parameters Author Amanda Kau ( ORCID : 0009–0004–4949–9284) Introduction The sizes of large language models (LLMs) have been steadily increasing over the last few years.

Retrieval-augmented-genArtificial-intelligenceLarge-language-modelsKnowledge-graphBilgisayar ve Bilişim Bilimleriİngilizce
Yayınlandı
Yazar Wenyi Pi

Enhancing Data Interactivity with LLMs and Neo4j Knowledge Graphs Author Wenyi Pi ( ORCID : 0009–0002–2884–2771) Introduction Since OpenAI launched ChatGPT, a large language model (LLM) based chatbot, in 2023, it has set off a technological wave.

Large-language-modelsArtificial-intelligencePrompt-engineeringBilgisayar ve Bilişim Bilimleriİngilizce
Yayınlandı

Author Dhruv Gupta ( ORCID : 0009–0004–7109–5403) Introduction Large Language Models (LLMs) have become the new face of Natural language processing (NLP). With their generative power and ability to comprehend human language, the human reliance on these models is increasing every day. However, the LLMs have been known to hallucinate and thus produce wrong outputs.

Knowledge-graphArtificial-intelligenceLarge-language-modelsBilgisayar ve Bilişim Bilimleriİngilizce
Yayınlandı

Automated Knowledge Graph Construction with Large Language Models — Part 2 Harvesting the Power and Knowledge of Large Language Models Author Amanda Kau ( ORCID : 0009–0004–4949–9284 ) Introduction Knowledge graphs (KGs) are a structured representation of data in a graphical format, in which entities are represented by nodes and are connected by edges representing relationships

MegalodonLong-textsTransformer-architectureBilgisayar ve Bilişim Bilimleriİngilizce
Yayınlandı

An improvement architecture superior to the Transformer, proposed by Meta Author · Qingqin Fang ( ORCID: 0009–0003–5348–4264) Introduction Recently, researchers from Meta and the University of Southern California have introduced a model called Megalodon. They claim that this model can expand the context window of language models to handle millions of tokens without overwhelming your memory.

Large-language-modelsArtificial-intelligenceTransformersNatural-language-processBilgisayar ve Bilişim Bilimleriİngilizce
Yayınlandı
Yazar Wenyi Pi

Understanding the Evolutionary Journey of LLMs Author Wenyi Pi ( ORCID : 0009–0002–2884–2771) Introduction When we talk about large language models (LLMs), we are actually referring to a type of advanced software that can communicate in a human-like manner. These models have the amazing ability to understand complex contexts and generate content that is coherent and has a human feel.

ModularNaiveAdvancedRetrieval-augmented-genBilgisayar ve Bilişim Bilimleriİngilizce
Yayınlandı

From Naive to Modular: Tracing the Evolution of Retrieval-Augmented Generation Author · Vaibhav Khobragade ( ORCID: 0009–0009–8807–5982) Introduction Large Language Models (LLMs) have achieved remarkable success.

Large-language-modelsRlhfFine-tuningBilgisayar ve Bilişim Bilimleriİngilizce
Yayınlandı
Yazar Xuzeng He

Supervised Fine-tuning, Reinforcement Learning from Human Feedback and the latest SteerLM Author · Xuzeng He ( ORCID: 0009–0005–7317–7426) Introduction Large Language Models (LLMs), usually trained with extensive text data, can demonstrate remarkable capabilities in handling various tasks with state-of-the-art performance. However, people nowadays typically want something more personalised instead of a general solution.

Natural-language-processiTransformersArtificial-intelligenceBilgisayar ve Bilişim Bilimleriİngilizce
Yayınlandı

Attention mechanism not getting enough attention Author Dhruv Gupta ( ORCID : 0009–0004–7109–5403) Introduction As discussed in this article, RNNs were incapable of learning long-term dependencies. To solve this issue both LSTMs and GRUs were introduced. However, even though LSTMs and GRUs did a fairly decent job for textual data they did not perform well.

Fake-newsArtificial-intelligenceLarge-language-modelsBilgisayar ve Bilişim Bilimleriİngilizce
Yayınlandı

Large Language Models for Fake News Generation and Detection Author Amanda Kau ( ORCID : 0009–0004–4949–9284) Introduction In recent years, fake news has become an increasing concern for many, and for good reason. Newspapers, which we once trusted to deliver credible news through accountable journalists, are vanishing en masse along with their writers.