A knowledge hypergraph generalizes the concept of a knowledge graph. Instead of triplets, relationships can involve multiple entities (not limited to three). This allows for more complex relationships.
Knowledge Hypergraph Triplets.
In a hypergraph, a single relationship can involve more than three entities.
(Eight species of bears, are, Extant, Widespread, Northern Hemisphere, Southern Hemisphere)
(Bears, found on, Continents, North America, South America, Eurasia)
(Modern bears, have characteristics, Large bodies, Stocky legs, Long snouts, Small rounded ears, Shaggy hair, Plantigrade paws, Five nonretractile claws, Short tails)
Video Tutorial.
Main Code: Knowledge Hyper Graph with LLM-RAG.
import ollama import chromadb
documents = [ "Bears are carnivoran mammals of the family Ursidae.", "They are classified as caniforms, or doglike carnivorans.", "Although only eight species of bears are extant, they are widespread, appearing in a wide variety of habitats throughout most of the Northern Hemisphere and partially in the Southern Hemisphere.", "Bears are found on the continents of North America, South America, and Eurasia.", "Common characteristics of modern bears include large bodies with stocky legs, long snouts, small rounded ears, shaggy hair, plantigrade paws with five nonretractile claws, and short tails.", "With the exception of courting individuals and mothers with their young, bears are typically solitary animals.", "They may be diurnal or nocturnal and have an excellent sense of smell.", "Despite their heavy build and awkward gait, they are adept runners, climbers, and swimmers.", "Bears use shelters, such as caves and logs, as their dens; most species occupy their dens during the winter for a long period of hibernation, up to 100 days.", ] Kg_triplets = [["Bears", "belong to", "Family Ursidae"], ["Bears", "classified as", "Caniforms"], ["Bears", "number of species", "Eight"], ["Bears", "habitat", "Northern Hemisphere"], ["Bears", "habitat", "Southern Hemisphere"], ["Bears", "found in", "North America"], ["Bears", "found in", "South America"], ["Bears", "found in", "Eurasia"], ["Modern bears", "characteristic", "Large bodies"], ["Modern bears", "characteristic", "Stocky legs"], ["Modern bears", "characteristic", "Long snouts"], ["Modern bears", "characteristic", "Small rounded ears"], ["Modern bears", "characteristic", "Shaggy hair"], ["Modern bears", "characteristic", "Plantigrade paws with five nonretractile claws"], ["Modern bears", "characteristic", "Short tails"], ["Bears", "social behavior", "Solitary except courting and mothers with young"], ["Bears", "activity pattern", "Diurnal"], ["Bears", "activity pattern", "Nocturnal"], ["Bears", "sense", "Excellent smell"], ["Bears", "capability", "Adept runners"], ["Bears", "capability", "Adept climbers"], ["Bears", "capability", "Adept swimmers"], ["Bears", "use", "Shelters such as caves and logs"], ["Bears", "denning behavior", "Winter hibernation for up to 100 days"],] # Convert the triplets to text def triplet_to_text(triplet): txt = str(triplet[0]) +" "+str(triplet[1]) +" "+str(triplet[2]) # print(txt) return txt triplet_texts = [triplet_to_text(triplet) for triplet in Kg_triplets] # Create database client = chromadb.PersistentClient(path="E:\\Niraj_Work\\DL_Projects\\llm_projects\\database_tmp") collection = client.create_collection(name="bear_hkg") metadata = {"hnsw:space":"cosine"} # store each document in a vector embedding database for d in range(0,len(Kg_triplets)): triplet_txt = triplet_to_text(Kg_triplets[d]) response = ollama.embeddings(model="mxbai-embed-large", prompt=triplet_txt) embedding = response["embedding"] collection.add( ids=[str(d)], embeddings=[embedding], documents=[triplet_txt] )
# an example prompt prompt = "How does the bear's body looks?" # generate an embedding for the prompt and retrieve the most relevant doc response = ollama.embeddings( prompt=prompt, model="mxbai-embed-large" ) results = collection.query( query_embeddings=[response["embedding"]], n_results=3 )
# data = results['documents'][0][0] data = "" supported_docs = results['documents'] if len(supported_docs)==1: data = results['documents'][0][0] else: for i in range(0, len(supported_docs)): data = data+" "+str(supported_docs[i]) data = data.strip() # generate a response combining the prompt and data we retrieved in step 2 output = ollama.generate( model="llama3", prompt=f"Using this data: {data}. Respond to this prompt: {prompt}" )
print(output['response'])
Additional Code: Constructing Hypergraph Triplets using LLM.
import ollama import chromadb
documents = [ "Bears are carnivoran mammals of the family Ursidae.", "They are classified as caniforms, or doglike carnivorans.", "Although only eight species of bears are extant, they are widespread, appearing in a wide variety of habitats throughout most of the Northern Hemisphere and partially in the Southern Hemisphere.", "Bears are found on the continents of North America, South America, and Eurasia.", "Common characteristics of modern bears include large bodies with stocky legs, long snouts, small rounded ears, shaggy hair, plantigrade paws with five nonretractile claws, and short tails.", "With the exception of courting individuals and mothers with their young, bears are typically solitary animals.", "They may be diurnal or nocturnal and have an excellent sense of smell.", "Despite their heavy build and awkward gait, they are adept runners, climbers, and swimmers.", "Bears use shelters, such as caves and logs, as their dens; most species occupy their dens during the winter for a long period of hibernation, up to 100 days.", ] single_doc = ' '.join(documents) # an example prompt prompt = "Give the list of all Knowledge Hypergraph triplets for the following text" +"\n"+single_doc
# generate a response combining the prompt and data we retrieved in step 2 output = ollama.generate( model="llama3", # prompt=f"Using this data: {data}. Respond to this prompt: {prompt}" prompt=prompt )
A knowledge graph represents information using a graph structure where entities are nodes and relationships between entities are edges. Each piece of information is typically represented as a triplet (subject, predicate, object).
Knowledge Graph Triplets.
Each triplet involves a single subject, predicate, and object. For eg.
(Bears, are, Carnivoran mammals)
(Bears, belong to, Family Ursidae)
(Bears, classified as, Caniforms)
(Caniforms, also known as, Doglike carnivorans)
(Bears, number of species, Eight)
Video Tutorial.
Code.
1. Main code for using Knowledge Graph with LLM-RAG
import ollama import chromadb
documents = [ "Bears are carnivoran mammals of the family Ursidae.", "They are classified as caniforms, or doglike carnivorans.", "Although only eight species of bears are extant, they are widespread, appearing in a wide variety of habitats throughout most of the Northern Hemisphere and partially in the Southern Hemisphere.", "Bears are found on the continents of North America, South America, and Eurasia.", "Common characteristics of modern bears include large bodies with stocky legs, long snouts, small rounded ears, shaggy hair, plantigrade paws with five nonretractile claws, and short tails.", "With the exception of courting individuals and mothers with their young, bears are typically solitary animals.", "They may be diurnal or nocturnal and have an excellent sense of smell.", "Despite their heavy build and awkward gait, they are adept runners, climbers, and swimmers.", "Bears use shelters, such as caves and logs, as their dens; most species occupy their dens during the winter for a long period of hibernation, up to 100 days.", ] Kg_triplets = [["Bears", "are", "carnivoran mammals"], ["Bears", "belong to", "family Ursidae"], ["Bears", "are classified as", "caniforms"], ["Caniforms", "are", "doglike carnivorans"], ["Bears", "have", "eight species"], ["Bears", "are", "widespread"], ["Bears", "appear in", "a wide variety of habitats"], ["Bears", "are found in", "Northern Hemisphere"], ["Bears", "are partially found in", "Southern Hemisphere"], ["Bears", "are found on", "North America"], ["Bears", "are found on", "South America"], ["Bears", "are found on", "Eurasia"], ["Modern bears", "have", "large bodies"], ["Modern bears", "have", "stocky legs"], ["Modern bears", "have", "long snouts"], ["Modern bears", "have", "small rounded ears"], ["Modern bears", "have", "shaggy hair"], ["Modern bears", "have", "plantigrade paws with five nonretractile claws"], ["Modern bears", "have", "short tails"], ["Bears", "are typically", "solitary animals"], ["Bears", "can be", "diurnal"], ["Bears", "can be", "nocturnal"], ["Bears", "have", "an excellent sense of smell"], ["Bears", "are", "adept runners"], ["Bears", "are", "adept climbers"], ["Bears", "are", "adept swimmers"], ["Bears", "use", "shelters"], ["Shelters", "include", "caves"], ["Shelters", "include", "logs"], ["Bears", "use", "shelters as dens"], ["Most species", "occupy", "dens during winter"], ["Bears", "hibernate for", "up to 100 days"],] # Convert the triplets to text def triplet_to_text(triplet): txt = str(triplet[0]) +" "+str(triplet[1]) +" "+str(triplet[2]) # print(txt) return txt # triplet_texts = [triplet_to_text(triplet) for triplet in Kg_triplets] # Create database client = chromadb.PersistentClient(path="E:\\Niraj_Work\\DL_Projects\\llm_projects\\database_tmp") collection = client.create_collection(name="bear_kg") metadata = {"hnsw:space":"cosine"} # store each document in a vector embedding database for d in range(0,len(Kg_triplets)): triplet_txt = triplet_to_text(Kg_triplets[d]) response = ollama.embeddings(model="mxbai-embed-large", prompt=triplet_txt) embedding = response["embedding"] collection.add( ids=[str(d)], embeddings=[embedding], documents=[triplet_txt] )
# an example prompt prompt = "How does the bear's body looks?" # generate an embedding for the prompt and retrieve the most relevant doc response = ollama.embeddings( prompt=prompt, model="mxbai-embed-large" ) results = collection.query( query_embeddings=[response["embedding"]], n_results=3 ) print("result = ",results) print(collection.get(include=['embeddings','documents','metadatas']))
# data = results['documents'][0][0] data = "" supported_docs = results['documents'] if len(supported_docs)==1: data = results['documents'][0][0] else: for i in range(0, len(supported_docs)): data = data+" "+str(supported_docs[i]) data = data.strip() # generate a response combining the prompt and data we retrieved in step 2 output = ollama.generate( model="llama3", prompt=f"Using this data: {data}. Respond to this prompt: {prompt}" )
print(output['response'])
2. Code to Construct Knowledge Graph.
import ollama import chromadb
documents = [ "Bears are carnivoran mammals of the family Ursidae.", "They are classified as caniforms, or doglike carnivorans.", "Although only eight species of bears are extant, they are widespread, appearing in a wide variety of habitats throughout most of the Northern Hemisphere and partially in the Southern Hemisphere.", "Bears are found on the continents of North America, South America, and Eurasia.", "Common characteristics of modern bears include large bodies with stocky legs, long snouts, small rounded ears, shaggy hair, plantigrade paws with five nonretractile claws, and short tails.", "With the exception of courting individuals and mothers with their young, bears are typically solitary animals.", "They may be diurnal or nocturnal and have an excellent sense of smell.", "Despite their heavy build and awkward gait, they are adept runners, climbers, and swimmers.", "Bears use shelters, such as caves and logs, as their dens; most species occupy their dens during the winter for a long period of hibernation, up to 100 days.", ] single_doc = ' '.join(documents) # an example prompt prompt = "Give the list of all Knowledge Graph triplets for the following text" +"\n"+single_doc
# generate a response combining the prompt and data we retrieved in step 2 output = ollama.generate( model="llama3", # prompt=f"Using this data: {data}. Respond to this prompt: {prompt}" prompt=prompt )
Retrieval-Augmented Generation (RAG) enhances Large Language Models (LLMs) by integrating retrieval mechanisms with generative capabilities. It improves response accuracy by accessing external databases for relevant information, overcoming LLM limitations in knowledge cut-off and hallucinations. RAG combines the strengths of retrieval (precise, up-to-date data) and generation (contextual, fluent language), making it essential for complex queries, factual correctness, and dynamic knowledge. It optimizes performance, especially in specialized or rapidly evolving fields, ensuring comprehensive, accurate, and contextually relevant outputs, thus significantly enhancing the utility and reliability of LLMs in practical applications.
In this tutorial, I delve into the key topics and advancements related to Retrieval-Augmented Generation (RAG) using executable codes. Each topic is meticulously explained through video tutorials, offering detailed yet accessible discussions and working demonstrations, accompanied by the corresponding code. Some code segments are sourced from relevant libraries for demonstration purposes. This comprehensive tutorial covers the following topics, providing an in-depth understanding of RAG and its practical applications.
Basics of RAG (Retrieval Augmented Generation) with LLM
How to use LLM + RAG to Construct Knowledge Graph.
How to construct Flow-Diagram by Using LLM + RAG.
Graph Based RAG (Retrieval Augmented Generation) Techniques.
Note: In addition to this, it also provides a linked tutorial on a pressing topic: "Use of Long Text Sequences with LLMs Trained on Shorter Text Sequences.". In the future, I will introduce new research advancements in the field of Large Language Models.
1. Basics of RAG (Retrieval Augmented Generation) with LLM.
Video Tutorial.
Basic RAG Code.
import ollama import chromadb
documents = [ "Quantum mechanics is a fundamental theory in physics that describes the behavior of nature at and below the scale of atoms.", "It is the foundation of all quantum physics, which includes quantum chemistry, quantum field theory, quantum technology, and quantum information science.", "Quantum mechanics can describe many systems that classical physics cannot.", "Classical physics can describe many aspects of nature at an ordinary (macroscopic and (optical) microscopic) scale, but is not sufficient for describing them at very small submicroscopic (atomic and subatomic) scales.", "Most theories in classical physics can be derived from quantum mechanics as an approximation valid at large (macroscopic/microscopic) scale.", "Quantum systems have bound states that are quantized to discrete values of energy, momentum, angular momentum, and other quantities, in contrast to classical systems where these quantities can be measured continuously.", "Measurements of quantum systems show characteristics of both particles and waves (wave–particle duality), and there are limits to how accurately the value of a physical quantity can be predicted prior to its measurement, given a complete set of initial conditions (the uncertainty principle)." ] # Create database client = chromadb.Client() collection = client.create_collection(name="docs") # store each document in a vector embedding database for i, d in enumerate(documents): response = ollama.embeddings(model="mxbai-embed-large", prompt=d) embedding = response["embedding"] collection.add( ids=[str(i)], embeddings=[embedding], documents=[d] )
# an example prompt prompt = "What are the key benefits of using quantum mechanics over classical physics?" # generate an embedding for the prompt and retrieve the most relevant doc response = ollama.embeddings( prompt=prompt, model="mxbai-embed-large" ) results = collection.query( query_embeddings=[response["embedding"]], n_results=1 ) data = results['documents'][0][0]
# generate a response combining the prompt and data we retrieved in step 2 output = ollama.generate( model="llama3", prompt=f"Using this data: {data}. Respond to this prompt: {prompt}" )
print(output['response'])
2. How to use LLM + RAG to Construct Knowledge Graph..
Video Tutorial.
Code to Generate Knowledge Graph Triplets.
import ollama import chromadb
documents = [ "Quantum mechanics is a fundamental theory in physics that describes the behavior of nature at and below the scale of atoms.", "It is the foundation of all quantum physics, which includes quantum chemistry, quantum field theory, quantum technology, and quantum information science.", "Quantum mechanics can describe many systems that classical physics cannot.", "Classical physics can describe many aspects of nature at an ordinary (macroscopic and (optical) microscopic) scale, but is not sufficient for describing them at very small submicroscopic (atomic and subatomic) scales.", "Most theories in classical physics can be derived from quantum mechanics as an approximation valid at large (macroscopic/microscopic) scale.", "Quantum systems have bound states that are quantized to discrete values of energy, momentum, angular momentum, and other quantities, in contrast to classical systems where these quantities can be measured continuously.", "Measurements of quantum systems show characteristics of both particles and waves (wave–particle duality), and there are limits to how accurately the value of a physical quantity can be predicted prior to its measurement, given a complete set of initial conditions (the uncertainty principle)." ]
# store each document in a vector embedding database for i, d in enumerate(documents): response = ollama.embeddings(model="mxbai-embed-large", prompt=d) embedding = response["embedding"] collection.add(ids=[str(i)], embeddings=[embedding], documents=[d] )
# an example prompt prompt1 = "What are the key benefits of using quantum mechanics over classical physics?" prompt2 = "List all entities, and generate the knowledge graph triplets by using all entities." # Generate Answers - for Prompt-1: # generate an embedding for the prompt and retrieve the most relevant doc response1 = ollama.embeddings( prompt=prompt1, model="mxbai-embed-large" ) results1 = collection.query( query_embeddings=[response1["embedding"]], n_results=1 ) data1 = results1['documents'][0][0]
# generate a response combining the prompt and data we retrieved in step 2 output1 = ollama.generate( model="llama3", prompt=f"Using this data: {data1}. Respond to this prompt: {prompt1}" ) print("Response for the question -1",output1['response']) # Generate Answers - for Prompt-2: # generate an embedding for the prompt and retrieve the most relevant doc response2 = ollama.embeddings( prompt=prompt2, model="mxbai-embed-large" ) results2 = collection.query( query_embeddings=[response2["embedding"]], n_results=1 ) data2 = results2['documents'][0][0]
# generate a response combining the prompt and data we retrieved in step 2 output2 = ollama.generate( model="llama3", prompt=f"Using this data: {data2}. Respond to this prompt: {prompt2}" ) print("Response for the question -2",output2['response'])
Code to Visualize the Knowledge Graph (by using above triplets).
import networkx as nx import matplotlib.pyplot as plt
# Step 1: Define your triplets triplets = [ ("Quantum mechanics", "a fundamental theory", "Theory"), ("Theory", "in physics", "Physics"), ("Physics", "describes the behavior of", "Nature"), ("Nature", "at and below the scale of", "Atoms"), ("Atoms", "is related to the scale of", "Scale"), ]
# Step 2: Create a directed graph G = nx.DiGraph()
# Step 3: Add edges from triplets for subject, predicate, obj in triplets: G.add_edge(subject, obj, label=predicate)
# Step 4: Draw the graph pos = nx.spring_layout(G, seed=42) # Position nodes using Fruchterman-Reingold force-directed algorithm # Draw nodes and edges nx.draw(G, pos, with_labels=True, node_size=3000, node_color="lightblue", font_size=10, font_weight="bold", arrowsize=20)
# Display the graph plt.title("Knowledge Graph Visualization") plt.show()
3. How to construct Flow-Diagram by Using LLM + RAG.
Video Tutorial.
Code.
import ollama import chromadb
documents = [ "Quantum mechanics is a fundamental theory in physics that describes the behavior of nature at and below the scale of atoms.", "It is the foundation of all quantum physics, which includes quantum chemistry, quantum field theory, quantum technology, and quantum information science.", "Quantum mechanics can describe many systems that classical physics cannot.", "Classical physics can describe many aspects of nature at an ordinary (macroscopic and (optical) microscopic) scale, but is not sufficient for describing them at very small submicroscopic (atomic and subatomic) scales.", "Most theories in classical physics can be derived from quantum mechanics as an approximation valid at large (macroscopic/microscopic) scale.", "Quantum systems have bound states that are quantized to discrete values of energy, momentum, angular momentum, and other quantities, in contrast to classical systems where these quantities can be measured continuously.", "Measurements of quantum systems show characteristics of both particles and waves (wave–particle duality), and there are limits to how accurately the value of a physical quantity can be predicted prior to its measurement, given a complete set of initial conditions (the uncertainty principle)." ] # Create database client = chromadb.Client() collection = client.create_collection(name="docs")
# store each document in a vector embedding database for i, d in enumerate(documents): response = ollama.embeddings(model="mxbai-embed-large", prompt=d) embedding = response["embedding"] collection.add(ids=[str(i)], embeddings=[embedding], documents=[d] )
# an example prompt prompt1 = "Generate a Mermaid diagram." # Generate Answers - for Prompt-1: # generate an embedding for the prompt and retrieve the most relevant doc response1 = ollama.embeddings( prompt=prompt1, model="mxbai-embed-large" ) results1 = collection.query( query_embeddings=[response1["embedding"]], n_results=1 ) data1 = results1['documents'][0][0]
# generate a response combining the prompt and data we retrieved in step 2 output1 = ollama.generate( model="llama3", prompt=f"Using this data: {data1}. Respond to this prompt: {prompt1}" ) print("Response for the question -1",output1['response'])
4. Graph Based RAG (Retrieval Augmented Generation) Techniques.
Video Tutorial.
Code.
import ollama import chromadb
documents = [ "The use of retrieval-augmented generation (RAG) to retrieve relevant information from an external knowledge source enables large language models (LLMs) to answer questions over private and/or previously unseen document collections.", "However, RAG fails on global questions directed at an entire text corpus, such as “What are the main themes in the dataset?”, since this is inherently a queryfocused summarization (QFS) task, rather than an explicit retrieval task.", "Prior QFS methods, meanwhile, fail to scale to the quantities of text indexed by typical RAG systems.", "To combine the strengths of these contrasting methods, we propose a Graph RAG approach to question answering over private text corpora that scales with both the generality of user questions and the quantity of source text to be indexed.", "Our approach uses an LLM to build a graph-based text index in two stages: first to derive an entity knowledge graph from the source documents, then to pregenerate community summaries for all groups of closely-related entities.", "Given a question, each community summary is used to generate a partial response, before all partial responses are again summarized in a final response to the user.", "For a class of global sensemaking questions over datasets in the 1 million token range, we show that Graph RAG leads to substantial improvements over a na¨ıve RAG baseline for both the comprehensiveness and diversity of generated answers." ] # Create database client = chromadb.Client() collection = client.create_collection(name="docs") # store each document in a vector embedding database for i, d in enumerate(documents): response = ollama.embeddings(model="mxbai-embed-large", prompt=d) embedding = response["embedding"] collection.add( ids=[str(i)], embeddings=[embedding], documents=[d] )
# an example prompt prompt = "How Graph RAG (Retrieval Augmented Generation, used with Large Language Model) generates a Global Summarization for the given context?" Entity_1 = "Graph RAG" Entity_2 = "Global Summarization" Community_Triplets = [("Graph RAG", "Uses", "Leiden Community Detection Algorithm"), ("LLM", "Extracts", "Entity Knowledge Graph"), ("Graph Index", "Partitioned By", "Community Detection Algorithms"), ("Community Summaries", "Used For", "Global Summarization"),] summary_text = "Graph RAG - Uses - Leiden Community Detection Algorithm; LLM - Extracts - Entity Knowledge Graph; Graph Index - Partitioned By - Community Detection Algorithms; Community Summaries - Used For - Global Summarization" # generate an embedding for the prompt and retrieve the most relevant doc response = ollama.embeddings( prompt=prompt, model="mxbai-embed-large" ) summary_text_embedding = ollama.embeddings( prompt=summary_text, model="mxbai-embed-large" ) results = collection.query( query_embeddings=[summary_text_embedding["embedding"]], n_results=1 ) data = results['documents'][0][0]
# generate a response combining the prompt and data we retrieved in step 2 output = ollama.generate( model="llama3", prompt=f"Using this data: {data}. Respond to this prompt: {prompt}" )
Ding, Yujuan, Wenqi Fan, Liangbo Ning, Shijie Wang, Hengyun Li, Dawei Yin, Tat-Seng Chua, and Qing Li. "A Survey on RAG Meets LLMs: Towards Retrieval-Augmented Large Language Models." arXiv preprint arXiv:2405.06211 (2024).
Wu, Kevin, Eric Wu, and James Zou. "How faithful are RAG models? Quantifying the tug-of-war between RAG and LLMs' internal prior." arXiv preprint arXiv:2404.10198 (2024).
Li, Jiarui, Ye Yuan, and Zehua Zhang. "Enhancing LLM Factual Accuracy with RAG to Counter Hallucinations: A Case Study on Domain-Specific Queries in Private Knowledge-Bases." arXiv preprint arXiv:2403.10446 (2024).
Edge, Darren, Ha Trinh, Newman Cheng, Joshua Bradley, Alex Chao, Apurva Mody, Steven Truitt, and Jonathan Larson. "From Local to Global: A Graph RAG Approach to Query-Focused Summarization." arXiv preprint arXiv:2404.16130 (2024).
The unbeatable aspects of Wasserstein Generative Adversarial Networks (WGANs) come from their significant improvements over traditional GAN architectures. They address critical challenges like mode collapse and enhance the generation of high-quality, diverse samples. The following are the key technical advancements in WGAN architecture that motivate me to create tutorials on WGAN:
Wasserstein Distance: Shifts from traditional metrics to the Wasserstein distance for more meaningful training gradients, reducing mode collapse and stabilizing network convergence.
Weight Clipping and Lipschitz Constraint: Initially, WGANs used weight clipping to meet the Lipschitz constraint for the Wasserstein distance, but this approach had drawbacks like capacity underuse and gradient problems. The WGAN-GP variant introduced a gradient penalty to overcome these issues, leading to better training stability and sample quality.
Gradient Penalty (WGAN-GP): Incorporates a gradient penalty in the loss function, promoting stable training and high-quality output by preventing excessive critic gradients.
Critic Role: Unlike traditional GANs' discriminators, WGAN critics assess generated sample quality on a continuous scale, enabling finer quality evaluation and aiding in model training dynamics.
Training Protocol: WGANs employ a distinct training method, often involving more frequent training of the critic than the generator to provide effective gradients, ensuring balanced learning and model stability.
These advancements make WGANs superior for generating realistic samples and ensuring smoother model training, maintaining their unique position in AI research and development.
Video Tutorials.
Part-1
Part-2
Part-3
Code - Training WGAN
# example of training a wgan on mnist from numpy import expand_dims import keras import keras.backend as K import tensorflow as tf import numpy as np from keras import Model from keras.optimizers import Adam from keras.layers import Input, Reshape, Flatten from keras.layers import Dense, BatchNormalization, Conv2D, Conv2DTranspose, LeakyReLU, Dropout batch_size = 32 input_shape = (28, 28, 1) latent_dim = 100 img_shape = (28, 28, 1) class WGAN_1: def __init__(self): print("welcome to WGAN coding") # write code for wasserstein loss. def wasserstein_loss(self, y_true, y_pred): return K.mean(y_true * y_pred)
def preprocess_real_part_training_dataset(self): # load mnist dataset (dataX, dataY), (testDX, testDY) = keras.datasets.fashion_mnist.load_data() # Select the first 1000 rows of training data and labels dataX = dataX[:1000] dataY = dataY[:1000] # Add an additional dimension for the grayscale channel by using expand_dims() from NumPy dataX = expand_dims(dataX, axis=-1) # convert from unsigned ints to floats and scale from [0,255] to [0,1] dataX = dataX.astype(np.float32) / 255.0 return dataX
# example of loading the generator model and generating images import numpy as np from keras.models import load_model from numpy.random import randn from keras.models import load_model from matplotlib import pyplot import matplotlib.pyplot as plt # load model model = load_model('g_model.h5') # Generate synthetic images num_images = 10 latent_dim = 100 noise = np.random.normal(0, 1, (num_images, latent_dim)) generated_images = model.predict(noise)
# Plot the generated images plt.figure(figsize=(10, 10)) for i in range(num_images): plt.subplot(1, num_images, i+1) plt.imshow(generated_images[i, :, :, 0], cmap='gray') plt.axis('off') plt.show()
Reference:
1. Wasserstein GAN; Martin Arjovsky (Courant Institute of Mathematical Sciences), Soumith Chintala, and Leon Bottou1 (Facebook AI Research)
3. Kwon, Dohyun, Yeoneung Kim, Guido Montúfar, and Insoon Yang. "Training Wasserstein GANs without gradient penalties." arXiv preprint arXiv:2110.14150 (2021).
4. Guo, Xin, Johnny Hong, Tianyi Lin, and Nan Yang. "Relaxed Wasserstein with applications to GANs." In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 3325-3329. IEEE, 2021.
Training large language models (LLMs) on longer
sequences poses challenges in computational resources, model complexity,
gradient propagation, and overfitting. These include increased memory requirements
due to self-attention mechanisms, longer training times, difficulty in scaling
Transformers for very long sequences, challenges in capturing long-term
dependencies, risk of vanishing or exploding gradients, and potential
overfitting to training data. Solutions like linear biases, RoFormer, and RoPE
improve handling of long-range dependencies, enhance model generalization, and
incorporate positional information for better performance in NLP tasks.
For Example:
Attention with linear Biases
Improved Handling
of Long-Range Dependencies. Traditional attention mechanisms struggle with
capturing long-range dependencies in text due to the quadratic increase in
computational complexity with sequence length. Linear biases help to mitigate
this by effectively incorporating positional information, thus enhancing the
model’s ability to maintain context over long distances within the text.
RoFormer
Improved Model Generalization: By more effectively encoding
positional information, RoFormer helps LLMs to generalize better across
different tasks and datasets. This results in enhanced performance on a wide
range of NLP tasks, including text classification, machine translation, and
semantic analysis.
Enhanced Positional Encoding: RoPE uniquely integrates
positional information with the token embeddings, preserving the relative
distances between tokens. This method enables the model to better understand
and utilize the order of words or tokens, which is crucial for many language
understanding and generation tasks.
Video Tutorial -1
Video Tutorial -2
Video Tutorial -3
References.
Su, Jianlin, Murtadha Ahmed, Yu Lu, Shengfeng
Pan, Wen Bo, and Yunfeng Liu. "Roformer: Enhanced transformer with rotary
position embedding." Neurocomputing 568 (2024): 127063.
Press, Ofir, Noah A. Smith, and Mike Lewis.
"Train short, test long: Attention with linear biases enables input length
extrapolation." arXiv preprint arXiv:2108.12409 (2021).
Vaswani, Ashish, Noam Shazeer, Niki Parmar,
Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser, and Illia
Polosukhin. "Attention is all you need." Advances in neural information processing systems 30 (2017).
GANs, called Generative Adversarial Networks, are special types of deep learning models that have two main parts: a generator and a discriminator. The generator creates fake data, and the discriminator checks if this data looks real or not compared to real data. By training against each other, GANs get better at making data that looks real, changing how we make new images, expand datasets, and learn without supervision. The following points show the significance of GAN in the area of AI.
Creative Applications: GANs create realistic images, music, text, and videos, enabling creativity in art, content, and virtual environments.
Data Augmentation: GANs generate synthetic data to enhance small datasets, improving the performance of machine learning models.
Defense Against Deepfakes: GANs are used to develop defenses and detect manipulated media content amidst the growth of deepfake technology.
Drug Discovery and Molecular Design: GANs play an increasing role in drug discovery, producing novel molecular structures with desired properties, potentially transforming the pharmaceutical industry.
Scope.
This article comprises interactive video tutorials and code demonstrations to elucidate the GAN architecture. The discussion covers various topics, culminating in the presentation of straightforwardly designed code.
GAN Part-1
GAN Part-2
GAN Part-3.
Keras implementation of GAN
The following contains the Kera implementation of Deep Convolution GAN (DCGAN). Please go through the above video tutorials, to properly understand and use the code.
The system is built using Python 3.10 and relies on several essential library dependencies:
Tensorflow (version 2.15)
tqdm (version 4.66.2)
h5py (version 3.10)
Keras (version 2.115)
Train DCGAN.
# example of training a gan on mnist from numpy import expand_dims from tqdm import tqdm import keras import tensorflow as tf import numpy as np from keras import Model from keras.optimizers import Adam from keras.layers import Input, Reshape, Flatten from keras.layers import Dense, BatchNormalization, Conv2D, Conv2DTranspose, LeakyReLU, Dropout batch_size = 32 input_shape = (28, 28, 1) latent_dim = 100 img_shape = (28, 28, 1) class GAN_1: def __init__(self): print("welcome to GAN coding") # This code prepares a TensorFlow dataset for training by shuffling the data, batching it into # consistent batch sizes, and prefetching batches to optimize data loading during training. def preprocess_real_part_training_dataset(self, batch_size): # load mnist dataset (dataX, dataY), (testDX, testDY) = keras.datasets.fashion_mnist.load_data() # Add an additional dimension for the grayscale channel by using expand_dims() from NumPy dataX = expand_dims(dataX, axis=-1) # convert from unsigned ints to floats and scale from [0,255] to [0,1] dataX = dataX.astype(np.float32) / 255.0 # testDX = testDX.astype(np.float32) / 255.0 trainX = tf.data.Dataset.from_tensor_slices(dataX).shuffle(1000) # Combines consecutive elements of this dataset into batches. trainX = trainX.batch(batch_size, drop_remainder=True).prefetch(1) return trainX
# example of loading the generator model and generating images import numpy as np from keras.models import load_model from numpy.random import randn from keras.models import load_model from matplotlib import pyplot import matplotlib.pyplot as plt # load model model = load_model('g_model.h5') # Generate synthetic images num_images = 10 latent_dim = 100 noise = np.random.normal(0, 1, (num_images, latent_dim)) generated_images = model.predict(noise)
# Plot the generated images plt.figure(figsize=(10, 10)) for i in range(num_images): plt.subplot(1, num_images, i+1) plt.imshow(generated_images[i, :, :, 0], cmap='gray') plt.axis('off') plt.show()
NOTE: This code is not intended for any commercial use. It is created solely for simple educational purposes.