LLMs and Gen AI , Part 1

LLMs and Gen AI 


LLMs (Large Language Models) focus on processing and understanding existing vast amounts of text data. They can use this understanding to perform various tasks like text generation, translation, and question answering, but the core remains working with existing information. 

They primarily rely on deep learning architectures like Transformers, which are adept at finding patterns and relationships within massive amounts of text data.
LLM helps in doing practically everything such as:

  • Content creation: Writing different creative text formats based on existing styles.
  • Chatbots and virtual assistants: Powering chatbots for natural conversations and answering questions.
  • Machine translation: Translating text between languages while preserving meaning and context.
  • Text summarization: Creating concise summaries of factual texts.
  • Some LLMs can be used for generative tasks, especially when conditional generation techniques are to be incorporated.

    For example, an LLM might be prompted to write a poem in a specific style, generating new text following that style.
    They 
    excel at working with and understanding existing textual information, using it for various tasks and sometimes even generating new text based on that understanding.

    Different types of LLMs:

    1. Generative Pre-Trained Transformer (GPT): It focuses mainly on creative text writing which can imitate writing styles.

      Eg: Open AI's GPT-3
            Google AI's Meena

    2. Factual Language Models (FLM): It works on giving accurate and informative responses, summarizing factual topics and answering the user's questions precisely.

      Eg: Gemini AI
             Fair's Bart

    3. Dialogue-focused LLMs: It is designed to understand and respond to natural language in a way that prompts human interaction.

      Eg: Microsoft's Xiacole
            Facebook's BlenderBot

    4. Code-focused LLMs: They're specifically designed to understand various computer languages and help in assisting developers write cleaner and more efficient code. 

      Eg: Codex
            Copilot

    5. Domain-Specific LLMs: These types of LLMs are trained on targeted datasets to excel at tasks and understand the unique language(jargon) used within a particular domain.

      Eg: BioBERT
            JurismAI




    Generative AI (Gen-AI) is a fascinating field of artificial intelligence focused on creating entirely new data.

    Unlike traditional AI systems that analyze existing information, Gen-AI aims to produce novel data that resembles and expands upon existing data sets.

    Generative AI often uses a system of two models working against each other:
    • Generative Model (G): This model acts like the creative artist, trying to generate new data instances that resemble the training data it's been fed on. It's like trying to create forgeries that are indistinguishable from the real thing.
    • Discriminative Model (D): This model plays the role of the critic. It analyzes both the real data (from the training set) and the forgeries created by the generative model, trying to tell them apart. As the training progresses, it gets better at spotting fakes. 
    These models are trained in an adversarial way.
    As the generative model gets better at creating realistic "fakes", the discriminative model gets better at spotting them. This pushes both models to improve iteratively, in a constant back-and-forth.

    Generative AI has the potential to revolutionize various fields, but it's important to be aware of the challenges and limitations.
    As research progresses, we can expect even more exciting applications to emerge.
    So far, it's a hypothetical model considering the amount of development it all needs but it's a highly evolving field.



    Though LLMs and Gen AI are essentially different in various things that they are trained for, in quite a few aspects, their applications overlap.




    Comments

    Popular posts from this blog

    #ChemFact1

    #ChemFact3