Course Schedule

  1. Week 1: Intro to GAI Tools
  2. Week 2: Solve General Problems with GAI Tools
  3. Week 3: Use GAI Tools as Effective Study Aid
  4. Week 4: Apply GAI Tools to Advanced Learning Use Cases (I) - Computing Domains
  5. Week 5: Apply GAI Tools to Advanced Learning Uses Cases (II) - Non-Computing Domains
  6. Week 6: Adverse Use and Societal Implications of GAI Tools

What is AI?

A branch of computer science devoted to developing data processing systems that perform functions normally associated with human intelligence, such as reasoning, learning, and self-improvement.

Quote

The science and engineering of making intelligent machines

  • John McCarthy

What is ML?

Learns the pattern through large amounts of data and make a decision on previously unseen data.

What is Generative AI?

System that can generate new data!! Built on complex neural network

Applications:

  • Text Generation
  • Image Generation
  • Audio Generation
  • Video Generation

Natural Language Processing

Subfield of AI that focuses n enabling computers to understand, interpret, generate, and manipulate language. The language model! e.g. Speech - to - text models! Translation! Information retrieval!

Underlying Concepts in NLP

  • Tokenisation: breaking down text into individual-tokens
  • N-grams: Continuous sequence of n-items
    • e.g. bigrams [“Singapore is”, “is beautiful”]
    • Helps to capture context but may fail to capture global context
  • Embeddings: Numerical representation of words or sentences
    • common models: Word2Vec, GloVe

n-gram Probabilistic Model - Language Generation

Simple approach to language model. Estimate the probability of the nth words based on the previous (n-1) words. Key Considerations:

  1. Unseen n-grams (words that were not seen during training)
    1. smoothing
  2. Data sparsity: some n-grams may not be seen during training
  3. Limited context will affect accuracy

Neural Network Based Models

Train neural network to predict next word in the sequence ?? Recognize patterns and relationships between words.

e.g. Recurrent neural networks (RNN)

  • Maintains an internal state that summarises the history of previous inputs
  • capture temporal dependencies
  • retain context from earlier elements in the sequence
  • Limitations:
    • Vanishing gradient problem
    • Sequential computationally expensive; impacts scalability
  • Improvements on RNN:
    • Long Short-Term Memory and Gated Recurrent Unit networks

Transformers and Large Language Models

Encoder-Decoder Networks

Decoder network will generate the output one word at a time. e.g. translation, text summarization

Encoder only: classification, sentiment analysis Decoder only: Text completion & Generation, image captioning

Limitations:

  • inefficient in dealing with long input sequences
  • Loss of information

Workarounds: The Attention Mechanism!

  • Gives the encoded network a hidden state which has some sort of attention score
  • Context vector has a weighted sum which gives the decoder state an idea of which context is more important

Advantages (compared to Long Short-Term Memory and Gated Recurrent Unit Networks):

  • Efficiency
  • Interpretability
  • Global context

Transformer Architecture

Reinforcement Learning:

Reinforcement Learning Example: ChatGPT

Learning objectives

  • GAI tools are expected to disrupt the ways students learn, presenting objectives as well as challenges.
  • AI would impact 60% of advanced economy jobs!! How to adapt?
  • Learn how to simulate critical thinking and problem solving skills with the help of AI.

Reflections

  • Wow it seems that this course is planned with the help of ChatGPT !?