Data Science Courses
Data Science Courses
  • 99
  • 1 009 757
Ali Ghodsi, Deep Learning, PAC Learnability in Deep Learning, Fall 2023, Lecture 20
This video provides an introduction to the fundamentals of statistical machine learning, focusing on PAC (Probably Approximately Correct) learnability and agnostic PAC learnability. It highlights the challenges faced when applying generalization theories to deep learning, offering insights into the limitations and nuances of these approaches.
Covered Topics:
- Basics of Statistical Machine Learning
- PAC Learnability
- Agnostic PAC Learnability
- Generalization Challenges in Deep Learning
Keywords:
Deep Learning, Statistical Machine Learning, PAC Learnability, Agnostic PAC, Machine Learning Theory, Generalization.
Переглядів: 1 778

Відео

Ali Ghodsi, Deep Learning, Graph Neural Newark (Part 2), Fall 2023, Lecture 19
Переглядів 1,8 тис.6 місяців тому
Exploring Graph Convolutional Networks, Graph Attention Network (GAT). Discover how GAT models apply attention mechanisms to graph-structured data, enabling precise weighting of node features for robust and adaptive network learning.
Ali Ghodsi, Deep Learning, Graph Neural Newark (Part 1), Fall 2023, Lecture 18
Переглядів 2,9 тис.6 місяців тому
Exploring Graph Convolutional Networks and ChebNet This video provides an overview of Graph Convolutional Networks (GCNs), focusing on the concept of Convolution on Graphs and its application in various GCN models such as ChebNet and Vanilla Spectral GCNs. Keywords: Convolution on Graphs, Graph Convolutional Network, ChebNet, ConvGNNs, Vanilla Spectral GCN, Chebyshev Polynomials, ChebNet Graph ...
Ali Ghodsi, Deep Learning, Diffusion Models, DDPMs, Fall 2023, Lecture 17
Переглядів 5 тис.7 місяців тому
This video delves into Denoising Diffusion Probabilistic Models (DDPM), a class of generative models that progressively refine data by reversing a diffusion process. It also touches on generative moment matching networks utilizing Maximum Mean Discrepancy (MMD) to measure the distance between distributions. Keywords: Diffusion Models, DDPM, Generative Models, MMD, Moment Matching Networks.
Ali Ghodsi, Deep Learning, GAN, Generative adversarial networks, AAE, Fall 2023, Lecture 16
Переглядів 1,3 тис.7 місяців тому
This video explores Generative Adversarial Networks (GANs), conditional GANs, mode collapse issues, Minibatch discrimination, and strategies to prevent mode collapse. It also discusses Adversarial Autoencoders (AAE) in various learning contexts and their use in clustering. Keywords: GAN, Conditional GAN, Mode Collapse, Minibatch Discrimination, Adversarial Autoencoder (AAE), Supervised AAE, Sem...
Ali Ghodsi, Deep Learning, Variational Autoencoder, VAE, Performer, Fall 2023, Lecture 15
Переглядів 2,9 тис.7 місяців тому
This lecture primarily focuses on Variational Autoencoders (VAE), beginning with an overview of variational inference and the computation of the Evidence Lower Bound (ELBO) for VAE, which starts at minute 24. Prior to this, the video discusses Performers, a variation of Transformers, based on the paper "Rethinking Attention with Performers." Keywords: Variational Autoencoders, VAE, variational ...
Ali Ghodsi, Deep Learning, RLHF, GhatGPT, Alignment in LLMs, Fall 2023, Lecture 14
Переглядів 1,2 тис.7 місяців тому
RLHF (Reinforcement Learning with Human Feedback) , Instruct GPT, ChatGPT training, efficient transformers, Performer, Large Language Models, Alignment in LLMs
Ali Ghodsi, Deep Learning, Deep Reinforcement Learning-Part 2, Deep RL, Fall 2023, Lecture 13
Переглядів 8257 місяців тому
Policy Grediant, REINFORCE algorithm, Play atari game by Deep learning, Game of Go
Ali Ghodsi, Deep Learning, Deep Reinforcement Learning-Part 1, Deep RL, Fall 2023, Lecture 12
Переглядів 1,4 тис.7 місяців тому
Introduction to Reinforcement Learning, Policy iteration, and value iteration. Markov decision process (MPD), monte carlo estimate, temporal difference
Ali Ghodsi, Deep Learning, Transformers, Fall 2023, Lecture 10
Переглядів 2,4 тис.7 місяців тому
Transformers, Encoder-Decoder, Positional embedding
Ali Ghodsi, Deep Learning, BERT and GPT, Fall 2023, Lecture 11
Переглядів 1,9 тис.7 місяців тому
Bidirectional Encoder Representations from Transformer (BERT), Generative Pre-Trained Transformer (GPT), GPT 2, GPT 3, GPT 4, T5
Ali Ghodsi, Deep Learning, Attention mechanism, self-attention, S2S, Fall 2023, Lecture 9
Переглядів 3,1 тис.8 місяців тому
Attention mechanism and self-attention, Sequence-to-sequence models This video provides an in-depth exploration of Attention Mechanism and Self-Attention, crucial concepts that have revolutionized the field of Natural Language Processing (NLP). Transformers, the game-changers in NLP, rely heavily on self-attention. Join us as we unravel the fundamentals of attention and self-attention in the co...
Ali Ghodsi, Deep Learning, Dropout, Batch Normalization, Fall 2023, Lecture 5
Переглядів 1,6 тис.8 місяців тому
Dropout, Batch normalization Batch normalization was initially inspired by the notion of internal covariate shift (ICS). However, it's now understood that Batch normalization doesn't address ICS, and we also recognize that ICS doesn't have detrimental effects. Instead, Batch normalization reduces Lipschitzness and enhances the smoothness of the loss function. This video succinctly covers dropou...
Ali Ghodsi, Deep Learning, Recurrent neural network (RNN), RNN, Fall 2023, Lecture 8
Переглядів 1,6 тис.8 місяців тому
Recurrent neural network (RNN), Backpropagation through time (BPTT), Vanishing and Exploding Gradient, Echo State Networks, Long delays, Gated RNNs, LSTM, Gated recurrent units, (GRU)
Ali Ghodsi, Deep Learning, Convolutional Neural Networks, CNN, Fall 2023, Lecture 6
Переглядів 1,6 тис.8 місяців тому
Convolutional Neural Networks, Convolution, Cross-correlation, pooling layer
Ali Ghodsi, Deep Learning, Regularization (Layer norm, FRN,TRU), Keras, Fall 2023, Lecture 7
Переглядів 1,1 тис.8 місяців тому
Ali Ghodsi, Deep Learning, Regularization (Layer norm, FRN,TRU), Keras, Fall 2023, Lecture 7
Ali Ghodsi, Deep Learning, Regularization, Fall 2023, Lecture 4,
Переглядів 2,1 тис.8 місяців тому
Ali Ghodsi, Deep Learning, Regularization, Fall 2023, Lecture 4,
Ali Ghodsi, Deep Learning, , Optimization, Fall 2023, Lecture 3
Переглядів 2,7 тис.9 місяців тому
Ali Ghodsi, Deep Learning, , Optimization, Fall 2023, Lecture 3
Ali Ghodsi, Deep Learning, feedforward neural network, Backpropagation, Fall 2023, Lecture 2
Переглядів 4,5 тис.9 місяців тому
Ali Ghodsi, Deep Learning, feedforward neural network, Backpropagation, Fall 2023, Lecture 2
Ali Ghodsi, Deep Learning, Motivation and course administrations, Fall 2023, Lecture 1
Переглядів 9 тис.9 місяців тому
Ali Ghodsi, Deep Learning, Motivation and course administrations, Fall 2023, Lecture 1
Ali Ghodsi, Lect 13 (Fall 2020): Deep learning, Transformer, BERT, GPT
Переглядів 9 тис.3 роки тому
Ali Ghodsi, Lect 13 (Fall 2020): Deep learning, Transformer, BERT, GPT
Ali Ghodsi, Lect 10 (Fall 2020): Deep learning, Attention mechanism
Переглядів 9 тис.3 роки тому
Ali Ghodsi, Lect 10 (Fall 2020): Deep learning, Attention mechanism
PyTorch tutorial
Переглядів 2,6 тис.3 роки тому
PyTorch tutorial
Pascal Poupart, Sum Product Networks, Oct 19 2017
Переглядів 2,8 тис.6 років тому
Pascal Poupart, Sum Product Networks, Oct 19 2017
Pascal Poupart, Sum Product Networks, Oct 17 2017
Переглядів 4 тис.6 років тому
Pascal Poupart, Sum Product Networks, Oct 17 2017
Ali Ghodsi, Deep Learning, Oct 12 28 2017, [Lect 6.1]
Переглядів 4,1 тис.6 років тому
Ali Ghodsi, Deep Learning, Oct 12 28 2017, [Lect 6.1]
Ali Ghodsi, Deep Learning, CNN, Oct 3, 2017 [Lect 6]
Переглядів 4,8 тис.6 років тому
Ali Ghodsi, Deep Learning, CNN, Oct 3, 2017 [Lect 6]
Ali Ghodsi, Deep Learning, Oct 5, 2017, [Lect 7]
Переглядів 4,2 тис.6 років тому
Ali Ghodsi, Deep Learning, Oct 5, 2017, [Lect 7]
Ali Ghodsi, Deep Learning, Sept 28 2017, [Lect 5]
Переглядів 4,6 тис.6 років тому
Ali Ghodsi, Deep Learning, Sept 28 2017, [Lect 5]
Ali Ghodsi, Deep Learning, Regularization, Sep 26, 2017 [Lect 4]
Переглядів 5 тис.6 років тому
Ali Ghodsi, Deep Learning, Regularization, Sep 26, 2017 [Lect 4]

КОМЕНТАРІ

  • @chaowang6903
    @chaowang6903 12 днів тому

    great stuff, do we still need egen decomposition to get lambda_max ?

  • @faezehabdolinejad
    @faezehabdolinejad 13 днів тому

    ممنونم استاد خیلی عالی بود

  • @chaowang6903
    @chaowang6903 14 днів тому

    Thank you so much for sharing your amazing course!

  • @akera2775
    @akera2775 27 днів тому

    Thanku sir for explaining this course in easy way Now I have started enjoying the course❤

  • @joshi98kishan
    @joshi98kishan Місяць тому

    Thank you professor. This lecture explains exactly what I was looking for - why principal components are the eigenvectors of the sample covariance matrix.

  • @purushottammishra3423
    @purushottammishra3423 Місяць тому

    I got answers to almost every"WHY?" that I had while reading books.

  • @amins6695
    @amins6695 Місяць тому

    Amazing Lectures

  • @HelloWorlds__JTS
    @HelloWorlds__JTS Місяць тому

    Immediately after explaining the importance of centering the data, he purposely neglects this in his first demo! But then he mentions this at 22:38, and his second demo he does center the data. Great instructions, thanks!

  • @VanshRaj-pf2bm
    @VanshRaj-pf2bm Місяць тому

    Ye lecture kis bachhe ke liye h

  • @mayankagrawal7865
    @mayankagrawal7865 2 місяці тому

    I am myself one of the person which you are claiming to be GAN generated. Please don't misled people.

  • @sripradhaiyengar9980
    @sripradhaiyengar9980 2 місяці тому

    Thank you thank you!

  • @PradeepKumar-tl7dd
    @PradeepKumar-tl7dd 3 місяці тому

    best video oh PCA

  • @mahdig4739
    @mahdig4739 3 місяці тому

    That was great Dr. Ghodsi! Many thanks!

  • @anadianBaconator
    @anadianBaconator 3 місяці тому

    I love this lecture! Finally I have a better overview of Transformers! Thank you so much prof!

  • @trontan642
    @trontan642 3 місяці тому

    Much more clear than my professor.

  • @MrFunasty
    @MrFunasty 3 місяці тому

    How can we add x(d by 1)and z(m by 1) when their shapes are different? 🙄

  • @CS_n00b
    @CS_n00b 4 місяці тому

    what is the guarantee that u1 sigma1 v1 are non negative if A is non negative?

  • @prateekpatel6082
    @prateekpatel6082 4 місяці тому

    wrong derivation of derivative of st w.r.t w , its a recursive equation , since s implicitly depends on w

  • @thomastsao7507
    @thomastsao7507 5 місяців тому

    excellent !

  • @Rasha-tc5bl
    @Rasha-tc5bl 5 місяців тому

    رائع جدا ..الله يعطيه العافيه ويسعده ويوفقه ويوفق عياله ان كان عنده عيال

  • @amirrezamohammadi
    @amirrezamohammadi 5 місяців тому

    Truly Enjoyed! Thanks

  • @bsementmath6750
    @bsementmath6750 5 місяців тому

    Prof. You used to be very verbose and invasive on the board. Why this hybrid mode of ppt and some board? Love from Pakistan!

  • @mahsakhoshnoodi2972
    @mahsakhoshnoodi2972 5 місяців тому

    Thank you for this informative lecture, I have a question though. Why the expectaion of epsilon^2 with the normal distribution of mean zero is going to be sigma^2?

    • @moodi2002
      @moodi2002 3 місяці тому

      If \( e_i \) is a random variable that is proportional to a Gaussian distribution \( N(0, \sigma^2) \), then we can write \( e_i = k \cdot X_i \), where \( X_i \) is a standard Gaussian random variable with mean \( 0 \) and variance \( \sigma^2 \), and \( k \) is a constant of proportionality. Since \( X_i \) follows a standard Gaussian distribution, its expectation \( \mathbb{E}[X_i] \) is \( 0 \), and its variance \( \text{Var}[X_i] \) is \( \sigma^2 \). Now, we want to find the expectation of \( e_i^2 \): \[ \mathbb{E}[e_i^2] = \mathbb{E}[(k \cdot X_i)^2] = k^2 \cdot \mathbb{E}[X_i^2] \] For a standard Gaussian random variable \( X_i \), the expectation of \( X_i^2 \) is the variance of \( X_i \) plus the square of its mean: \[ \mathbb{E}[X_i^2] = \text{Var}[X_i] + (\mathbb{E}[X_i])^2 = \sigma^2 + 0^2 = \sigma^2 \] So, substituting this into the expression for \( \mathbb{E}[e_i^2] \), we get: \[ \mathbb{E}[e_i^2] = k^2 \cdot \mathbb{E}[X_i^2] = k^2 \cdot \sigma^2 \] Therefore, the expectation of \( e_i^2 \) is \( k^2 \cdot \sigma^2 \).

  • @prateekpatel6082
    @prateekpatel6082 5 місяців тому

    subtle mistake for perceptron learning : we dont update gradient for correct classified point , the update happens only on mis classified point .

  • @user-fq3ms4bz2p
    @user-fq3ms4bz2p 5 місяців тому

    I think the proff wrote the wrong formula for marginal P(x=x0).

  • @user-gd8bt9qs4l
    @user-gd8bt9qs4l 5 місяців тому

    c'est un grand

  • @longh
    @longh 5 місяців тому

    thank you, professor! The explanation is very intuitive.

  • @chaowang6903
    @chaowang6903 6 місяців тому

    Great lecture on why we use testing errors for true error estimation

  • @ai__76
    @ai__76 6 місяців тому

    Thanks very much

  • @ai__76
    @ai__76 6 місяців тому

    Thanks for the useful course

  • @garmdarehalborz5441
    @garmdarehalborz5441 6 місяців тому

    Great, Thanks a lot

  • @zeynolabedinsoleymani4591
    @zeynolabedinsoleymani4591 6 місяців тому

    What is the "intuition" behind using Chebyshev polynomials in GNN rather than other orthonormal basis functions? Why it works well?

    • @Falconoo7383
      @Falconoo7383 4 місяці тому

      Because computation is very expensive.

  • @homataha5626
    @homataha5626 6 місяців тому

    Are the slides available?

  • @AmrMoursi-sm3cl
    @AmrMoursi-sm3cl 6 місяців тому

    Thanks for sharing this amazing information ❤❤❤❤

  • @user-us1jf8zd8e
    @user-us1jf8zd8e 6 місяців тому

    Detail mathematical formula explanation start @47:00

  • @kiannaderi8374
    @kiannaderi8374 6 місяців тому

    thank you

  • @mohamedmarzouk2537
    @mohamedmarzouk2537 6 місяців тому

    Thank you, very helpful and informative

  • @omidbazgir9891
    @omidbazgir9891 7 місяців тому

    Amazing lecture! Thanks for uploading the videos! Is there anyway we can have access to the coding assignments?

  • @asntrk1
    @asntrk1 7 місяців тому

    Variational Autoencoder: 24:52

  • @AmrMoursi-sm3cl
    @AmrMoursi-sm3cl 7 місяців тому

    1000000 Thanks ❤

  • @MrMIB983
    @MrMIB983 7 місяців тому

    I'm so excited to finally see a Diffusion models lecture by Professor Ali. Thank you.

  • @user-us1jf8zd8e
    @user-us1jf8zd8e 7 місяців тому

    Supervise PCA start @57:19

  • @MrMIB983
    @MrMIB983 7 місяців тому

    Amazing, this topic is new in the professor's course. It would be awesome to have an only RL course from professor Ali. Also looking forward to seeing amazing diffusion models lectures.

  • @vivekrai1974
    @vivekrai1974 7 місяців тому

    5:35 Why would it be one row of W prime? In the first case, we got a column vector of W because we multiplied a one-hot encoded vector with W. However, multiplying with h should not give one particular row as h is not a one-hot encoded vector.

  • @HarpaAI
    @HarpaAI 7 місяців тому

    🎯 Key Takeaways for quick navigation: 00:07 📚 *Introduction to GPT and BERT* - GPT and BERT are both Transformer-based models. - GPT stands for Generative Pre-Trained Transformer, while BERT stands for Bidirectional Encoder Representations from Transformers. 05:26 🧠 *How BERT Works* - BERT is a stack of encoders with multiple layers and attention heads. - It is trained by masking words in sentences and predicting the masked words, making it bidirectional in nature. 10:17 🏭 *Applications of BERT* - BERT can be used in various applications by fine-tuning its pretrained model. - It's especially useful for tasks like sentiment analysis and can handle domain-specific tasks. 14:55 🧬 *Domain-Specific BERT Models* - There are domain-specific BERT models trained for specific fields like bioinformatics and finance. - These models can be used in applications within their respective domains. 25:09 📝 *Introduction to GPT* - GPT is a stack of decoder layers, where each decoder is similar to the transformer decoder but without cross-attention. - GPT is trained to predict the next word in a sequence. 29:48 🚀 *GPT's Evolution* - GPT models have evolved over time, with each version becoming larger and more powerful in terms of parameters. - GPT-4, for instance, has an enormous 175 billion parameters, making it highly capable in natural language understanding and generation. 30:28 🧠 *Introduction to GPT 4 and its size* - Introduction to GPT 4 and its undisclosed size. - Speculation on the impact of model size on performance. 34:04 🌐 *T5: Combining BERT and GPT* - T5 is a combination of BERT and GPT. - Transformation of various NLP problems into text-to-text format. - The application of T5 to a wide range of NLP tasks. 44:12 🔐 *Challenges in Aligning Language Models with User Intent* - The challenge of aligning language models with user instructions. - The importance of alignment for ethical and practical reasons. - The need to avoid harmful or offensive responses. 49:30 🎯 *Steps for Alignment and Reinforcement Learning* - Overview of the three steps for alignment: Supervised, Fine-Tuning, and Reinforcement Learning. - Introduction to reinforcement learning from human feedback. - The importance of understanding reinforcement learning for alignment. Made with HARPA AI

  • @praveenkumar-tu1sj
    @praveenkumar-tu1sj 7 місяців тому

    It is simple sir please accept my humble request 🙏

  • @praveenkumar-tu1sj
    @praveenkumar-tu1sj 7 місяців тому

    Kindly help me sir i am failing miserably

  • @praveenkumar-tu1sj
    @praveenkumar-tu1sj 7 місяців тому

    Thanks in advanced.so that can replace actual data to get required predication i think it is very easy for a person who has knowledge of CNN/NN THANKS SIR. say example, I have lid driven cavity problem I get velocities u and v of bith sizes are 2 dimensional (say 33 by 33 example) , it is time dependent so I want to use cnn to predict u for t=25 providing u for t= 10,15,20. And I will give actual u at t=25 for comparison and add statistical regressions , loss gain, training plots. Thanks sir. Please kindly help and I would be grateful and appreciative for kindness and support.

  • @adhirajbanerjee7288
    @adhirajbanerjee7288 7 місяців тому

    any lnks to the slides in this course ?

  • @timandersen8030
    @timandersen8030 7 місяців тому

    Thank you for a new 2023 course! You are one of the best to teach the subject!!!