TY-Blog

Lucas-TY

AI

Vision

by Lucas-TY on | AI

An overview of computer vision and discusses various CNN architectures used in this field, such as Dense Layers and Convolution Layers. It highlights the key features and advantages of architectures like AlexNet, VGG, and ResNet. The concept of receptive field and its impact on the network's performance is explained. The page also mentions the challenges and solutions related to sparsity in network structures. Finally, it introduces the concept of residual learning and its benefits in optimizing the network's performance.

Transformer

by Lucas-TY on | AI

An overview of the Transformer model, a powerful architecture in the field of AI. It explains the concept of self-attention, which allows the model to focus on relevant parts of a sequence. The page covers topics such as positional encoding, applications of self-attention in speech, image, and graph processing, as well as the use of Transformers in sequence-to-sequence tasks like chatbots and question answering. The page also mentions the encoder component of the Transformer model and discusses fully connected networks, layer normalization, and the differences between layer normalization and batch normalization.

Similarity-Search

by Lucas-TY on | AI

Discusses similarity search techniques, including TF-IDF and BM25 for sparse search, and Sentence-Bert for dense search. TF-IDF calculates the product of term frequency and inverse document frequency, while BM25 considers term frequency and document length. Sentence-Bert uses cosine similarity and the Hugging Face transformer to calculate similarities between sentence embeddings.

AI-Introduction-5

by Lucas-TY on | AI

This document provides an introduction to random variables (RVs) and their distributions, including continuous and discrete RVs. It also explains Naive Bayes, a classification algorithm, and discusses the concepts of independence, the chain rule of probability, Bayes net inference, and Bayes net learning. Additionally, it covers unsupervised learning, specifically clustering using K-means, and supervised learning, including linear regression and Naive Bayes for classification.

AI-Introduction-4

by Lucas-TY on | AI

Introduces linear functions and linear regression in the context of AI. It discusses the concept of data correlation, affine functions, feature transform, and linear regression. It also explains the definition and purpose of loss functions in evaluating model fit.

AI-Introduction-3

by Lucas-TY on | AI

Provides an introduction to different data encoding techniques in AI, including "one-hot encoding" and "bag of words" (BoW). It explains the advantages and limitations of BoW and provides example code for implementing it. The document also discusses BoW frequency and normalization techniques.

AI-Introduction-2

by Lucas-TY on | AI

Focusing on learning agents and data. It discusses the reasons for using machine learning, the representation of data, different models for supervised learning, unsupervised learning, and the challenges associated with different types of data.

AI-Introduction-1

by Lucas-TY on | AI

Focusing on knowledge bases, propositional logic, and semantics. It discusses knowledge representation, reasoning, and functions such as ASK and TELL. It also covers syntax, semantics, and symbols used in propositional logic, as well as forward and backward chaining methods.