Hello π
Hello π#
Welcome to my knowledge hub!
It contains
past articles around machine learning and data science and
reviews of helpful resources and learning frameworks
The goal of this knowledge hub is to share what I have learned and empower people to create impactful research for mixed audiences.
Click on any of the images to pick an area of exploration:
π Computer Vision |
π Natural Language Processing |
π Data Storytelling |
π Data Pipelining |
Below is an excerpt from an article I wrote:
What is AI?
Artificial Intelligence, simply put, is the quest to make machines achieve general intelligence, where these machines can
receive data in varying size and complexity,
process it into generalized information, and
use this information to create ideas and act without human assistance.
What is ML?
Machine learning is a subset of artificial intelligence that focuses on the processing of data into information. There are two ways that machine learning produces information:
A human creates a pattern and the model evaluates how well the data fits to the pattern based on statistical inference, and
The model creates the pattern and the model and human evaluate how well the data fits into the pattern
Approach 1: Bayesian/Frequentist Statistical Modelling
These types of models can have
static estimated parameters when using an unchanging fixed dataset, or
dynamic estimated parameters.
For models with dynamic estimated parameters, an objective/loss function is used that minimizes the error in the model. In this case the variables/features stay fixed, but the observations/instances change over time, altering the data distribution in the process.
These types of machine learning models can be great in that they can be interpreted by people. However, they do not do well on unstructured data and may be limited by the researcherβs choice of variables and research objectives β they have to attempt different variable combinations to get statistically valid results that also satisfy research objectives.
Approach 2: Deep Learning
These types of models have only dynamic estimated parameters, in part, because the model itself selects variables (inputs into a paticular neuron), and also experiments with different combinations of these variables to get optimal estimated parameters.
The only key criteria that the human can control are the hyper-parameters that include the
activation function β gives the first set of estimated parameters which are usually normalized and it tends to be non-linear,
loss function,
model selection metric β i.e. we say that the model has succeeded in its task and it can stop training when a certain threshold of the model selection metric is used,
desired dimensions of the output
number of epochs β the number of times the model will go over the data in order to identify the patterns.
A great benefit of this approach is that this model type is able to handle large amounts of unstructured data very well. However, because the variables and the parameter estimation method is unknown, interpreting how it obtained its results becomes very challenging.
Natural Language Processing is an application of Deep Learning, alongside Computer Vision β think of Computer Vision as the βeyesβ of the computer and Natural Language Processing as the βears and mouthβ of the computer. Both applications specialize in unstructured data, which is increasing in size as more and more people communicate on many digital channels today.