Introduction to Azure AI Fundamentals (AI-900)
- Overview of the Azure AI Fundamentals certification targeting aspiring AI engineers and data scientists.
- Emphasizes foundational knowledge of Azure AI services without requiring in-depth ML expertise.
- Preparation includes labs, lectures, and practice exams tailored for success.
Certification Path and Study Recommendations
- Suggested learning routes:
- Obtain Complete Microsoft Azure Developer Associate (AZ-204) Study Guide for cloud foundational knowledge.
- Consider DP-900 (Azure Data Fundamentals) for a data-focused background.
- AI-900 serves as a strong base before advanced certifications like Azure AI Engineer or Azure Data Scientist.
- Study time varies:
- Beginners: 20-30 hours.
- Intermediate (with some Azure or ML experience): 8-10 hours.
- Experienced cloud professionals: ~5 hours.
- Recommended study plan: 30–60 minutes daily over 14 days.
- Using both theoretical lectures and hands-on labs enhances learning retention.
Exam Structure and Key Domains
- Exam duration: 60 minutes; passing score roughly 70%.
- Question types: multiple choice, drag and drop, hot area; no penalty for wrong answers.
- Domains and weighting:
- AI workloads and considerations (15-20%)
- Fundamental machine learning principles on Azure (20-25%)
- Computer vision features (15-20%)
- Natural language processing (NLP) features (15-20%)
- Generative AI workloads (15-20%)
- Focus lies mostly on describing capabilities and concepts rather than complex problem-solving.
Core Azure AI Concepts Covered
Artificial Intelligence, Machine Learning, and Deep Learning
- AI: Computer systems mimicking human cognitive functions.
- ML: Algorithms improving from experience without explicit programming.
- Deep Learning: Neural networks with multiple hidden layers for complex data processing.
Azure Cognitive Services
- Prebuilt AI capabilities enabling image analysis, language understanding, speech processing, and decision-making.
- Key services include Computer Vision, Form Recognizer, Text Analytics, Speech Services, Q&A Maker, and Language Understanding (LUIS).
Computer Vision and Related Services
- Image classification, object detection, facial recognition, and optical character recognition (OCR).
- Azure Custom Vision service allows custom image model creation with easy tagging and training.
Natural Language Processing (NLP)
- Services for sentiment analysis, key phrase extraction, entity recognition, translation, and conversational AI.
- Language Understanding Intelligent Service (LUIS) for intent classification and entity extraction.
Responsible AI Principles
- Microsoft’s six principles: Fairness, Reliability & Safety, Privacy & Security, Inclusiveness, Transparency, and Accountability.
- Practical examples and frameworks for ethical AI development.
Machine Learning Pipelines and Automation
- Azure Machine Learning Studio enables building, training, deploying ML models.
- Automated ML (AutoML) simplifies model development with automatic feature engineering and model selection.
- Visual Designer facilitates no-code pipeline creation.
- Programmatic pipeline management via Python SDK.
Generative AI and Azure Open AI Service
- Introduction to generative AI models like GPT, DALL·E for text, code, and image generation.
- Explanation of Transformer architectures, tokens, embeddings, positional encoding, and attention mechanisms.
- Azure OpenAI Studio for managing and testing large language models.
Co-pilots and Prompt Engineering
- Co-pilots leverage generative AI to assist with document creation, coding, and data synthesis.
- Prompt engineering refines inputs to achieve precise outputs from AI models.
- Grounding techniques use contextual information to improve AI responses.
Practical Lab Examples and Setup
- Hands-on exercises using Azure ML Studio, cognitive services SDKs, and custom vision projects.
- Setting up compute resources, notebooks (Jupyter Labs), and key Azure services.
- Demonstrations of training, testing, and deploying ML models.
- Integration of Q&A Maker and Azure Bot Service for conversational applications.
Exam Preparation Tips
- Study both the conceptual and practical components.
- Use available practice exams and labs to reinforce knowledge.
- Focus on understanding service capabilities, AI concepts, and responsible AI.
- Manage time effectively during the exam, expect scaled scoring.
This comprehensive summary equips learners with a structured understanding to confidently prepare for and pass the AI-900 Azure AI Fundamentals certification exam while gaining practical skills in Azure’s AI ecosystem.
For a broader understanding of cloud certifications and practical projects, consider the Comprehensive AWS and Azure Cloud Computing Course Summary and Key Projects.
hey this is Andrew Brown and I'm bringing you another certification course and this time it's the Azure AI
fundamentals also known as the AI 900 and if you're looking to pass a certification we have everything that
you need here such as Labs lectures and a free practice exam so you can go ASAT exam get that uh certification put on
your resume and Linkedin to go get that job you've been looking for um if you want to support more free courses like
this one the best way is to purchase the additional uh paid materials where you can get access uh to more practice exams
and other resources if you don't know me um I've taught a bit of everything uh here on the cloud that's been with adabs
Azure gcp devops terraform kubernetes you name it I've taught it but uh you know you know the drill here let's get
into it and learn more about the Azure AI fundamentals hey this is Andrew Brown
from exam Pro and we are at the start of our journey here learning about the AI 900 asking the most important question
which is what is the AI 900 so the Azure AI fundamental certification is for those seeking an ml role such as AI
engineer or data scientist the certification will demonstrate if a person can Define and understand Azure
AI services such as competive services and Azure applied AI Services AI Concepts knowledge mining responsible AI
basics of NL pipelines classical ml models autom ml generative AI workloads which is newly added content and Azure
AI studio so you don't need to know super complicated ml knowledge here but it definitely helps to get you through
there so this certification is generally referred to by its course code the AI 900 and it's the natural path for the
Azure AI engineer or Azure data scientist certification this generally is an easy course to pass and it's great
for those new to cloud or ml related technology looking at our road map you might be asking okay well what are the
paths and what should I learn first so here are a few suggested routes if you already have your a900 that's that's a
great starting point before you take your AI 900 if you don't have your a z900 you can jump right into the AI 900
but I strongly recommend you go get that a z900 because it gives you General foundational knowledge it's just another
thing that you should not have to worry about which is just how to use Azure at a fundamental level do you need the
dp900 to take the AI 900 no but a lot of people seem to like to go this route where they want to have that data
Foundation before they move on to the AI 900 because they know that the broad knowledge is going to be useful they so
it's app pairing that you see a lot of people getting the AI 900 and the dp900 together for the AI 900 the path is a
little bit more clear it's either going to be data scientists or AI engineer so for the AI engineer you have to know how
to use the AI services in and out for data scientists it's more focused on setting up actual pipelines and things
like that within Azure machine learning so you just have to decide which path is for you the data scientist is definitely
harder than the AI engineer so if you aren't ready for the data science some people like taking the AI engineer first
and then doing the data scientists so this is kind of like a warmup again it's not 100% necessary but it's just based
on your personal learning style and a lot of times people like to take the data engineer after the data scientists
just to round out their complete knowledge now if you already have the a900 and the administrator associate you
can safely go to the data scientist if you want to risk it because this one is really hard so if you've passed the
a1004 you know you're going to probably have a lot more confidence learning up about all the concepts at this level
here but of course it's always recommended to go grab these foundational CTS because sometimes
course materials just do not cover the information and so the obvious stuff is going to get left out okay so moving
forward here how long should you study to pass for the AI 900 if you're entirely new to ml Ai and Cloud
providers such as Azure you should anticipate dedicating around 15 hours to grasp the basics this estimate can vary
base on your familiarity with these concepts for complete beginners the time commitment might extend to 20 to 30
hours for the intermediate level so people that have passed the a900 or dp900 you're looking at around 8 to 10
hours if you have one or more years of experience with Azure or another cloud service provider like a WS or gcp you're
looking at about 5 hours or less the average stunny time is about 8 hours this is where you should be committing
50% of the time to the lecture in labs and 50% for the practice exams the recommended study time is 30 minutes to
an hour a day for 14 days this should get you through it but just don't overstudy and just don't spend too
little time what does it take to pass the exam well you got to watch the lectures and memorize key information do
Hands-On labs and follow along with your own Azure account I'd say that you could probably get away with just watching all
the videos in this one without having to do the labs but again it really does reinforce that information if you do
take the time there is some stuff that is in Azure AI Studio or machine learning you might be wary of launching
instances because we do have to run instances and they will cost money unless you delete the instances after
use resulting in very small costs so if you feel that you're not comfortable with that by just watching the videos
you should be okay but when you get into the associate tier you absolutely have to expect to pay something to learn and
take that risk you want to do paid online practice exams that simulate the real exam as I've mentioned before I do
provide a free practice exam and have paid practice exams that accompany this course that are on my platform exam Pro
and that's how you can help support more of these free courses so can you pass this certification without taking a
practice exam well azzure is a little bit harder if this isn't a WS exam I would say yes but for Azure exams like
AI 900 dp900 and sc900 probably not it's kind of risky I think you should do at least one practice exam or go through
the sample one there's a sample one probably laying around on the Azure website let's take a look at the exam
guide breakdown here and then in the following video we'll look at in more detail so it's broken down into the
following domain so the exam has five domains of questions and each domain has its own waiting which determines how
many questions in a domain that will show up so 15 to 20% will be described Ai workloads and considerations 20 to
25% will consist of describe fundamental principles of machine learning on Azure 15 to 20% will consist of described
features of computer vision workloads on Azure 15 to 20% will be described features of natural language processing
workloads on Azure and 15 to 20% will be described features of generative AI workloads on Azure I want you to notice
it's says describe these domains this is good because that tells you it's not going to be super hard if you start
seeing things that say Beyond describe and identify then you know it's going to be a bit harder so where do you take
this exam well you can take it in person at a test center or online from the convenience of your own home so there's
two popular test centers there's CER aort and there's Pearson view you can also take it at a local test center if
there are nearby locations the term Proctor means a supervisor or person that is monitoring you while you're
taking the exam if I had the the option between in person or online I would always choose the in person because it's
a controlled environment and it's way less stressful online there are many things that can go wrong but it's up to
your personal preference and your situation the passing grade here is 700 out of a th000 so that's around 70% I
would say around because you could possibly fail with 70% because these things work on scaled scoring for
response types there's about 37 to 47 questions and you can afford to get about 10 to 13 questions wrong so some
questions are worth more than one point some questions cannot be skipped and the format of questions can be multiple
choice multiple answer drag and drop and hot area there shouldn't be any case studies for foundational level exams and
there's no penalty for wrong questions so for the duration you get 1 hour that means about 1 minute per question the
time for this exam is 60 Minutes your C time is 90 minutes C time refers to the amount of time that you should take to
allocate for that exam so this includes time to review the instructions read and accept the NDA complete the exam and
provide feedback at the end this certification is going to be valid forever and it does not expire Microsoft
fundamental certifications such as the a900 or ms9900 do not expire as long as the technology is still available or
relevant so we'll proceed to the full exam guide [Music]
now hey this is Andrew Brown from exam Pro and what we've pulled up here is the official exam outline on the Microsoft
website if you want to find this year s you just have to type in AI 900 Azure or Microsoft you should be able to easily
find it the page looks like this what I want you to do is scroll on down because we're looking for the AI 900 set of
guide and from there we're going to scroll on down to the skills measured section and you might want to bump up
the text Azure loves updating their courses with minor updates that don't generally affect the outcome of the
study here but it does get a lot of people worried because they always say well is your course out of date so no
they're just making minor changes because they'll do this like five times a year and so if there was a major
revision what would happen is they would change it so instead of being the AI 900 it would be like the AI 9001 or 9002
similar to how the AI 102 was previously AI 100 but now it's the AI 102 so just watch out for those and if it's a major
revision then yes it would probably need a completely new course so there aren't any major changes with the new update
other than the update for the generative AI workloads on Azure section A couple of name changes and a few things being
removed everything else remains is relatively the same with very minor changes so the concepts and such are
still up to date overall I think the exam is easier than the four so let's go through some of the topics and work our
way through here so describe Ai workloads and considerations so here we're just kind of describing the
generalities of AI so content moderation workloads involve filtering out inappropriate or harmful content from
user generated inputs ensuring a safe and positive user experience personalization workloads analyze user
behavior and preferences to tailor content recommendations or experiences to individual users computer vision
workloads involve the analysis of images and videos to recognize patterns objects faces and actions identify natural
language processing knowledge mining document intelligence and features of generative AI workloads note that these
are all just Concepts you don't need to know how to use the services at a high level then you have the responsible AI
section so Microsoft has these six principles that they really want you to know and they push it throughout all
their AI services so those are the six you'll need to know and they're not that hard to
learn moving on we have described fundamental principles of machine learning on
Azure so here it's just describing regression classification clustering and features of deep learning we have a lot
of practical experience with these in the course so you will understand at the end what these are used for next we have
core machine learning Concepts we can identify features and labels in a data set so that's the data labeling service
describe how training validation data sets are used in machine learning so we'll touch on that describe
capabilities of Automated machine learning automl simplifies building and picking the best models while data and
compute Services provide the power you need for training with Azure machine learning it helps with managing and
deploying your models letting you put your machine learning projects into action smoothly under computer vision
workloads we have image classification object detection optical character recognition facial detection and facial
analysis Solutions next we have Azure AI Vision Azure AI face detection and Azure AI
video indexer the Azure AI Services Encompass a wide range of tools designed to facilitate the development of
intelligent applications these Services used to be called computer vision custom vision face service and form recognizer
but have Evol or been grouped under broader service categories to streamline their application and integration into
projects for NLP we have key phrase extraction entity recognition sentiment analysis language modeling speech
recognition synthesis this one doesn't really appear much it's kind of a concept not so much something we have to
do and then there's translation so now we have Azure tools and services for NLP workloads these
include the Azure AI language service Azure AI speech service and Azure AI translator service these used to be
separate Services I believe like the text analytics service Lewis speech service and translator text service but
they have been added to the Azure AI umbrella of AI services and now we'll be moving on to the generative AI workloads
on Azure we'll be covering features of generative AI models common scenarios for generative Ai and responsible AI
considerations for generative Ai and also some of the cool features that Azure open ey service has to offer such
as natural language generation code generation and image generation so that's about a general
breakdown of the AI 900 exam [Music] guide hey this is angrew Brown from exam
Pro and we are looking at the layers of machine learning so here I have this thing that looks like kind of an onion
and what it is it's just describing the relationship between these uh ml terms uh uh related to Ai and we'll just work
our way through here starting at the top so artificial intelligence also known as AI is when machines that perform jobs
that mimic human behavior so it doesn't describe uh how it does that but it's just the fact that that's what AI is uh
one layer underneath we have machine learning so machines that get better at a task without explicit programming uh
then we have deep learning so these are machines that have an artificial neural network inspired by the human brain to
solve complex problems and if you're talking about someone that actually assembles either ml or or deep learning
uh models or algorithms that's a data scientist so a person with multi-disciplinary skills and math
statistics predictive modeling machine learning to make future predictions so what you need to understand is that AI
is just the outcome right and so AI could be using ml underneath or deep learning or a combination of both or
just IFL statements okay [Music] all right so let's take a look here at
the key elements of AI so AI is the software that imitates human behaviors and capabilities and there are key
elements according to Azure or Microsoft as to what makes up AI so let's go through this list quickly here so we
have machine learning which is the foundation of an AI system that can learn and predict like a human you have
anomaly detection so detect outliers or things out of place like a human computer vision be able to see like a
human natural language processing also known as NLP be able to process human languages in referr contexts you know
like a human at conversational AI be able to hold a conversation with a human so you know I wrote here according to
Microsoft and Azure because you know the the global definition is a bit different but I just wanted to put this here
because I've definitely seen this as an exam question and so we're going to have to go with azure's definition here okay
let's define what is a data set so a data set is a logical grouping of units of data that are closely related to or
share the same data structure and there are publicly available data sets that are used in uh learning of Statistics
data analytics and machine learning I just want to cover a couple here so the first is the mnist database so images of
handwritten digits used to test classify cluster image processing algorithms commonly used when learning uh how to
build computer vision ml models to translate handwritten into or handwriting into digital text so it's
just a bunch of handwritten uh numbers and letters and then another very popular data set is the common objects
in context Coco data set so this is a data set which contains many common images using a Json file Coco format
that identify objects or segments within an image uh and so this data set has a lot of stuff in it so object
segmentations recognition and it contexts super pixel stuff segmentation they have a lot of images and a lot of
objects uh so there's a lot of stuff in there so why am I talking about this and in particular Coco data sets well when
you use um Azure machine Learning Studio it has a DAT data labeling service and um the thing is is that uh it can
actually export out into Coco formats that's why I wanted you to get exposure to what Coco was and the other thing is
is that when you're building out Azure machine learning uh pipelines you uh they actually have open data sets which
we'll see later in the course um that shows you that you can just use very common ones and so uh you might see mest
and uh the other one there uh so I just wanted to get you some exposure [Music]
okay let's talk about data labeling so this is the process of identifying raw data so images text files videos and
adding one or more meaningful and informative labels to provide context so a machine learning model can learn so
with supervised machine learning labeling is a prerequisite to produce training data and each piece of data
will generally be labeled by a human the reason why I say generally here is because with azure's uh data labeling
Service uh they can actually do ml assisted labeling uh so with unsupervised machine learning labels
will be produced by the machine and may not be human readable uh and then one other thing I want to touch on is the
term called Ground truth so this is a proper u a properly labeled data set that you can use as the objective
standard to train and assess a given model is often called Ground truth the accuracy of your train model will depend
on the accuracy of your ground Truth Now using um azures tools I never seen use the word ground truth I see that a lot
in AWS and even this graphing here is from AWS but uh I just want to make sure you are familiar with all that stuff
[Music] okay let's compare supervised unsupervised and reinforcement learning
starting at the top we got supervised learning this is where the data has been labeled for training and it's considered
Tas driven because you are trying to make a prediction get a value back so when the labels are known and you want a
precise outcome when you need a specific value returned and so you're going to be using classification and regression in
these cases for unsupervised learning this is where data that has not been labeled uh the ml model needs to do its
own labeling this is considered data driven it's trying to recognize a structure or a pattern and so this is
when the labels are not known and the outcome does not need to be prise when you're trying to make sense of data so
you have clustering dimensionality reduction and associ if You' never heard this term before the
idea is it's trying to reduce the amount of Dimensions to make it easier to work with the data so make sense of the data
right uh we have reinforcement learning so this is where there is no data there's an environment and an ml model
generates data uh and and makes many attempts to reach a goal so this is considered uh decisions driven and so
this is for game AI learning tasks robot navigation when you've seen someone code a video game that can play itself that's
what this is if you're wondering this is not all the types of machine learning uh and these are in specific unsupervised
and supervised is considered classical machine learning because they he heavily rely on statistics and math to produce
the outcome uh but there you [Music] go so what is a neural network well it's
often described as mimicking the brain it's a neuron or node that represents an algorithm so data is inputed into a
neuron and based on the output the data will be passed to one of many connected neurals the connections between neurons
is weighted I really should have highlighted that one that's very important uh the network is organized
into layers there will be an input layer uh one to many hidden layers and an output layer so here's an example of a
very simple neural network notice the NN a lot of times you'll see this in ml as an abbreviation for neural networks and
sometimes neural networks are just called neural Nets so just understand that's the same term here what is deep
learning this is a neural network that has three or more hidden layers it's considered deep learning because at this
point it's uh it's not human readable to understand what's going on with within those
layers what is forward feed so neural networks where they have connections between nodes that do not form a cycle
they always move forward so that just describes uh a a forward pass through the network you'll see fnn which stands
for forward feed neural network just to describe that type of network uh then there's back back propagation which are
in forward feed uh networks this is where we move backwards through the neural net adjusting the weights to
improve the outcome on next iteration this is how a neural net learns the way the back propagation knows to do this is
that there's a loss function so a function that compares the ground truth to the prediction to determine the error
rate how bad the network performs so when it gets to the end it's going to perform that calculation and then it's
going to do its back propagation and adjust the weights um then you have activation functions I'm just going to
uh clear this up here so activation functions uh they're an algorithm applied to a hidden layer uh node that
affects connected output so for this entire hidden layer they'll all have the same uh one here and it just kind of
affects uh how it learns and like how the waiting works so it's part of back propagation and just the learning
process there's a concept of D so when the next layer increases the amount of nodes and you have spars so when the
next layer decreases the amount of nodes anytime you see something going from a dense layer to a sparse layer that's
usually called dimensional dimensionality reduction because you're reducing the amount of Dimensions
because the amount of nodes in your network determines the dimensions you have
okay [Music] what is a GPU well it's a general
processing unit that is specially designed to quickly uh render high resolution images and videos
concurrently gpus can perform parallel operations on multiple sets of data so they are commonly used for non-graphical
tasks such as machine learning and scientific computation so a CPU has an average of four to 16 processor cores a
GPU can have thousands of processor cores so something that has 408 gpus could have as many as 40,000 cores
here's an image I grabbed right off the Nvidia website and so it really illustrates very well uh like how this
would be really good for machine learning or U neural networks because no networks have a bunch of nodes they're
very repetitive tasks if you can spread them across a lot of cores that's going to work out really great so gpus are
suited uh for repetitive and highly paralleled Computing tasks such as rendering Graphics cryptocurrency mining
deep learning and machine [Music] learning we're talking about Cuda before
we can let's talk about what Nvidia is so Nvidia is a company that manufactures graphical processor units for gaming and
professional markets if you play video games you've heard of Nvidia so what is Cuda it is the compute unified device
architecture it is a parallel Computing platform in API by Nvidia that allows developers to use Cuda enable gpus for
general purpose Computing on gpus so gpg all major deep learning Frameworks are integrated with Nvidia deep uh
learning SDK the Nvidia uh deep learning SDK is a collection of Nvidia libraries for deep learning one of those libraries
is the Cuda deep neural network library so CNN so Cuda or CNN provides highly tuned implementations for standard
routines such as forward and back uh convolution convolution is really great for um uh uh computer vision pooling
normalization activation layers uh so you know in the Azure certification uh for the AI 900 uh they're not going to
be talking about Cuda but if you understand these two things you'll understand why gpus uh really matter
[Music] okay all right let's get a uh easy introduction into machine learning
pipeline so this one is definitely not an exhaustive one and we're definitely going to see more complex ones uh
throughout this course but let's get to it here so starting on the left hand side we might start with data labeling
this is very uh important when you're doing supervis learning because you need to label your data so the ml model can
learn by example during training uh this stage and the feature engineering stage are is considered pre-processing because
we are preparing our data to be trained for the model uh when we move on to feature engineering the idea here is
that ml models can only work with numerical data so you'll need to translate it into a format that it can
understand so extract out the important data that the ml model needs to focus on okay uh then there's the training steps
so your model needs to learn how to become smarter it will perform multiple iterations getting smarter with each
iteration uh you might also have a hyperparameter tuning uh step here it says tunning but
it should say tuning um but the ml model can have different parameters so you can use ml to try out many different
parameters to optimize the outcome when you get to deep learning it's impossible to tweak the parameters by hand so you
have to use hyperparameter tuning then you have serving sometimes known as deploying uh but you know when we say
deploy we talk about the entire pipeline not necessarily just the the ml model step so we need to make an ml model
accessible so we serve it by hosting in a virtual machine or container uh when we're talking about Azure um machine
learning it's either going to be an Azure kubernetes service or Azure container instance and you have uh
inference so inference is the active request uh of requesting to make a prediction so you send your payload with
either CSV or whatever and you get back the results you have a real time endpoint and batch processing so real
time is just they can batch can be real as well but generally it's slower but the idea is that do I am I making a
single item prediction or am I giving you a bunch of data at once and again this is a very simplified ml pipeline
I'm sure we'll revisit ml pipeline later in this [Music]
course so let's compare the uh the terms forecasting and prediction so forecasting you make a prediction with
relevant data it's great for analysis of Trends uh and it's not guessing and when you're talking about prediction this is
where you make a prediction without relevant data you use statistics to predict future outcomes it's more of
guessing and it uses decision Theory so imagine you have a bunch of data and the idea is you're going to infer from that
data okay maybe it's a maybe it's B maybe it's C and for prediction you don't have really much data so you're
going to have to uh kind of invent it and the idea is that you'll figure out what the outcome is there these are
extremely broad terms but you know just so you have a highle view of these two things
okay so what are performance or evaluation metrics well they are used to evaluate
different machine learning algorithms the idea is uh you know when your machine learning makes a prediction
these are the metrics you're using to evaluate to determine you know is your ml model working as you intended so for
different types of problems different metrics matter this is absolutely not an exhaustive list I just want you to get
you exposure to these uh words and things uh so that when you see them you go okay I'll come back here and refer to
this uh but lots of these you're just it's not it's not necessarily to remember but classification metrics you
should know so classification we have accuracy precision recall F1 score rock and a for regression metrics we have MSE
R msce Mae ranking metrics we have MMR dcg and dcg statistical metrics we have correlation computer vision metrics we
have psnr ssim IOU NLP metrics we have perplexity blue medor Rogue deep learning related metrics we have
Inception score I cannot say this person's name but I'm assuming it's a person but uh this Inception distance
and there are two categories of evaluation metrics we have internal evaluations so metrics used to evaluate
the internals of an ml model so accuracy F1 score Precision recall I call them the famous four using all kinds of uh
models and uh external evaluation metrics used to evaluate the final prediction of an ml model so yeah uh
don't get too worked up here I know that's a lot of stuff uh the ones that matter we will see again again
[Music] okay let's take a look at Jupiter notebook so these are web-based
applications for authoring documents that combin live code narrative text equations visualizations uh so if you're
doing data science or you're building ml models you absolutely are going to be working with jupyter notebooks they're
always integrated into uh cloud service providers ml tools um uh so jupyter notebook actually came about from
IPython so IPython is the precursor of it and they extracted that feature out it became jupyter notebook I IPython is
now a kernel uh to run uh python so when you execute out python code here it's using IPython which is just a version of
python uh jupyter notebooks were overhauled and better integrated into an IDE called Jupiter Labs which we'll talk
about here in a moment and you generally want to open notebooks in Labs the Legacy webbased interface is known as
Jupiter classic notebooks so this is what the old one looks like you can still open them up but everyone uses
Jupiter Labs now okay so let's talk about Jupiter labs jupyter labs is the next generation web-based user interface
all familiar features of the classic Jupiter notebook uh is in a flexible powerful user interface it has notebooks
a terminal a text editor a file browser Rich outputs Jupiter Labs will eventually replace the classic uh
Jupiter notebooks so there you [Music] go we keep mentioning regression but
let's talk about it in uh more detail here so we kind of understand the concept so regression is the process of
finding a function to equate a labeled data set notice it says labeled that means it's going to be for supervised
learning into a continuous variable number so another way to say it is predict this variable in the future so
the future is just means like that continuous variable doesn't have to be time but that's just a good example of
regression so what will the temperature be next week so will it be 20 Celsius how would we determine that well we
would have vectors so dots that are plotted on a graph that has multiple Dimensions the dimensions could be
greater than just X and Y you could have uh many uh and then you have a regression Line
This is the line that's going through our data set and uh and that's going to help us uh figure out um how to predict
the value so how would we do that well we would need to calculate the distance of a vector from the regression line
which is called an error and so different regression algorithms use uh the error to predict different variable
future variables so just to look at this graphic here so here is our regression line and here is a a DOT like a a vector
a piece of information and this distance from the line the the actual distance is what we're going to use in our ml model
to figure out if we were to plot another line up here right you know we would compare this line to all the other lines
okay and that's how we'd find similarity and what we'll commonly see for this is mean squared error root mean squared
error mean absolute error so MSE mrse and Mae [Music]
okay let's take a closer look at the concepts of classification so classification is the process of finding
a function to divide a labeled data set so again this is supervised learning into classes or categories so predict a
category to apply to the inputed data so will it rain next Saturday will it be sunny or rainy so we have our data set
and the idea is we're drawing through this a classification line to divide the data set so we're regression we're
measuring the line two or the the the vectors to the line and this one it's just what side of the line is it on if
it's on this side then it's sunny if it's on this side it's rainy okay for classification algorithms we got log
logistic regression decision trees random forests neural networks uh naive Bays K nearest neighbor also known as
knnn and support Vector machines svms [Music] okay let's take a closer look at
clustering so clustering is the process of grouping unlabeled data so unlabeled data means it's unsupervised learning
based on similarity and differences so the outcome could be group data based on similarities or differences I guess it's
the same description up here uh but imagine we have a graph and we have data and the idea is we draw boundaries
around that to see uh similar groups so maybe we're recommending purchases to Windows computers or recommending
purchase to Mac computers now remember this is unlabeled data so the label is being inferred or um or they're just
saying these things are similar right so clustering algorithms we got K means k medoids a density Bas hierarchial
[Music] okay hey this is Andrew Brown from exam Pro and we're looking at the confusing
Matrix and this is a table to visualize the model predictions the predicted versus the ground truth labels the
actual also known as an error Matrix and they are useful for classification problems to determine if our um if our
classification is working as we think it is so imagine we have a question how many bananas did this person eat or
these people eat and so we have this kind of a box here where we have predicted versus actual and it's really
comparing the ground truth and what the model predicted right and so on the exam they'll ask you questions like okay well
imagine that uh and they might not even say yes or no maybe like zero and one and so what they're saying is you know
imagine you have you want to tell us the true positives right and so the idea is they won't show you the labels here but
you know one and one would be a true positive and zero and Z would be a false negative okay another thing they'll ask
you about these uh confusion matrixes is uh the size of them so the idea is that we're looking right now at a um oops
just going to erase that there but we are looking at a binary classifier because we have one label and uh uh just
two labels right one and two okay but you could have three say one two and three so how would you calculate that
well there would just be a third cell over here uh you know and it's just going to be actual predicted because
we're only going to have ground Truth Versus prediction and so that's how you'll know it will be six the size will
be six might not say cells but it'll just say six [Music]
okay so to understand anomaly detection let's define quickly what is an anomaly so an abnormal thing that is marked by
deviation from the norm or standard so anomaly detection is the process of finding outliers within a data set
called an anomaly so detecting when a piece of data or access pattern appear suspicious or malicious so use
cases for Nom detection can be data cleaning intrusion detection fraud detection system Health monitoring event
detection and sensory or sensor networks ecosystem disturbances detection of critical and cascading flaws Anomaly
detections by hand is a very tedious process so using ml for anomaly detection is more efficient and accurate
and Azure has a service called anomaly detector detects anomalies and data to quickly find uh quick identify and
troubleshoot [Music] issues so computer vision is when we use
machine learning neural networks to gain high level understanding of digital images or videos so for computer vision
deep learning algorithms we have convolutional neural networks these are for image and video recognition they're
inspired after how the human eye actually processes information and sends it back to the brain to be processed you
have recurrent neural networks rnns which are generally used for handri recognition or speech recognition of
course these algorithms have other applications but these are the most common use cases for them for types of
computer vision we have image classification so look at an image or video and classify its place in a
category object detection so identify objects within an image or video and apply labels and location boundaries
semantic segmentation so identify segments or objects by drawing pixel masks around them so great for objects
and movement image analysis so analyze uh an image or video to apply descriptive context uh labels so maybe
an employee is sitting at a desk in Tokyo would be uh something that image analysis would do optical character
recognition or OCR find texts in images or videos and extract them into digital text for editing facial detection so
detect faces in a photo or video and draw a location boundary uh and label their expression so for computer vision
to some things around Azure or Microsoft Services there's one called seeing AI it's an AI app developed by Microsoft
for iOS and so you use your device camera to identify OB uh people and objects and the app is audibly describes
those objects for people with visual impairments it's totally free if you have an IOS app I have an Android phone
so I cannot use it but I hear it's great some of the Azure computer vision service offering is computer vision so
analyze images and videos extract descriptions tags objects and texts custom Vision so custom uh image
classification object detection models using your own images face so detect and identify people uh emotions and images
form recognizer so translate scan documents into key value or tabular editable
[Music] data so natural language processing also known as NLP is machine learning that
can understand the context of a corpus Corpus being a body of related text so nlps enable you to analyze and interpret
text within documents and email messages interpret or contextualize spoken tokens so for example maybe customer sentiment
analysis whether customer is happy or sad synthesize speech so a voice assistant uh assistant talking to you
automatically translates spoken or written phrases and sentences between languages interpret spoken or written
commands and determine appropriate actions a very famous example for a voice assistant specifically or virtual
assistant for Microsoft is Cortana uh it uses the Bing search engine to perform tasks such as setting reminders and
answering questions for the user uh and if you're on a Windows 10 machine uh it's very easy to activate Cortana by
accident uh when we're talking about azures MLP offering we have text and analytics so sentiment analysis to find
out what customers think find topic uh topic relevant phrases using key phrase extraction identify the language of the
text with the language detection detect and categorize entities in your text with named entity recognition for
translator we have real-time text translation multilanguage support uh for speech service we have transcribe
audible speech into readable searchable texts and then we have language understanding also known as
Lewis uh natural language processing service that enables you to understand human
language in your own application website chatbots iot device and more when we talk about conversational AI it usually
generally uses NLP so that's where you'll see that overlap next [Music]
okay let's take a look here at conversational AI which is technology that can participate in conversations
with humans so we have chat Bots voice assistants and interactive voice recognition systems which is like the
the second version to interactive voice response system so you know when you call in and they say press these numbers
that is a response system and a recognition system is when they can actually take human uh Speech and
translate that into action so the use cases here would be online customer support replaces human agents for uh for
replying about customer FAQs maybe shipping questions anything about customer support accessibility so voice
operate UI for those who are uh visually impaired HR processes so employee training onboarding updating employee
information I've never seen it used like that but that's what they say as a use case Healthcare accessible affordable
healthare so maybe you're doing a claim process I've never seen this but maybe in the US where you do more your claims
and everything is privatized it makes more sense Internet of Things So iot devices so Amazon Alexa Apple Siri
Google home and I suppose Cortana but it doesn't really have a particular device so that's why I didn't list it there
computer software so autocomplete search on phone or desktop so that would be Cortana something it could do uh for the
two services that are around conversation AI for Azure we have Q&A maker so create a conversational
question and answer bot from your existing content also known as a knowledge base and Azure bot service
intelligent serverless bot service that scales on demand used for creating publishing managing bots so uh the idea
is you make your Bot here and then you deploy it with this [Music]
okay let's take a look here at responsible AI which focuses on ethical transparent and accountable uses of AI
technology Microsoft put into practice responsible AI via its six Microsoft AI principles this whole thing is invented
by Microsoft uh and so you know it's not necessarily a standard but it's something that Microsoft is pushing hard
to uh have people adopt okay so we the first thing we have is fairness so this is an AI system which should treat all
people fairly we have reliability and safety an AI system should perform reliably and safely privacy and security
AI system should be secure and respect privacy inclusiveness AI system should Empower everyone and engage people
transparency AI systems should be understandable accountability people should be accountable for AI systems and
we need to know these in uh uh greater detail so we're going to have a a short little video on each of these
[Music] okay the first on our list is fairness so AI systems should treat all people
fairly so an AI system can reinforce existing social social stere uh stereotypical bias can be introduced uh
during the development of a pipeline so an A system that are used to allocate or withhold opportunities resources or
information uh in domains such as criminal justice employee employment and hiring finance and credit so an example
here would be an ml model designed to select a final applicant for hiring pipeline without incorporating any bias
based on gender ethnicity or may result in unfair Advantage so Azure ml can tell you how each feature can influence a
model's prediction for bias uh one thing that could be of use is fair learn so it's an open source python project to
help data science is to improve fairness in the AI systems at the time of I made this course a lot of their stuff is
still in preview so you know it's the fairness component is is not 100% there but it's great to see that they're
getting that along okay so we are on to our second AI principle for Microsoft and this one is
AI systems should perform reliably and safely so AI software must be rigorously tested to ensure they work as expected
before release to the end user if there are scenarios where AI is making mistakes it is important to release a
report Quantified risks and harms to end users so they are informed of the shortcomings of an AI solution something
you should really remember for the exam they'll definitely ask that AI wear concern uh for reliability safety for
humans is critically important autonomous vehicles a health diagnosis a suggestion prescriptions and autonomous
weapon systems they didn't mention this in their content and I was just like doing some additional research research
I'm like yeah you really don't want mistakes when you have automated weapons or ethnically you shouldn't have them at
all but hey that's uh that's just how the world works but yeah this is this category
[Music] here we're on to our third Microsoft AI principle AI system should be secure and
respect privacy so AI can require vast amounts of data to train deep machine ml models the nature of an ml model may
require personally identifiable information so piis uh it is important that we ensure
protection of user data that it is not leak or disclosed in some cases ml models can be run locally on a user's
device so their uh piis remain on their device avoiding the the vulnerability this is called this is like Edge
Computing so that's the concept there AI security principles to malicious actors so data origin and lineage data use
internal versus external data corruption considerations anomaly detection so there you
[Music] go we're on to the fourth Microsoft AI principle so AI systems should Empower
everyone and engage people if we can design AI Solutions for the minority of users they can design a solution for the
majority of users so when we're talking about minority groups we're talking about physical ability gender sexual
orientation ethnicity other factors this one's really simple uh in terms of practicality it doesn't 100% make sense
because if you've worked with um uh groups that are deaf and blind developing technology for them a lot of
times they need specialized Solutions uh but the approach here is that you know if we can design for the minority we can
design for all that is uh the principle there so that's what we need to know okay
let's take a look here at transparency so AI systems should be understandable so interpretability and intelligibility
is when the end user can understand the behavior of UI so transparency of AI systems can result in mitigating
unfairness help developers debug their AI systems ging more trust from our users those build a those who build AI
systems should be open about why they're using AI open about the limitations of the AI systems adopting an open source
AI framework can provide transparency at least from a technical perspective on the internal workings of an AI
system we are on to the last Microsoft AI principle here people uh should be accountable for AI systems so the
structure put in place to consistently enacting AI principles and taking them into account AI systems should work
within Frameworks of governments organizational principles ethical and legal standards that are clearly defined
principles guide Microsoft and how they develop sell and Advocate when working with third parties and this push towards
regulation towards a principles so this is Microsoft saying hey everybody adopt our model um there aren't many other
model so I guess it's great that Microsoft is taking the charge there I just feel that it needs to be a bit more
welldeveloped but what we'll do is look at some more practical examples so we can better understand how to apply their
principles okay so if we really want to understand how to apply the Microsoft AI principles
they've great created this nice little tool via a free web app for practical scenarios so they have these cards you
can read through these cards they're color coded for different scenarios and there's a website so let's go take a
look at that and see what we can learn [Music] okay all right so we're here on the
guidelines for human AI interaction so we can better understand the uh how to put into practice the Microsoft AI
principles they have 18 cards and let's work our way through here and see the examples the first one our list make
clear what the system can do help the users understand what the AI system is capable of doing so here PowerPoint
quick start Builders an on uh Builds an online outline to help you get started researching a subject it display uh
suggested topics that help you understand the features capability then we have the Bing app shows examples of
types of things you can search for um Apple watch displays all metrics it tracks and explains how going on the
second card we have make clear how well the system can do what it can do so here we have office new uh companion
experience ideas dock alongside your work work and offers one-click assistance with grammar design Data
Insights richer images and more the unassuming term ideas coupled with label previews help set expectations and
presented suggestions the recommender in apple music uses language such as we will think you'll like to communicate
uncertainty the help page for Outlook web mail explains the filtering into focused and other and we'll start
working right away but we'll get better with use making clear the mistakes uh will happen and you teach the product
and set overrides onto our red cards here we have time Services based on context time when to act or interrupt
based on the user's current task and environment when it's time to leave for appointments Outlook sends a time to
leave notification with directions for both driving and public transit taking into account current location event
location real-time traffic information um and then we have after using Apple Maps routing it remembers
when you're parked your car when you open the app after for a little while it suggests routing to the location of the
park car all these Apple examples make me think that Microsoft has some kind of partnership with apple I guess I guess
Microsoft or or Bill Gates did own Apple shares so maybe they're closer than we think uh show contextually relevant
information time when to act or interrupt based on user's current task and environment powered by Machine
learning acronyms in word helps you understand shorthand employed uh in your own work environment relative to current
Open document uh on Walmart.com when the user is looking at a product such as gaming
console recommends accessories and games that would go with it when a user searches for movies Google shows results
including showtimes near the users's location for the current data onto our fifth card here match based uh we didn't
we didn't miss this one right yeah we did okay so we're on the fifth one here match relevant social norms ensure
experience is delivered in a way the users would expect given the social cultural context when editor identifies
ways to improve writing style presents options politely consider using that's the Canadian way being polite uh Google
photos is able to recognize pets and use the wording important cats and dogs recognizing that for many pets are an
important part of one's family and you know what uh when I uh started renting my new house uh I I said you know is
there a problem with dogs and my landlord said well of course pets are part of the family and that was
something I like to hear uh Cortana uses semiformal tone apologizing when unable to find a uh contact which is polite and
socially appropriate I like that okay mitigate social biases ensure AI system languages and behaviors do not
reinforce undesirable unfair stereotypes and biases my anal summarizes how you spend your time at work then suggest
ways to work smarter one ways to mitigate bias is by using gender neutral icons to represent important people
sounds good to me a Bing search for CEO or doctor shows images of diverse people in terms
of gender and an ethnicity sounds good to me the predictive uh keyboard for Android suggests both genders when
typing a pronoun starting with the letter H we're on to our yellow cards uh so support efficient invocation so make
it easy to invoke or request system Services when needed so flashfill is a helpful timesaver in Excel that can be
easily invoked with on canvas interactions and uh that keep you in flow on amazon.com oh hey there got
Amazon in addition to the system giving recommendations as you browse you can manually invoke additional
recommendations from the recommender for your menu uh design ideas in Microsoft PowerPoint can be invoked uh with the
with the Press of a button if needed I cannot stand it when that pops up I always have to tell it to leave me alone
okay support efficient dismal uh efficient disle dismissal oh support efficient
dismissal okay make it easy to dismiss or ignore or undesired AI system Services okay this sounds good to me
Microsoft forms allows you to create custom surveys quizzes polls questionnaires and forms some choices
questions trigger suggested options position beneath the relevant question the suggestion can be easily ignored and
dismissed Instagram allows the user to easily hide or report ads that have been suggested by AI by tapping the ellipses
at the top of the right of the ad Siri can be easily dismissed uh uh by saying never mind
I'm always telling my Alexa never mind support efficient uh correction make it easy to edit refine or recover the AI
system uh when the when the AI system is wrong so alt Auto alt text automatically generates alt text for photographs by
using intelligent services in the cloud descriptions can be easily Modified by clicking the alt text button in the
ribbon once you set a reminder with Siri the UI displays a tap to edit link when Bing automatically corrects spelling
errors in search queries it provides the option to revert to the query as originally typed with one click on to
card number 10 Scope Services when in doubt so engage in
disambiguate disambiguation or gracefully degrade the AI system service when uncertain about a
user's goal so when Auto replacing word is uncertain of a correction it engages in disambiguation by displaying multiple
options you can select from Siri will let you know it has trouble hearing if you don't respond or talk or or speak
too softly big Maps will provide multiple routing options when uh when unable to recommend best one we're on to
card number 11 make clear why the system did what it did enable users to access an explanation of why the AI system
behaved as it did office online recommends docu documents based on history and activity descriptive text
above each document makes it clear why the recommendation is shown product recommendations on Amazon .c include why
Rec recommended link that shows that what products in the user shopping history informs the recommendations
Facebook enables you to access an explanation about why you are seeing each ad in the news
feed onto our green cards so remember recent interactions so maintain short-term memory and allow the user to
make efficient references to that memory when attaching a file Outlook offers a list of recent files including recently
copied file links Outlook also remembers people you have interacted with recently and displays uh them when addressing a
new email uh Bing search remembers some recent queries and search can be
continued uh conversationally how old is he after a search for kyanu Reeves Siri carries over the context from one
interaction to the next a text message is Created from the person you told Siri to message to onto card number 13 lucky
number 13 learn from user Behavior personalize the user experience by learning from their actions over time
tap on a search bar in office applications and search lists uh the top three commands on your screen that
you're most likely to need to personalize the technology called zero query doesn't even need to type in the
search bar to provide a personalized predictive answer amazon.com gives personalized product recomm
recommendations based on previous purchases onto card 14 update and adapt Cur uh C uh cautiously limit disruptive
changes when updating adaptive adapting the systems behaviors so PowerPoint designer improves slides for office 65
subscribers by automatically generating design ideas from to choose from designer has integrated new capabilities
such as smart Graphics icon suggestions and existing user experience ensuring the updates are not
disruptive office tell office tell me feature shows dynamically recommended items and a designated try area to
minimize disruptive changes onto card number 15 encourage granular feedback back enable the users to provide
feedback indicating their preferences during regular interactions with the AI system so ideas and Excel empowers you
to understand your data through high level visual summaries Trends and patterns encourages feedback on each
suggestion by asking is this helpful not only does Instagram provide the option to hide specific ads but it also
solicits feedback to understand why the ad is not relevant and Apple's music app love dislike buttons are prominent
easily accessible number 16 convey the consequences of us actions immediately
update or convey how user actions will impact future behaviors of the AI system you can get stock in G Geographic data
types and Excel it is easy as typing text into a cell and converting it to stock data type or geograph geography
data type when you perform the conversion action an icon immediately appears in the converted cells uh upon
tapping the like dislike button for each recommendation it at in apple music a pop-up informs the user that they'll
receive more or fewer similar recommendations onto card number 17 we're almost near the end provide Global
controls allow the user to globally customize the system system monitors and how it behaves so editor expands on
spelling and grammar checking capabilities of word to include more advanced proofing and editing designed
to ensure document is readable editor can flag a range of critique types and allow to customize the thing is is that
in word it's so awful spellchecking I don't understand like it's been years and the the spell checking never gets
better so they got to emplore better spellchecking AI I think bang search provides settings that impact the the
types of results the engine will return for example safe search uh then we have Google photos
allows user to turn location history on enough for future photos it's kind of funny seeing like Bing in there about
like using AI because at one point it's almost pretty certain that Bing was copying just Google search indexes to
learn how to index I don't know that's Microsoft for you uh we're on on to card 18 notify users about changes inform the
user when AI system adds or updates his capabilities uh the what's new dialogue in office informs you about changes by
giving an overview of the latest features and updates including updates to AI features in Outlook web the help
tab includes a what's new section that covers updates so there we go we made it to the end of the list I hope that was a
fun listen for you and and there I hope that we could kind of match up the uh the responsible AI I kind of wish what
they would have done is actually mapped it out here and said where it match but I guess it's kind of an isolate service
that kind of ties in so I guess there we go [Music]
okay hey this is Andrew Brown from exam Pro and we are looking at Azure cognitive services and this is a
comprehensive family of AI services and cognitive apis to help you build intelligent apps so create customizable
pre-trained models built with breakthrough AI researches I put that in quotations I'm kind of throwing some
shade at uh Microsoft at Azure just because it's their marketing material right uh deploy cognitive Services
anywhere from Cloud to the edge uh with containers get started quickly no machine learning expertise required but
I think it it helps to have a bit of background knowledge uh develop with strict ethical standards uh Microsoft
loves talking about the responsible um their responsible AI stuff empowering responsible use with industry leading
tools and guidelines so let's do a quick breakdown of the types of services in this family so for decision we have
anomaly detector identify potential problems early on content moderator detect potentially offensive or unwanted
content personalizer create Rich personalized experiences for every user for languages we have language
understanding also known as uh Luis Lewis I don't know why I didn't put the initialism there but don't worry we'll
see it again build natural language understanding into app spots and iot devices Q&A maker create a
conversational question and answer layer over your data text analytics detect sentiment so sentiment is is like
whether customers are happy sad glad uh keep phrases and named entities translator detect and translate more
than 90 supported languages for speech we have speech to text so transcribe audible um speech into readable search
text text to speech convert text to lifelike speech for natural interfaces speech translation so integrate realtime
speech translation into your apps uh speaker recognition uh identify and verify uh the People speaking based on
audio for vision uh we have computer vision so analyze content in images and videos custom Vision so analyze or sorry
customize image Rec image recognition to fit your business needs um pH detect uh and identify people and emotions in
images so there you [Music] go so Azure cognitive Services is an
umbrella AI service that enables customers to access multiple AI services with an AI key and API endpoint so what
you do is you go create a new cognitive service and once you're there it's going to generate out two keys and an endpoint
and that is what you're using generally for authentication uh with the various AI services programmatically and that is
something that is key to the service that you need to [Music]
know so knowledge mining is a discipline in AI that uses a combination of intelligent services to quickly learn
from vast amounts of information so it allows organizations to deeply understand and easily explore
information uncover hidden insights and find relationships and patterns at scale so we have ingest enrich and explore as
our three steps so for ingest content from a range of sources using connectors to First and thirdparty data stores so
we might have structured data such as databases csvs uh the csvs would more be semi-structured but we're not going to
get into that level of detail unstructured data so PDFs videos images and audio for enrich the content with AI
capabilities that let you extract information find patterns and deepen understanding in so cognitive services
like Vision language speech decision and search and explore the newly indexed data via search spots existing
businesses applications and data visualizations enrich uh structured Data customer relationship management wrap
systems powerbi this whole knowledge mining thing is a thing but like I believe that the whole model around this
is so that Azure uh shows you how you can use the cognitive services to solve uh things without having to invent new
Solutions so let's look at a bunch of use cases that Azure has uh and see what where we can find some in uh useful use
so the first one here is for Content research so when organizations uh task employees review and research of
technical data it can be tedious to read page after page of dense texts knowledge mining helps employees quickly review
these dense materials so you have a document and in the enrichment step you could be doing printed text recognition
key phrase extraction sharpener uh sharpen skills technical keyword sanitation format definition minor large
scale vocabulary matcher you put it through a search service and now you have search reference library so it
makes things a lot easier to work with uh now we have uh audit risk compliance management so developers could use
knowledge mining to help attorneys quickly identify entities of importance from Discovery documents and flag
important ideas across documents so we have documents so Clause extraction Clause uh classification gdpr risk named
uh identity extraction keyphrase extraction language detection automate translation uh then you put it back into
a search index and now you can use it a management platform or a word plugin and so we have business Process
Management in Industries where bidding competition is fierce or when the diagnosis of a problem must be quick or
in near real time companies use knowledge mining to avoid costly mistakes so the client Drilling and uh
uh and completion reports document processor AI services and custom models Q for human validation intelligent
automation yes send it to a backend system or a data Lake Andor a data Lake and then you do your analytics dashboard
then we have customer support and feedback uh analysis so for many companies customer support is costly and
in uh efficient knowledge mining can help customer support teams quickly find the right answers for a customer inquiry
or assess customer sentiment at scale so you have your Source data you do your document cracking use cognitive skills
so pre-trained services or custom uh you have enriched documents from here you're going to do your projections and have a
knowledge store you're going to have a search index and then do your analytics something like powerbi uh we have
digital assessment management I know there's a lot of these but it really helps you understand how cognitive
services are going to be useful uh given the amount of unstructured data created daily many companies are struggling to
make use of or find information within their files knowledge mining through a search index makes it easy for end
customers and employees to locate what they're looking for faster so you in just like art metadata and the actual
images themselves for the top ler we're doing geopoint extractor biographical enricher then down below we're tagging
we're custom object detector similar image tagger we put it in a search index they love those search indexes and now
you have an art Explorer uh we have contract management this is the last one here many companies
create products for multiple sectors hence the business opportunities with different vendors and buyers increase
exponentially knowledge mining can help organizations to scour thousands of pages of sources to create accurate bids
so here we have RFP documents um this will actually probably come back later in the original set but we will Ex we'll
do risk extraction print text recognition keyphrase extraction organizational extraction engineering
standards we'll create a search index and put it here this will bring back data also Metadate extraction will come
back here and then this is just like a continuous pipeline [Music]
okay hey this is Andrew Brown from exam Pro and we are looking at face service and Azure face service provides an AI
algorithm that can detect recognize and analyze human faces and images uh such as a face in an image face with specific
attributes face landmarks similar faces and the same face as a specific identity across a gallery of images so here is an
example of an image uh that I ran that will do in the follow along and uh what it's done is it's drawn a bounding box
around the image and there's this ID and this is a unique identifier uh string for each detected face in an image and
these can be unique across the gallery which is really useful as well another cool thing you can do is a face
landmarks so the idea is that you have a face and it can identify very particular components of it and up to 27 predefined
landmarks is what is provided with this face Service uh another interesting thing is face attribut so you can uh
check whether they're wearing access accessories so think like earrings or lip rings uh determine its age uh the
blurriness of the image uh what kind of emotion is being uh experienced the exposure of the image you know the
contrast uh facial hair gender glasses uh your hair in general the head pose there's a lot of information around that
makeup which seems to be limited like when we ran it here in the lab all we got back was eye makeup and lip makeup
but hey we get some information whether they're wearing a mask uh noise so whether there's artifacts like visual
artifacts or occlusion so whether an object is blocking parts of the face and then they simply have a Boolean value
for whether the person smileing or not which I assume is a very common component so that's pretty much all we
really need to know about the faith service and there you [Music]
go hey this is Andrew Brown from exam Pro and we are looking at the speech and translate service so a your's translate
service is a translation service as the name implies and it can translate 90 languages and dialects and I was even
surprised to find out that it can translate into kingon um and it uses a neural machine translation nmt replace
reping its Legacy statistical machine translation smt so what my guess here is that statistical meaning that it used
classical machine learning back in 2010 and and then they decided to switch it over to neural networks um which of
course would be a lot more accurate uh as your translate service can support a custom translator so it allows you to
extend the service for translation based on your business domain use cases so if you use a lot of technical words and
things like that then you can fine-tune that or particular phrases then there's the other service Azure speech service
and this is a uh a speech uh synthesis Serv service so what can do is speech to text text to speech and speech
translation so it's synthesizing creating new voices okay so we have speech to text so real time speech to
text batch uh batching Multi-Device conversation conversation transcription and you can create custom speech models
then you have text to speech so this utilizes a speech synthesis markup language so it's just a way of
formatting it and it can create cust some voices uh then you have the voice
assistant so integrates with the bot framework SDK and speech recognition so speaker verification and identification
so there you [Music] go hey this is Andrew Brown from exam
Pro and we are looking at text analytics and this is a service for NLP so natural language processing for text Mining and
text analysis so text analytics can perform sentiment analysis so find out what people think about your brand or
topics so features provide sentiment labels such as negative neutral positive then you have opinion minin which is an
aspect based sentiment analysis it's for granular information about the opinions related to aspects then you have key
phrase extraction so quickly identify the main Concepts in text you have language detections that detect the
language uh of an inputed text that it's written in and you have name entity recognition so Neer so identify and
categorize entities in your text as people places objects and quantities and subset of ner is personally identifiable
information so piis let's just look at a few of these more in detail some of them are very obvious but some of these would
help to have an example so the first we're looking at is key phrase extraction so quickly identify the main
Concepts and text so keyphrase extraction works best when you have uh when you give it bigger amounts of text
to work on this is the opposite of sentiment analysis which performs better on smaller amounts of text so document
sizes can be 5,000 or fewer characters per document and you can have up to a thousand items per collection so imagine
you have a movie review with a lot of text in here and you want to uh extract out the key phrases so here it
identified Borg ship Enterprise um surface travels things like that uh then you have named entity recognition so
this detects words and phrases mentioned in unstructured data that can be associated with one or more semantic
types and so here's an example I think this is medicine based and so the idea is that it's identifying it's
identifying um these words or phrases and then it's applying a semantic type so it's saying
like this is a diagnosis this is a medication class and stuff like that uh semantic type could be more broad so
there's location event but location twice your person diagnosis age and there is a predefined set I believe that
is in um Azure that you should expect but they have a generic one and then there's one that's for health we're
looking at sentiment analysis this graphic makes it uh make a lot more sense when we're splitting between
sentiment and opinion mining but the idea here is that sentiment analysis will apply labels and confidence scores
to text at the sentence and document level and so labels can include negative positive mixed or neutral and we'll have
a confidence score ranging from 0 to one and so over here we have a sentiment analysis of this line here and it's
saying that this was a negative sentiment but look there's something that's positive and there's something
that's negative so was it really negative and that's where opinion mining gets really useful because it has more
granular data where we have a subject and we have an opinion right and so here we can see the room was great but the
staff was unfriendly negative so we have a bit of a split there [Music]
okay hey this is Andrew Brown from exam Pro and we are looking at optical character recognition also known as OCR
and this is the process of extracting printed or handwritten text into a digital and editable format so OCR can
be applied to photos of street signs products documents invoices bills s Financial reports articles and more and
so here's an example of us extracting out nutritional data or nutritional facts off the back of a food product so
Azure has two different kinds of apis that can perform OCR they have the OCR API and the read API so the OCR API uses
an older recognition model it supports only images it executes synchros synchronously returning immediately when
uh it detects text it's suited for for Less texts it supports more languages it's easier to implement and on the
other side we have the read API so this is an updated recognition model supports images and PDFs executes
asynchronously paralyzes tasks per line for faster results suited for lots of texts supports fewer languages and it's
a bit more difficult to implement and so when we want to use this service we're going to be using uh computer vision SDK
okay hey this is Andrew Brown from exam Pro and we're taking a look here at form
recognizer service this is a specialized OCR service that translates printed text into digital and editable content it PR
preserves the structure and relationships of the form like data that's what makes it so special so form
recognizer is used to automate data entry in your applications and enrich your document search capabilities it can
identify key value pairs selection marks table structures uh it can produce output structures such as original file
relationships bounding boxes confidence score and form recognizer is composed of a custom document processing models
pre-built models for invoices receipts IDs business cards the model layouts let's talk about the layout here so
extract text selection marks table structures along with bounding box coordinates from documents for
recognizer can extract text selection marks and table structures the row and column numbers associated with the text
using highdefinition optical character enhancement models that is totally useless that
text hey this is Andrew Brown from exam Pro and we are looking at form recognizer
service and this is a Specialized Service uh for OCR that translates printed text into digital edible content
but the magic here is that preserves the structure and relationship of form likee data so there's an invoice you see those
magenta lines it's saying identify that form like data so form recogn is used to automate data entry in your applications
and enrich your document search capabilities and it can identify key value pairs selection marks table
structures and it can output structures such as original file relationships bounding box boxes confidence scores
it's composed of a customer Uh custom document processing model pre-built models for invoices receipts IDs and
business cards it's based on this layout model uh and there you [Music]
go so let's touch upon custom models so custom models allow you to extract text key value pair selection marks and
tabular data from your forms these models are trained with your data so they're tailored to your forms you only
need five samples uh sample input forms to start a trained document processing model can output structured data that
includes the relationship in the original form document after you train the model you can test and retrain it
and eventually use it reliably extract data from uh more forms according to your needs you have two learning options
you have unsupervised learnings to understand the layout and Rel ship between Fields entries and your forms
and you have supervised learning to extract values of Interest using the labeled form so we've covered
unsupervised and supervisor learning so you're going to be very familiar with these two
[Music] okay so Forum recognizer service has many pre-built models that are uh easy
to get uh started with and so let's go look at them and see what kind of fields it extracts out by default so the first
is receipts so sales receipts from Australia Canada great Britain India United States will work great here and
the fields it will extract out is receipt type Merchant name Merchant phone number Merchant address
transaction date transaction time total subtotal tax tip items name quantity price total price if there's information
that is on a a receipt that you're not getting out of these fields and that's where you make your own custom model
right for business cards it's only available for English business cards but uh we can extract out contact names
first name last name company names departments job titles emails websites addresses mobile phones faxes work
phones uh and other phone numbers not sure how many people are using uh business cards these days but hey they
have it as an option for invoices extract data from invoices in various formats and return structured data so we
have customer name customer ID purchase order invoice ID invoice date due date vendor name vendor address vendor
address receipt customer address customer address receipt billing address billing address receipt shipping address
subtotal total tax invoice total amount due service address uh remittance address uh start service start date and
end date uh previous unpaid balance and then they even have one for line items so items amount description quantity
unit price product code unit date tax and then for IDs which could be worldwide passports us driver license
things like that um you would have Fields such as country region date of birth date of expire expiration document
name first name last name nationality sex machine readable zone I'm not sure what that is document type and address
and region uh and there are some additional features with some of these models we didn't really cover them it's
not that important but um yeah there we [Music] go hey this is Andrew Brown from exam
Pro and we were looking at natural understanding or LS or Louise depends on how You' like to say it and this is a no
code ml service to build language uh natural language into Apps Bots and iot devices to quickly create Enterprise
ready custom models that continuously improve so Lewis I'm going to just call it lwis cuz that's what I prefer is
access via its own isolate domain at lewis. and it utilizes NLP and nlu so nlu is the ability to perform or ability
to transform a linguistic statement to a representation that enables you to understand your users naturally and it
is intended to focus on intention and extraction okay so where the users want or sorry what the users want and what
the users are talking about so uh the loose application is composed of a schema and the schema is autogenerated
for you when you use the Lewis AI web interface so you definitely AR going to be writing this by hand but it just
helps to see what's kind of in there if you do have some programmatic skills you obviously you can make better use of the
service than just the web interface but the schema defines intentions so what the users are asking for a l app always
contains a nonone intent we'll talk about why that is in a moment and entities what parts of the intent is
used to determine the answer then you also have utterances so examples of the user input that includes intent and
entities to train the ml model to match predictions against the real user input so an intent requires one or more
example utterance for training and it is recommended to have 15 to 30 example utterances to explicitly train uh to
ignore an utterance you use the nonone intent so intent classifies user utterances and entities extract data
from utterances so hopefully that understands I always get this stuff mixed up it always takes me a bit of
time to understand there is more than just these things there's like features and other things but you know for the AI
900 we don't need to go that deep okay uh just to get to visualizing this to make a bit easier so imagine we have
this uh this utterance here these would be the identities so we have two and Toronto this is the example utterance
and then the idea is that you'd have the intent and the intent and if you look at this keyword here this really helps
where it says classify it's that's what what it is it's a classification of this example utterance and that's how the ml
model is going to learn [Music] okay hey this is Andre Brown from exam
Pro and we are looking at Q&A maker service and this is a cloud-based NLP service that allows you to create a
natural conversational layer over your data so Q&A maker is hosted on its own isolate domain at Q&A maker. it will
help the most uh it will help you find the most appropriate answer from any input from your custom knowledge base of
information so you can commonly uh it's commonly used to build conversation clients which include social apps chat
Bots speech enabled uh desktop applications uh Q&A maker doesn't store customer data all the customer data
stored in the region the customer deploys the the dependent Services instances within okay so let's look at
some of the use cases for this so when you have static information you can use Q&A maker uh in your knowledge base of
answers this knowledge base is customed to your needs which you've built with documents such as PDF and URL
when you want to provide the same answer to a repeat question command when different users submit the same question
the answers is returned when you want to filter stack information based on meta information so metatag data is provide
uh provides additional filtering options relevant to your client application users and information common meded
information includes chitchat content type format content purpose content freshness and there's a use case when
you want to manage a bot conversation that includes static information so your knowledge base takes uh takes a user
conversation text or command and answers it if the answer is part of a predetermined conversation flow
represented in the knowledge base with multiple TurnKey context the bot can easily provide this flow so Q&A maker
Imports your content into a knowledge base of questions and answer Pairs and Q&A maker can build your knowledge base
from an existing document manual or website you're all docx PDF I thought this was the coolest thing so you can
just basically have anyone write a doc X as long as it has a heading and a and text and I think you can even extract at
images and it'll just turn it into uh the bot it just saves you so much time it's crazy it will use ml to extract the
question and answer pairs the content of the question of answer pairs include all the alternate forms of the question
metad dag tags used to filter choices during the search followup prompts to continue to search refinement uh
refinement K maker stores answers text in markdown once your knowledge base is imported you can finetune the imported
results by editing the question and answer pairs as seen here uh there is the chat box so you can converse with
your Bot through a chat box I wouldn't say it's particularly a feature of Q&A maker but I just want you to know that's
how You' interact with it so when you're using the Q&A maker AI the Azure bot service the bot composer um or via
channels you'll get an embeddable one you'll see this box where you can start typing in your questions and and get
back the answers to test it here an example is a multi-term conversation so somebody asked a question a generic
question and that said hey are you talking about adus or Azure which is kind of like a follow-up prompt and
we'll talk about multi-turn here in a second but uh that's something I want you to know about okay so chitchat is a
feature in Q&A maker that allows you to easily add prepopulated sets of top Chit Chats into your knowledge base the data
set has about 100 scenarios of chitchat in voices of multiple personas so the idea is like if someone says something
random like how are you doing what's the weather today things that your Bot wouldn't necessarily know it has like
canned answers and it's going to be different based on how you want the response to be okay uh there's a concept
of layered ranking so uh Q&A maker system is a layered ranking approach the data is stored in Azure search which
also serves as the first ranking layer the top result for uh from Azure search are then passed through Q&A makers NLP
reranking model to produce the final results and confidence score T touching on multi- turn conversation is a
follow-up prompt and context to manage the multiple turns known as multi-t for your Bot from one question to another
when a question can't be answered in a single turn that is when you're using multi-term conversation so Q&A maker
provides multi-term prompts and active learning to help you improve your questions based on key and answer Pairs
and gives you the opportunity to connect questions and answer pairs the connection allows the client application
to provide a top answer and provide more questions to refine the search for a final answer after the knowledge base
receives questions from users at the publish endpoint cre maker applies active learning to these rule work
questions to suggest changes to your knowledge base to improve the quality all
right [Music] hey this is Andrew Brown from exam Pro
and we are looking Azure bot service so the Azure bot service is an intelligent servess bot service that scales on
demand used for creating publishing and managing bots so you can register and publish a variety of bots from the Azure
portal so here there's a bunch of ones I've never heard of um probably with thirdparty providers partnered with
Azure and then there's the ones that we would know like the Azure healthbot the Azure bot or the uh web app bot which is
more of a generic one so Azure Bop service Bop bot service can integrate your Bot with other Azure Microsoft or
thirdparty service uh Services via channel so you can have a direct line Alexa Office 365 Facebook keik line
Microsoft teams Skype uh twio and more all right and two things that are commonly associated with the Azure bot
service is the bot framework and Bot composer in fact it was really hard just to make make this slide here because
they just weren't very descriptive on it cuz they wanted to push these other two things here but let's talk about the bot
framework SDK so the bot framework SDK which is now version four is an open source SDK that enables developers to
model and build sophisticated conversations the bot framework uh along with the Azure bot service provides an
end-to-end workflow so we can design build test publish connect and evaluate our uh Bots okay with this with this
framework developers can create Bots that use speech understand uh natural language handle questions answers and
more the bot framework includes a module accessible SDK for building Bots as well as tools templates and related AI
Services then you have bot framework composer and uh this is built on top of the bot framework SDK it's an open
source IDE for developers to author test provision and manage conversational experiences you can download it's an app
on Windows OS X and Linux is probably built uh using uh like uh web technology and so here is the actual uh app there
and so you can see there's kind of a bit of a flow and things you can do in there so you can either use C or not to build
your Bot you can deploy the bot to the Azure web apps or Azure functions you have templates to build Q&A maker bot
Enterprise or personal assistant bot language bot calendar or people bot uh you can test and debug via the bot
framework emulator uh and it has a built-in package manager there's a lot more to these things but again at the AI
900 this is all we need to know um but yeah there you [Music]
go hey this is Andrew Brown from exam Pro and we are looking at Azure machine learning service I want you to know
there's a classic version of the service it's still accessible in the portal this is not an exam we are going to 100%
avoid it uh it has severe limitations we cannot transfer anything over from the classic to the new one uh so the one
we're going to focus on is the Azure machine learning service you do create Studios within it so you'll hear me say
Azure machine Learning Studio and I'm referring to the new one a service that simplifies running AI ml work related
workloads allowing you to build flexible automated ml pipelines use python or Run Deep learning workloads such as
tensorflow we can make Jer notebooks in here so build and document your machine learning models as you build them share
and collaborate Azure machine learning SDK for python so an SDK designed specifically to interact with the Azure
machine learning Services it does ML Ops machine learning operations so end to end automation of ml model pipelines
cicd training inference aure machine learning designer uh uh so this is a drag and drop interface to visually
build test deploy machine learning models uh technically pipelines I guess as a data labeling service so assemble a
team of humans to label your training data responsible machine learning so model fairness through disparity metrics
and mitigate unfairness at the time of this Ser this is not very good but it's supposed to tie in with the responsible
AI that Microsoft is always promoting okay so once we launch our own uh Studio within Azure machine learning service
you're going to get this nice big bar uh or navigation left hand side it shows you there's a lot of stuff that's in
here so let's just break it down and what all these things are so for authoring we got notebooks these are
jupyter notebooks and IDE to write python code to build ml models they kind of have their own preview which I don't
really like but there is a way to bridge it over to jupyter notebooks or into Visual Studio code we have automl
completely automated process to build and train ml models it's you're limited to only three types of models but still
that's great we have the designer so visual drag and drop designer to construct end to-end ml pipelines for
assets we have data sets so data you can upload which we will be used which will be used for for training experiments
when you run a training job they are detailed here uh pipelines ml workflows that you have built or have used in the
designer model so a model registry containing train models that can be deployed endpoints so when you deploy a
model it's hosted on accessible endpoint so you're going to be able to uh access it via a rest API or maybe the SDK uh
for manage we got compute the underlying Computing instances used uh for notebooks training and inference
environments are reproducible uh python environment for machine learning experiments data stores
a data repository where your data resides data labeling uh so you have a human with ML assisted labeling to label
your data for supervised Learning Link services external service you can connect to the workspace such as Azure
synapse [Music] analytic let's take a look at uh the
types of compute that is available in our Azure machine Learning Studio we got four categories uh we have compute
instances to uh development workstations that data scientists can use to work with data and models compute clusters to
scalable clusters of VMS for on demand processing experimentation code deployment targets for Predictive
Services that use your trained models and attach compute links to existing Azure compute resources such as uh Azure
VMS uh and Azure data brick clusters now uh what's interesting here is like with this compute you can see that you can
open it in Jupiter Labs jupyter vs code R studio and terminal but you can you can work with uh your computer instances
your development workstations uh directly in the studio which that's the way I do it um what's interesting is for
inference that's when you want to make a prediction you use Azure kubernetes service or Azure container instance and
I didn't see it show up under here so I'm kind of confused whether that's where it appears U maybe we'll discover
as we do the follow alongs that they do appear here but uh I'm not sure about that one but yeah those are the four
there okay So within Azure machine Learning Studio we can do some data labeling so
we create data labeling jobs to prepare your ground Truth for supervised learning we have two options human in
the loop labeling you have a team of humans that will apply labeling these are humans you grant access to labeling
uh machine learning assisted daily labeling you will use ml to perform uh labeling so you can export the label
data from for machine learning experimentation at any time uh your users often export multiple times and
train different models rather than wait for all the images to be labeled images labeled can be exported in Coco format
that's why we talked about Coco uh a lot earlier in our data set section as your machine learning data set and this is
the data set format that makes it easy to use for training and Azure machine learning so generally you want to use
that format the idea is you would choose a labeling task type uh and that way you would have this UI and then people would
go in and and just click buttons and do the labeling okay so aure ml data store securely
connects you to storage service services on Azure without putting your authentication credentials and the
Integrity of your original data source at risk so here is the example of data source that are available to us in the
studio and let's just go quickly through them so we have Azure blob storage this is data that is stored as objects
distributed across many machines Azure file share am mountable file share via SMB and NFS protocols Azure data Lake
storage Gen 2 um this is blob storage design for vast amounts of big data analytics Azure SQL this is a fully
managed mssql relational database azure postest datas this is an open source relational database often considered an
object related database preferred by developers a your MySQL another open source relational database the most
popular one and considered a pure relational database okay so aure ml data sets makes it easy
to register your data sets for use with your ml workload so what you do is you'd add a data set and you get a bunch of
metadata associated with it uh and you can also upload uh addition like the data set again to have multiple versions
so you'll have a current version and a latest version uh it's very easy to get started working with them because
they'll have some sample code that's for the Azure ml SDK uh to import that into uh into your Jupiter notebooks uh for
data sets you can generate profiles that will give you summary statistics distribution of data and more you will
have to use a compute instance to generate that data so you press the generate profile and you'd have that
stored I think it's in Blob storage there are open data sets these are publicly hosted data sets that are
commonly used for learning how to build ml models so if you go to open data sets you just choose one and so this is a
curated list of open data sets that you can quickly add to your data store great for learning how to use autom ml or
aszure machine learning designer or any kind of uh ml uh workload if you're new to it that's why we covered mnist and
Coco earlier just because those are some common data sets there but there you [Music]
go taking a look here at Azure ml experiments this is a logical grouping of azure runs and runs act uh is the act
of running an ml task on a virtual machine or container so here's a list of them and uh it can run various uh types
of ml tasks so scripts could be pre-processing autom ml uh a training pipeline but what it's not going to
include is inference and what I mean is once you've deployed your model or Pipeline and you uh uh make predictions
via a request it's just not going to show up under here [Music]
okay okay so we have azure ml pipelines which is an exec workflow of a complete machine learning task not to be confused
with Azure pipelines which is part of azure devops or data Factory which has its own pipelines it's a total a totally
separate thing here so subas are encapsulated is a series of steps within the pipeline independent steps allow
multiple data scientists to work on the same pipeline at the same time without over taxing compute resources separate
steps also make it easy to use different compute type sizes for each step when you rerun a pipeline the Run jumps to
the steps that need to be rerun such as the updated training script steps do not need to be rerun and they will be
skipped after a uh pipeline has been published you can configure a rest endpoint which allows you to rerun the
pipeline from any platform or stack there's two ways uh to build pipelines you can use the Azure ml designer or
programmatically using Azure machine learning python SDK so here is example of some code just make a note here I
mean it's not that important but no it's just you create steps okay and then you assemble all the steps into a pipeline
line here all [Music] right so as your machine learning
designer lets you quickly build as your ml pipelines without having to write any code so here is what it looks like and
over there you can see our pipeline is quite Visual and on the left hand side you have a bunch of assets you can drag
out that are pre-built there so uh it's a really fast way for building a pipeline so you do have to have a good
understanding of um ml pipelines nend to end to make good use of it uh once you you trained your pipeline you can create
an inference pipeline so you drop down and you say whether you want it to be real or batch and you can TW toggle
between them later so I mean there's a lot to this service but for the AI 900 we don't have to go uh diving too deep
okay so Azure ml models or the model registry allows you to create manage and track your registered models as
incremental versions under the same name so each time you register a model with the same name as an existing one the
registry assures that it's a new version Additionally you can provide metadata tags and use tags when you search for
models so yeah it's just really easy way to share and deploy or download your models
[Music] okay Azure MLM points allow you to deploy machine learning models as a web
service so the workflow for deploying model is register the model prepare an entry script prepare an inference
configuration deploy the model locally to ensure everything works compute uh choose a compute Target redeploy the
model uh to the cloud test the resulting web service so we have uh two options here real time endpoints so an endpoint
that provides remote access to invoke the ml model uh service running on either Azure kubernetes service AKs or
Azure container instance ACI then we have pipeline endpoints so endpoint that provides remote access to invoke an ml
pipeline you can parameterize the pipeline endpoint for manage repeatability in batch scoring and
retraining scenarios um and so you can deploy a model to an endpoint it will either be
deployed to a AKs or ACI as we said earlier uh and the thing is is that when you do do that just understand that
that's going to be shown under the AKs or ACI um within the Azure portal it's not Consolidated under the Azure machine
Learning Studio when you've deployed a realtime endpoint you can test the endpoint by sending either a single
request or batch request so they have a nice form here where it's single or it's um like here it's a CSV that you can
send so there you go so Azure has a built-in jupyter like
notebook editor so you can build and train your ml models and so here is an example of it I personally don't like it
too much but that's okay because we have some other options to make it easier but what you do is you choose your compute
uh instance to run the notebook you'll choose your kernel uh which is a pre-loaded programming language and
programming libraries for different use cases but that's a Jupiter kernel uh concept there uh so you can open the
notebook in a more familiar ID such as vs code jupyter notebook classic or Jupiter lab so you go there drop it down
choose it and open it up and now you're in a more familiar territory the vs code one is exactly the same experience as
the um the one in Azure or Azure ml Studio I personally don't like it I think most people are going to be using
the notebooks but it's great that they have all those [Music]
options so Azure automated machine learning also known as autom ml automates the process of creating an ml
model so with Azure automl you supply a data set choose a task type uh and then automl will train and tune your model so
here are task types let's quickly go through them so we have classification when you need to make a prediction based
on several classes so binary classification multiclass classification regression when you need to predict a
continuous number value and then a Time series forecasting when you need to predict the value based on time so just
look at them a little bit more in detail so classification is a type of supervised learning in which the model
learns using training data and appli those learnings to new data so here is an example uh or this is just the option
here and so the goal of classification is to predict which categories new data will fall into based on learning from
its training data so binary classification is a record uh is labeled out of two possible labels so maybe it's
true or false zero or one it's just two values multiclass classification is a record is labeled out of range of out of
a range of labels uh and so it could be like happy sad mad or rad and just you know I can
see there's a smelling mistake there but yeah there should be an F so let's just correct that there we go uh you can also
apply deep learning and so if you turn deep learning on you probably want to uh use a gpus uh compute instance just
because or um compute cluster Because deep learning really prefers uh gpus okay looking at regression it's also a
type of supervised learning where the model learns uh using training data and applies those learnings to new data but
it's a bit different where the goal of regression is to predict a variable in the future uh then you have time series
forecasting and this sounds a lot like um uh regression because it is so forecast Revenue inventory sales or
customer demand an automated time series experiment that is treated as a multivariant regression problem past
time series values are pivoted to become additional dimensions for the regressor together with other predictors and
unlike classical time series methods has an advantage of naturally incorporating multiple contextual variables and their
relationship to one another during training so use cases here or Advanced configurations I should say holiday
detection and featurization time series uh deep learning uh neural networks so you got Auto arima profit forecast TCN
uh many model supports through grouping rolling origin cross validation configurable Labs rolling window
aggregate features so there you [Music] go so within automl we have data guard
rails and these are run by automl when automatic featurization is enabled it's a sequence of checks to ensure
highquality input data is being used to train the model so just to show you some information here so the idea is it could
apply validation split handling so the input data has been split for validation to improve the performance then you have
missing feature value uh imputation so no features missing values were detected in training data High cardinality
feature detection your inputs were analyzed and no no high cardinality features were detected High cardinality
means like if you have too many dimensions it becomes very dense or hard to process the data um so that's
something good to check [Music] against let's talk about autom ml's
automatic featurization so during model training with autom ML one of the following scaling or normalization
techniques will be applied to each model the first is standard scale wrapper so standardized features by removing the
mean and scaling to unit variance minmax scaler transform features by scaling each feature by the column is minimum
maximum Max ABS scaler scale each feature by its maximum absolute value robust scar scales features by the
quanti quantile range PCA linear dimensionality reduction using single uh value decomposition of the data to
project it to lower dimensional space uh uh Dimension uh reduction is very useful if your data is too complex if let's say
you have data and you have too many labels like 20 20 30 40 labels for per like for categories to pick out of you
want to reduce the dimensions so that your machine learning model is not overwhelmed so then you have truncated
SVD wrappers so the Transformer performs linear dimensionality reduction by means of truncated single singular value
decomposition contrary to PCA the estimator does not Center the data uh before Computing the singular value
decomposition which means it can work with spicy sparse matrices efficiently sparse normalization each sample that is
each row of the data Matrix which uh with at least one zero component is rescaled independently of other samples
so that is Norm so one L or two L2 I can't remember it's I2 or L anyway i1 and and I2 okay so the thing
is is that on the exam they're probably not going to be asking these questions but I just like to get you exposure but
I just want to show you that automl is doing all this this is like pre-processing stuff you know like this
is stuff that you'd have to do and so it's just taking care of the stuff for you
[Music] okay so within Azure automl they have a feature called Model selection and this
is the task of selecting a statistical model from a set of candidate models and Azure autom ml will use different uh or
many different ml algorithms that will recommend the best performing candidate so here is a list and I want to just
point out down below there's three pages there's 53 models it's a lot of models and so you can see that the it chose as
its top candidate was called voting Ensemble that's an ensemble uh um algorithm that's where you take two weak
ml models combine them together to make a more uh uh Stronger one and notice here it will show us the results and
this is what we're looking for which is the primary metric the highest value should indicate that that's the model we
should want to use you can get an explanation of the model called uh that's known as
explainability and now if you're a data scientist you might be a bit smarter and say well I know this one should be
better so I'll use this and tweak it but you know if you don't know what you're doing you just go with the top one
[Music] okay so we just saw that we had a top candidate model and there could be an
explanation to understand as to the effectiveness of this this is called MXL so machine learning explainability this
is the process of explaining interpreting ml or deep learning models MX mlx can help machine learning
developers to better understand interpret models Behavior so after your top candidate model is selected by Azure
automl you can get an explanation of internals of various factors so model performance uh data set Explorer
aggregate feature importance individual feature importance so I mean yeah this is aggregate so what it's looking at and
it's actually cut off here but it's saying that these are the most important ones that affect how the models outcome
so I think this is the diabetes dat a data set so BMI uh would be one that would be a huge influencer there okay
[Music] so the primary metric is a parameter that determines the metric to be used
during the model training for optimization so for classification we have a few and regression and time
series we have a few but you'll have these task types and underneath you'll choose the additional configuration and
that's where you can override the primary metric uh it might just Auto detected for you so you don't have to
because it might sample some of your data set to just kind of guess U but you might have to override it yourself uh
just going through some scenarios um and we'll break it down into two categories so here we have suited for larger data
sets that are well balanced well balanced means that your data set like is evenly distributed so if you have uh
uh CL classifications for A and B let's say you have 100 and 100 they're well balanced right you don't have one data
set much a subset of your dat set much larger than the other that's labeled so for accuracy this is great for image
classification sentiment analysis CH prediction for average Precision score weighted it's for sentiment analysis
Norm macro recall term prediction for precision score weighted uh uncertain as to what that would be good for maybe
sentiment analysis suited for smaller data sets that are inbalance so that's where your data set like you might have
like 10 records for one and 500 for the other on the label so you have Au weighted fraud detection image
classification anomaly detection spam detection onto regression scenarios uh will break it down into ranges so when
you have a very wide range uh Spearman correlation works really well R2 score this is great for airline uh delay
salary estimation bug res resolution time we're looking at smaller ranges where you're talking about normalize
root square mean to error so price predictions um review tips score predictions for normalized mean absolute
error um it's going to be just another one here they don't give a description for time series it's the same thing it's
just in the context of Time series so forecasting all [Music]
right another option we can change is the validation type when we're setting up our ml model so Val validation model
validation is when we compare the results of our training data set to our test data set model validation occurs
after we train the model and so you can just drop it down there we have some options so Auto kfold cross validation
Monte Carlo cross validation train validation split I'm not going to really get into the details of that I don't
think it'll show up on the AI 900 exam but I just want you to be aware that you do have those options
[Music] okay hey this is Andrew Brown from exam Pro and we are taking a look here at
custom vision and this is a fully managed no code service to quickly build your own classification and object
detection ml models the service is hosted on its own isolate domain at www.com vision. so the first idea is you
upload your images so bring your own labeled images or custom Vision to quickly add tags to any unlabeled data
images you use the labeled images to teach custom Vision the concepts you care about which is training and you use
a simple rest API that calls uh to quickly tag images with your new custom computer vision model so you can
evaluate [Music] okay so when we launch custom Vision we
have to create a project and with that we need to choose a project type and we have classification and object detection
reviewing classification here you have the option between multi-label so when you want to apply many tags to an image
so think of an image that contains both a cat and a dog you have multi class so when you only have one possible tag to
apply to an image so it's either an apple banana and orange it's not multiples of these things you have
object detection this is when we want to detect various objects in an image uh and you also need to choose a domain a
domain is a Microsoft managed data set that is used for training the ml model there are different domains that are
suited for different use cases so let's go take a look first at image classification domains so here is the
big list the domains being over here okay and we'll go through these here so General is optimized for a broad
range of image classification tasks if none of the uh uh if none of the other specified domains are appropriate or
you're unsure of which domain to choose select one of the general domains so G uh or A1 is optimized for better
accuracy with comparable inference time as general domain recommended for larger data sets or more difficult user
scenarios this domain requires a more training time then you have A2 optimized for better accuracy with faster advert
times than A1 and general domains recommended for more most data sets this domain requires less training time than
General and A1 you have food optimized for photographs or dishes of as you would see them on a restaurant menu if
you want to classify photographs of individual fruits or vegetables use food domains uh so then we have optimize for
recognizable landmarks both natural and artificial this domain works best when Landmark is clearly visible in the
photograph this domain works even if the land mark is slightly um obstructed by people in front of
it then you have retail so optimize for images that are found in a shopping cart or shopping uh website if you want a
high Precision classifying uh classifying between dresses pants shirts use this domain compact domains
optimized for the constraints of real-time classification on the edge okay then uh we have object detection
domain so this one's a lot shorter so we'll get through a lot quicker so optimize for a broad range of object
detection tasks if none of the uh other domains are appropriate or you're unsure of which domain choose the general one
A1 optimized for better accuracy and comparable inference time than the general domain recommended for most
accurate region locations larger data sets or more difficult use case scenarios the domain requires more
training and results are not deterministic expect uh plus minus 1% mean average Precision difference uh
with the same training data provided you have logo optimized for finding Brands uh logos and images uh products on
shelves so optimized for detecting and classifying products on the shelves so there you
[Music] go okay so let's get some uh more practical knowledge of the service so
for image classification you're going to upload multiple images and apply a single or multiple labels to the entire
image so here I have a bunch of images uploaded and then I have my tags over here and they could either be multi or
singular for object detection you apply tags to objects in an image for data labeling and you hover uh your cursor
over the image custom Vision uses ml to show boundaries uh bounding boxes of possible objects have not yet been
labeled if it does not detect it you can also just click and drag to draw out whatever Square you want so here's one
where I tagged it up quite a bit you have to have at least 50 images on every tag to train uh so just be aware of that
when you are tagging your images uh when you're training your model is ready when you and you have two options so you have
quick training this trains quickly but it will be less accurate you have Advanced Training this increases compute
time to improve your results so for Advanced Training BAS basically you just have this thing that you move to the
right uh with each iteration of training our ml model will improve the evaluation metric so precision and recall it's
going to vary we're going to talk about the metrics here in a moment but the probability threshold value determines
when to stop training when our evaluation metric meets our desired threshold so these are just additional
options where when you're training you can move this left to right uh and these left to right
okay and then when we get our results back uh we're going to get um some metrics here so uh we have evaluation
metric so we have Precision being exact and accurate selects items that are relevant recall such sensitivity or
known as true positive rate how many relevant items returned average Precision it's important that you
remember these because they might ask you that on the exam so for uh cut when we're looking at object detection and
we're looking at the evaluation metric outcomes for this one we have Precision recall and mean average Precision uh
once we've deployed our pipeline it makes sense that we go ahead and give it a quick test to make sure it's working
correctly so you press the click quick test button and you can upload your image and it will tell you so this one
says it's Warf uh when you're ready to publish you just hit the publish button and then you'll get uh some prediction
URL and information so you can invoke it uh one other feature that's kind of useful is the Smart labeler so once
you've loaded some training data within it can now make suggestions right so you can't do this right away but once it has
some data it's like it's like kind of a prediction that is not 100% guaranteed right and it just helps you build up
your training uh data set a lot faster uh very useful if you have a very large data set this is known as ml assisted
labeling [Music] okay hey this is Andrew Brown from exam
Pro and in this section we'll be covering the newly added section to the AI 900 that focuses on generative AI
generative AI including Technologies like chat GPT is becoming more recognized outside of tech circles while
it may seem magical in its ability to produce humanlike content it's actually based on Advanced mathematical
techniques from statistics data science and machine learning understanding these Core Concepts can help Society Envision
new AI possibilities for the future first let's compare the differences between regular AI versus generative ai
ai refers to the development of computer systems that can perform tasks typically requiring human intelligence these
include problem solving decision-making understanding natural language recognizing speech and images and more
the primary goal of traditional AI is to create systems that can interpret analyze and respond to human actions or
environmental changes efficiently and accurately it aims to replicate or simulate human intelligence in machines
AI applications are vast and include areas like expert systems natural language processing speech recognition
and Robotics AI is used in various Industries for tasks such as customer service chatbots recommendation systems
in e-commerce autonomous vehicles and medical diagnosis on the other hand generative AI is a subset of AI that
focuses on creating new content or data that is novel and realistic it does not just interpret or analyze data but
generates new data itself it includes generating text images music speech and other forms of media it often involves
Advanced machine learning techniques particularly deep learning models like generative adversarial networks
variational autoencoders and Transformer models like GPT generative AI is used in a range of applications including
creating realistic images and videos generating human-like text composing music creating virtual environments and
even drug Discovery some examples include tools like GPT for text generation doly for image creation and
various deep learning models that compose music so let's quickly summarize the differences of regular AI with
generative AI across three features functionality data handling and applications regular AI focuses on
understanding and decision making whereas generative AI is about creating new original outputs in terms of data
handling regular AI analyzes and bases decisions on existing data while generative AI uses the same data to
generate new previously unseen outputs and for applications regular I scope includes data analysis automation
natural language processing and Healthcare in contrast generative AI leans towards more creative and
Innovative applications such as content creation synthetic data generation deep fakes and
Design the next topic we'll be covering is what is a large language Model A large
language model such as GPT Works in a way that's similar to a complex automatic system that recognizes
patterns and makes predictions training on large data sets initially the model is trained on massive amounts of text
Data this data can include books articles websites and other written material during this training phase the
model learns patterns and language such as grammar word usage sentence structure and even style and tone understanding
context the model's design allows it to consider a wide context this means it doesn't just focus on single words but
understands them in relation to the words and sentences that come before and after this context understanding is
important for generating coherent and relevant text predicting the next word when you give the model a prompt which
is a starting piece of text it uses what it has learned to predict the next most likely word it then adds this word to
the prompt and repeats the process continually predicting the next word based on the extended sequence
generating text this process of predicting the next word continues creating a chain of words that forms a
coherent piece of text the length of this generated text can vary based on Specific Instructions or limitations set
for the model refinement with feedback the model can be further refined and improved over time with feedback this
means it gets better at understanding and generating text as it is exposed to more data and usage in summary a large
language model works by learning from a vast quantity of text Data understanding the context of language and using this
understanding to predict and generate new text that is coherent and contextually appropriate which can be
further refined with feedback as shown in the workflow [Music]
image next let's talk about Transformer models so a Transformer model is a type of machine learning model that's
especially good at understanding and generating language it's built using a structure called the Transformer
architecture which is really effective for tasks involving natural language processing like translating languages or
writing text trans Transformer model architecture consists of two components or blocks first we have the encoder this
part reads and understands the input text it's like a smart system that goes through everything it's been taught
which is a lot of text and picks up on the meanings of words and how they're used in different contexts then we have
the decoder So based on what the encoder has learned this part generates New pieces of text it's like a skilled
writer that can make up sentences that flow well and make sense there are different types of Transformer models
with specific jobs for example Bert is good at understanding the language it's like a librarian who knows where every
book is and what's inside them Google uses it to help its search engine understand what you're looking for GPT
is good at creating text it's like a skilled author who can write stories articles or conversations based on what
it has learned so that's an overview of a transformer model next we'll be talking about the main components of a
transformer model the next component of a transformer model we'll be covering is
the the tokenization process tokenization in a Transformer model is like turning a sentence into a puzzle
for example you have the sentence I heard a dog bark loudly at a cat to help a computer understand it we chop up the
sentence into pieces called tokens each piece can be a word or even a part of a word so for our sentence we give each
word a number like this I might be one her might be two a might be three do might be four bark might be five loudly
might be six at might be seven is already token is three tap might be eight now our sentence becomes a series
of numbers this is like giving each word a special code the computer uses these codes to learn about the words and how
they fit together if a word repeats like a we use its code again instead of making a new one as the computer reads
more text it keeps turning new words into new tokens with new numbers if it learns the word meow it might call it
nine and skateboard could be 10 by doing this with lots and lots of text the computer builds a big list of these
tokens which it then uses to stand and generate language it's a bit like creating a dictionary where every word
has a unique number the next component of a transformer model will be covering our
embeddings so to help a computer understand language we turn words into tokens and then give each token a
special numeric code called an embedding these embeddings are like a secret code that captures the meaning of the word as
a simple example suppose the embeddings for our tokens consist of vectors with three elements for example four for dog
has the embedding vectors 10 32 five for bark has the vectors 10 22 eight for cat the vectors are 10 3 1 nine for meow the
vectors are 10 2 1 and 10 for skateboard as the vectors 3 3 one which is quite different from the rest words that have
similar meanings or are used in similar ways get codes that look alike so dog and bark might have similar codes
because they are related but skateboard might be off in a different area because it's not much related to these other
words words this way the computer can figure out which words are similar to each other just by looking at their
codes it's like giving each word a home on a map and words that are neighbors on this map have related meanings the image
shows a simple example model in which each embedding has only three dimensions real language models have many more
Dimensions tools such as word TWC or the encoding part of a transformer model help AI to figure out where each word
dot should go on this big map let's go over positional encoding from a Transformer model positional
encoding is a technique used to ensure that a language model such as GPT doesn't lose the order of words when
processing natural language this is important because the order in which words appear can change the meaning of a
sentence let's take the sentence I heard a dog bark loudly at a c from our previous example without positional
encoding if we simply tokenize this sentence and convert the tokens into embedding vectors we might end up with a
set of vectors that lose the sequence information positional encoding adds a positional Vector to each word in order
to keep track of the positions of the words however by adding positional en coding vectors to each words embedding
we ensure that each position in the sentence is uniquely identified the embedding for I would be modified by
adding a positional Vector corresponding to position one labeled I 1 the embedding for herd would be altered by a
vector for position two labeled herd 2 the embedding for a would be updated with a vector for position three labeled
a 3 and reused with the same positional vector for its second occurrence this process continues for each word token in
the sentence with dog four bark five loudly six at seven and Cat 8 all receiving their unique positional
encodings as a result the sentence I heard a dog bark loudly at a cat is represented not just by a sequence of
vectors for its words but by a sequence of vectors that are influenced by the position of each word in the sentence
this means that even if another sentence had the same words in a different order its overall representation would be
different because the positional encodings differ reflecting the different sequence of words so that's an
overview of positional encoding the next component of a transformer we'll be covering is
attention attention in AI especially in Transformer models is a way the model figures out how important each word or
token is to the meaning of a sentence particularly in relation to the other words around it let's reuse the sentence
I heard a DOT bark loudly at a cat to explain this better self attention imagine each word word in the sentence
shining a flashlight on the other words the brightness of the light shows how much one word should pay attention to
the others when understanding the sentence for bark the light might shine brightest on dot because they're closely
related encoder's role in the encoder part of a transformer model attention helps decide how to represent each word
as a number or vector it's not just the word itself but also its context that matters for example bark in the bark of
a tree would have a different representation than bark and I heard a DOT bark because the surrounding words
are different decoder's role when generating new text like completing a sentence the decoder uses attention to
figure out which words it already has are most important for deciding what comes next if our sentence is I heard a
dog the model uses attention to know that her and dog are key to adding the next word which might be bark multi-head
attention this is like having multiple flashlights each highlighting different aspects of the words maybe one
flashlight looks at the meaning of the word another looks at its role in the sentence like subject or object and so
on this helps the model get a richer understanding of the text building the output the decoder builds the sentence
one word at a time using attention at each step it looks at the sentence so far decides what's important and then
predicts the next word it's an ongoing process with each new word influencing the next so attention in Transformer
models is like a guide that helps the AI understand and create Language by focusing on the most relevant parts of
the text considering both individual word meanings and their relationships within the
sentence let's take a look at the attention process token embeddings each word in the sentence is represented as a
vector of numbers or its embedding predicting the next token the goal is to figure out what the next word should be
also represented as a vector as signing weights the attention layer looks at the sentence so far and decides how much
influence each word should have on the next one calculating attention scores using these weights a new Vector for the
next token is calculated which includes an attention score multi-head attention does this several times focusing on
different aspects of the words choosing the most likely word a neural network takes these vectors with attention
scores and picks the word from the vocabulary that most likely comes next adding to the sequence The Chosen word
is added to the existing sequence and the process repeats for each new word so let's use gp4 is an example for how this
entire process works explained in a simplified manner a Transformer model like gp4 works by taking a text input
and producing a well structured output during training it learns from a vast array of text Data understand
understanding how words are typically arranged in sentences the model knows the correct sequence of words but hides
future words to learn how to predict them when it tries to predict a word it Compares its guess to the actual word
gradually adjusting to reduce errors in practice the model uses its training to aside importance to each word in a
sequence helping it guess the next word accurately the result is that gp4 can create sentences that sound like they
were written by a human however this doesn't mean the model knows things or is intelligent in the human sense it's
it's simply very good at using its large vocabulary and training to generate realistic text base on word
relationships so that's an overview of attention in a Transformer [Music]
model hey this is Andrew Brown from exam Pro and in this section we'll be going over an introduction to Azure openai
service Azure open AI service is a cloud-based platform designed to deploy and manage Advanced language models from
open AI this service combines open I's latest language model development with the robust security and scalability of
azure's cloud infrastructure Azure open AI offers several types of models for different purposes gp4 models these are
the newest in the line of GPT models and can create text and programming code when given a prompt written in natural
language GPT 3.5 models similar to GPT 4 these models also create text and code from natural language props the GPT 3.5
turbo version is specially designed for conversations making it a great choice for chat applications and other
interactive AI tasks embedding models these models turn written text into number sequences which is helpful for
analyzing and comparing different pieces of text to find out how similar they are doll e models these models can make
images from descriptions given in words the doll e models are still being tested and are shown in the Azure Open aai
studio so you don't have to set them up for use manually key Concepts in using Azure open AI include prompts and
completions tokens resources deployments prompt engineering and various models prompts and completions users interact
with the API by providing a text command in English known as a prompt and the model generates a text response or
completion for example a prompt to C to five and a loop results in the model returning appropriate code tokens asure
open AI breaks down text into tokens which are words or character chunks to process requests the number of tokens
affects response latency and throughput for images token cost varies with image size and detail setting with low detail
images costing fewer tokens and high detail images costing more resources Azure open aai operates like other Azure
products where users create a resource within their Azure subscription deployments to use the service users
must deploy a model via deployment apis choosing the specific model for their needs prompt engineering crafting
prompts is crucial as they guide the model's output this requires skill as prompt construction is nuanced and
impacts the model's response models various models offer different capabilities and pricing Dolly creates
images from text while whisper transcribes and translates speech to text each has unique features suitable
for different tasks so that's an overview of azure open ey service the next topic we'll be covering
is azure open AI Studio developers can work with these models in Azure open AI Studio A web-based environment where AI
professionals can deploy test and manage llms that support generative AI app development on Azure access is currently
limited due to the high demand upcoming product improvements and Microsoft's commitment to responsible AI presently
collaborations are being prioritized for those who already have a partnership with Microsoft are engaged in lower risk
use cases and are dedicated to including necessary safeguards in Azure open AI Studio you
can deploy large language models provide F shot examples and test them in Azure open AI Studios chat playground the
image shows Azure open eyes chat playground interface where users can test and configure an AI chat bot in the
middle there's a chat area to type user messages and see the assistant's replies on the left there's a menu for
navigation and a section to set up the assistant including a reminder to save changes on the right adjustable
parameters control the eyes response Behavior like length Randomness and repetition users into queries adjust
settings and observe how the AI responds to fine-tune its performance so that's an overview of azure open AI
Studio let's take a look at the pricing for the model models in Azure open AI service starting off with the language
models we have GPT 3.5 Turbo with a context of 4K tokens cost 0.15 for prompts and
0.002 for completions per 1,000 tokens another version of GPT 3.5 turbo can handle a larger context of 16k tokens
with PRT and completion costs increased up to 0.003 and
0.004 respectively gpt3 5 Turbo 11106 with a 16k context has no available pricing gp4 turbo and gp4 turbo Vision
both have an even larger Contex size of 128k tokens but also have no listed prices the standard gp4 model with an 8K
token Contex costs 3 cents for prompts and 6 cents for completions and a larger context version of gp4 with 32k tokens
cost 6 cents for prompts and 12 cents for completions there are other models such
as the base models fine-tuning models image models embedding models and speech models they
all have their respective pricing but we won't be going through each of them in a lot of detail but essentially they are
all on a paper use pricing model it could be payer hour or paper token and so on the higher quality the model the
more expensive it will likely be so that's an overview of azure open AI Service
[Music] pricing hey this is Andrew Brown from exam Pro and the next topic will be
going over co-pilots co-pilots are a new type of computing tool that integrates with applications to help users with
common tasks using generative AI models they are designed using a standard architecture allowing developers to
create custom co-pilots tailored to specific business needs and applications co-pilots might appear as a chat feature
beside your document or file and they utilize the content within the product to generate specific results creating a
co-pilot involves several steps training a large language model with a vast amount of data utilizing services like
Azure open AI service which provide pre-trained models that developers can either use as his refin tune with their
own data for more specific tasks deploying the model to make it available for use within applications building
co-pilots that prompt the models to generate usable content enabling business users to enhance their
productivity and creativity through AI generated assistance co-pilots have the potential to revolutionize the way we
work these co-pilots use generative AI to help with first draft information synthesis strategic planning and much
more let's take a look at a few examples of co-pilot starting with Microsoft co-pilot so Microsoft co-pilot is
integrated into various applications to assist users in creating documents spreadsheets presentations and more by
generating content summarizing information and aiding in strategic planning it is used across Microsoft
Suite of products and services to enhance user experience and efficiency next we have the Microsoft being search
engine which which has an integrated co-pilot to help users when browsing or searching the Internet by generating
natural language answers to questions by understanding the context of the questions providing a richer and more
intuitive search experience Microsoft 365 co-pilot is designed to be a partner in your workflow integrated with
productivity and communication tools like PowerPoint and Outlook it's there to help you craft effective documents
design spreadsheets put together presentations manage emails and streamline other tasks GitHub co-pilot
is tool that helps software developers offering real-time assistance as they write code it offers more than
suggesting code Snippets it can help in Thoroughly documenting the code for better understanding and maintenance
additionally co-pilot contributes to the development process by providing support for testing code ensuring that
developers can work more efficiently and with fewer errors so that's an overview of
[Music] co-pilot hey this is Andrew Brown from exam Pro and the next topic will be
covering is prompt engineering prompt engineering is a process that improves the interaction between humans and
generative AI it involves refining the props or instructions given to an AI application to generate higher quality
responses this process is valuable for both the developers who create AI driven applications and the end users who
interact with them for example developers May build a generative AI application for teachers to create
multiple choice questions related to text students read during the development of the application
developers can add other rules for what the program should do with the prompts it receives system messages prompt
engineering techniques include defining a system message the message sets the context for the model by describing
expectations and constraints for example you're a helpful assistant that responds in a cheerful friendly manner these
system messages determine constraints and styles for the model's responses writing good prompts to maximize the
utility of AI responses it is essential to be precise and explicit in your props a well structured propt such as create a
list of 10 things things to do in Edinburgh during August directs the AI to produce a targeted and relevant
output achieving better results zero shot learning refers to an AI model's ability to correctly perform a task
without any prior examples or training on that specific task one shot learning involves the AI model learning from a
single example or instance to perform a task here is an example of prompt engineering with a user query and system
response so the user inputs can my camera handle the rainy season if I go to the Amazon rainforest next week some
The Prompt engineering components could be the weather resistance feature check users equipment database rainforest
climate data product specifications travel tips for photographers etc for the llm processing the AI system
integrates the user question with data about the Amazon's climate specifically during the rainy season and the product
information about the camera's weather resistance features it also references a database of the user's equipment to
ensure it's talking about the correct item and may include travel tips that are useful for photographers heading to
similar climate and the output results in your current camera model the proot markv is designed
with a weather sealed body suitable for high humidity and Rain conditions which matches the expected weather in the
Amazon rainforest for next week however for added protection during heavy rains consider using a rain cover next let's
take a look at the prompt engineering workflow this image describes a simplified step process for working with
AI models and prompt engineering why task understanding know what you want the AI to do two craft prompts write
instructions for the AI three prompt alignment make sure instructions match what the AI can do for optimizing prompt
improve the instructions for better AI responses five AI model processing the AI thinks about the instructions six
generating output the AI gives an answer or result seven output refinement fix or tweak the I's answer hey to iterative
improvement keep improving the instructions and answers so that's an overview of prompt engineering
[Music] the next topic we'll be covering is grounding grounding impr prompt
engineering is a technique used in large language models where you provide specific relevant context within a
prompt this helps the AI to produce a more accurate and related response for example if you want an llm to summarize
an email you would include the actual email text in the prompt along with a command to summarize it this approach
allows you to Leverage The llm for tasks it wasn't explicitly trained on without the need for training the model so
what's the difference between prompt engineering and grounding prompt engineering broadly refers to the art of
crafting effective prps to produce the desired output from an AI model grounding specifically involves
enriching prps with relevant context to improve the model's understanding and responses grounding ensures the AI has
enough information to process the prop correctly whereas prop engineering can also include techniques like format
style and the Strategic use of examples or questions to guide the AI the image outlines of framework for grounding
options in prompt engineering within the context of large language models grounding options these are techniques
to ensure llm outputs are accurate and adhere to responsible AI principles prompt engineering placed at the top
indicating its broad applicability this involves designing prompts to direct the AI toward generating the desired output
fine-tuning a step below in complexity where llms are trained on specific data to improve their task performance
training the most resource intensive process at the triangle base suggesting it's used in more extensive
customization needs llm Ops and responsible AI these foundational aspects emphasize the importance of
operational efficiency and ethical standards across all stages of LL and application development so that's an
overview of [Music] grounding hey this is Andrew Brown from
exampro and in this demo we'll be going over a short demo on what you can do with co-pilot with gp4 on Microsoft bang
so to get here you'll need to search for something like co-pilot Bing and click on the TR copilot and you should be able
to access this page so on here you have some suggested or popular prompts that people commonly use such as create an
image of a concept kitchen generate ideas for whacking new products how would you explain AI to a sixth grader
WR python code to calculate all the different flavor combination as for my ice cream parlor and so on you can
choose the conversation style ranging from more creative for more original and imaginative ideas more balanced or more
precise for more factual information we'll be going with somewhere in the middle so more balanced just for this
example on the bottom here you can type in any prompt you want so for example we can type something simple like summarize
the main differences between supervised and unsupervised learning for the AI 900 exam you'll see that it will start
generating an answer for you so for supervised learning data labeling in supervised learning the training data is
pre-labeled with the correct output values and it provides other objectives and
examples as well and for unsupervised learning No Labels unsupervised learning operates without labeled data it seeks
to discover patterns structures or relationships within the raw data notice how it uses sources from the internet
and if you want to learn more you can click on these links that it provides to directly go to the source of the
information which is very convenient so let's quickly check one out and it seems like the information we got was pretty
good and credible and on the bottom it also provides us some suggestions for
follow-up questions you may want to ask in the future that is related to the previous
PRT another cool feature of co-pilot is that it's integrated with dolly3 which is a image generation service so for
example you can say something like create an image of a cute dog running through a green field on a sunny
day so now you'll have to wait a little bit for it to generate the image that you described in your
prompt and there we go we have an adorable little puppy running through the
fields you also have the power to modify images if you're not satisfied with the result so they've provided a few options
for you here so for example we can add a rainbow in the background change it into a cat or make the sky pink and purple
let's try changing it to a cat so it's going to generate and change it from a dog to a cat and there we go it's now a
cute little cat running through the field you could also write code using
co-pilot so for example I can type in write a python function to check if a given number is
prime it'll start generating a piece of code for me it can write code in multiple
languages not just python so let's try out something with JavaScript let's try create a JavaScript
function to reverse a string and of of course we'll need to wait for the code to
generate so there we go here's our code for the function to reverse a string just as we asked for it so that's a
really quick and general demo for co-pilot with [Music]
gp4 hey this is Andrew Brown from exam Pro and in this follow along we're going to set up a studio with an Azure machine
learning uh Service uh so that it will be the basis for all the follow alongs here so I want you to do is go all the
way to the top here and type in Azure machine learning and you're looking for this one that looks like a science uh
bottle here and we'll go ahead and create ourselves our machine learning uh studio and so I'll create a new one here
and I'll just say um my studio and we'll hit okay and we'll name the workpace so we'll say my uh work
workplace we'll maybe say ml workplace here uh for containers there are none so
it'll create all that stuff for us I'll hit create and
create and so what we're going to do here is just wait for that creation okay all right so after a short little
wait there it looks like our studio set up so we'll go to that resource launch the studio and we are now in so uh
there's a lot of stuff in here but generally the first thing you'll ever want to do is get yourself a notebook
going so in the top left corner I'm going to go to notebooks and what we'll need to do is load some files in here
now they do have some sample files like how to use uh Azure ml so if we just quickly go through here um you know
maybe we'll want to look at something like uh Ms nist here and we'll go ahead and open this
one and maybe we'll just go ahead and clone uh this and we'll just clone it over
here okay and the idea is that we want to get this notebook running and so notebooks
have to be backed by some kind of compute so up here it says no compute found and Etc so what we can do here and
I'm just going to go back to my files oh it went back there for me but what I'm going to do is go all the way down
actually I'll just expand this up here makes it a bit easier close this tab out but uh what we'll do is go down to
compute and here we have our four types of comput so compute instances is when we're running notebooks compute clusters
is when we're doing training and inference clusters is when we have uh a inference pipeline uh and then attach
computer is bringing uh things like hdn sites or data bricks into here but for compute instances is what we need we'll
get ahead and go new you'll notice we have the option between CPU and GPU GPU is much more expensive see it's like 90
cents per hour for a notebook we do not need anything uh super powerful notice it'll say here development on notebooks
IDs lightweight testing here it says classical ml model training autom ml pipelines Etc
so I want to make this a bit cheaper for us here uh because we're going to be using the notebook to run uh cognitive
services and those cost next to nothing like they don't take much compute power uh and for some other ones we might do
something a bit larger for this this is good enough so I'll go ahead and hit next I'm just going to say my uh
notebook uh instance here we'll go ahead and hit create and so we're just going to have
to wait for that to finish creating and running and when it is I'll see you back here in a moment all right so after a
short little wait there it looks like our server is running and you can even see here it shows you you can launch in
jupyter Labs Jupiter vs code R Studio or The Terminal but what I'm going to do is go back all the way to our notebooks
just so we have some consistency here I want you to notice that it's now running on this compute if it's not you can go
ahead and select it uh and it also loaded in Python 3.6 there is 3.8 right now it's not a big deal which one you
use um but that is the kernel like how it will run this stuff now this is all interesting but I don't want to uh run
this right now what I want to do is get those cognitive Services uh into here so what we can do is just go up here and
we'll choose editors and edit in Jupiter lab and what that should do is open up a new tab
here uh is it opening if it's not opening what we can do is go to compute sometimes it's a bit
more responsive if we just click there it's the same way of getting to it um I don't know why but just sometimes that
link doesn't work uh when you're in the notebook and what we can do is well we're in here now we can see that this
is where uh this example project is okay um but what we want to do is get those cognitive services in here so I don't
know if I showed it to you yet but I have a repository I just got to go find it it's somewhere on my
screen um here it is okay so I have a repo called the free a a uh the free a it should be AI 900 I think I'll go
ahead and change that or that is going to get confusing okay so what I want you to do
here is um we'll get this loaded in so this is a public directory I'm just thinking there's a couple ways we can do
it we can go and uh use the terminal to grab it what I'm going to do is I'm just going to go download the
zip and this is just one of the easiest ways to install it and we need um to place it somewhere so here are my
downloads and I'm just going to drag it out here okay and uh what we'll do is upload that
there so I can't remember if it lets you upload entire folders we'll give it a go see if it lets us maybe rename this to
the free a or AI 900 there we'll say open uh yeah so it's individual file so it's not that big of a deal but we can
go and ahead and select it like that and maybe we'll just make a new folder in here we'll say this cognitive
[Music] services okay and uh what we'll do here is just
keep on uploading some stuff so we have assets so I have a couple loose Falls there and I know we have a
crew oops we'll have crew oops is not as responsive
um we want OCR uh I believe we have one called movie uh
reviews so we'll go into OCR here and upload the files that we have so we have a few files
there and we'll go back a directory here and I know movie uh reviews are just static
files and we have an objects folder so we will go back here to objects and then we'll go back and to
crew and we need a folder called Warf a folder called Crusher a folder called data and so for
each of these we have some images I think we're on Warf right yep we are
okay great so we will quickly upload all these well technically we don't really need to upload any of these well these
images we don't but I'm going to put them here anyway I just remember that uh these we just upload directly to the
service but because I'm already doing anyway I'm just going to put them here even though we're not going to do
anything with them all right and so now we are all set up
to do some cognitive services so I'll see you in the next video all right so now that we have our work environment
set up what we can do is go ahead and get cognitive Services hooked up because um we need that service in order to
interact with it because if we open up any of these you're going to notice we have a cognitive key and endpoint that
we're going to need so what I want you to do is go back to your Azure portal and at the top here we'll type in
cognitive Services now the thing is is that all these services are individualized but at
some point they did group them together and you're able to use them through unified um key and API Point that's what
this is and that's the way we're going to do it so we'll say add and uh it brought us to the
marketplace so I'm just going to type in cognitive services and then just click this one
here here and we'll hit create and uh we'll make a new one here I'm just going to call my uh Cog
Services say Okay um I prefer to be in Us East I will leave in US West it's fine and so in here we'll just say my
Cog services and if it doesn't like that I'll just put some numbers in there we
go we'll do standard so we will be charged something for that let's go take a look at the
pricing so you can see that the pricing is uh quite variable here but uh it's like
you'd have to do a thousand transactions before you are build uh so I think we're going to be okay for billing uh we'll
checkbox this here we'll go down below it's telling us about responsible AI notice uh sometimes services will
actually have you checkbox it but in this case it just tells us there and we'll go ahead and hit
create and I don't believe this took very long so we'll give it a second here yep it's
all deployed so we'll go to this resource here and what we're looking for are our keys and end
points uh and so we have two keys and two end points we only need a single key so I'm going to copy this endpoint over
we're going to go over to Jupiter lab and I'm just going to paste this in here I'm just going to put it in all the ones
that need it so this one needs one this one needs one this one needs
one and this one needs one and we will show the key here I guess it doesn't show but it copies of
course I will end up deleting my key before you ever see it but this is something you don't want to share
publicly and usually you don't want to embed Keys directly into a notebook but uh this is the only way to do it so it's
just how it is with Azure um so yeah all our keys are installed going back to the cognitive Services uh nothing super
exciting here but but it does tell us what services work with it you'll see there's an aster beside custom Vision
because we're going to access that through another app um but uh yeah cognitive servic is all set up and so
that means we are ready to uh start doing some of these Labs [Music]
okay all right so let's take a look here at computer vision first and computer vision is actually used for a variety of
different Services as you will see it's kind of an umbrella for a lot of different things but the one in
particular that we're looking at here is to describe image in stream if we go over here to the documentation this
operation generates description of image in a human reable language with complete sentences the description is based on a
collection of content tags which also returned by the operation okay so let's go see what that looks like in action so
the first thing is is that um we need to install this Azure cognitive Services Vision computer vision now we do have a
kernel and these aren't installed by default they're not part of the um uh machine learning uh the Azure machine
learning uh SDK for python I believe that's pre-installed but uh these AI services are not so what we'll do is go
ahead and run it this way and you'll notice where it says pip install that's how it knows to install and once that is
done we'll go run our requirements here so we have the OS which is for usually handling op like OS layer stuff we have
matte matte plot lib which is to visually plot things and we're going to use that to show images and draw borders
we need to handle images is I'm not sure if we're using numpy here but I have numpy loaded and then here we have the
Azure cognitive Services Vision computer vision we're going to load the client and then we have the credentials and
these are generic credentials for the cognitive Services credentials it's commonly used for most of these services
and some exceptions they the apis do not support them yet but I imagine they will in the future so just notice that when
we run something it will show a number if there's an aster it means it hasn't ran yet so I'll go ahead and hit play up
here so it goes an aster and we get a two and we'll go ahe head and hit play again and now those are loaded in and so
we'll go ahead and hit play okay so here we've just packaged our credentials together so we passed
our key into here and then we'll now load in the client uh and so we'll pass our endpoint and our key okay so we hit
play and so now we just want to load our image so here we're loading assets data.jpg let's just make sure that that
is there so we go assets and there it is and we're going to load it as a stream because you have to pass streams along
so we'll hit play and you'll see that it now ran and so now we'll go ahead and make that
call okay great and so we're getting some data back and notice we have some properties person wall indoor man
pointing captions it's not showing all the information sometimes you have to extract it out but we'll take a look
here so uh this is a way of showing mat pla lib in line I don't think we have to run it here but I have it in here anyway
and so what it's going to do is it's going to um show us the image right so it's going to print us the image and
it's going to grab whatever caption is returned so see how there's captions so we're going to iterate through the
captions it's going to give us a confidence score saying it thinks it's this so let's see what it comes out
with okay and so here it says Brent spider Spiner looking at a camera so that is the actor who plays data on Star
Trek has a confidence score of 57. 45% even though it's 100% correct uh they probably don't know contextual things
like um uh in the sense of like pop culture like they don't know probably start Tre characters but they're going
to be able to identify celebrities because it's in their database so that is um uh uh the first introduction to
computer computer vision there but the key things you want to remember here is that we use this describe an image
stream uh and that we get this confidence score and we get this contextual information okay and so
that's the first one we'll move on to um maybe custom Vision next
all right so let's take a look at custom Vision so we can do some um classification and object detection so
um the thing is is that it's possible it's possible to launch custom Vision through the marketplace so if we go
we're not going to do it this way if you type in custom Vision it never shows up here but if you go to the marketplace
here and type in custom vision and you go here you can create it this way but the way I like to do it I
think it's a lot easier to do is we go up the top here and type in custom vision. and you'll come to this website
and what you'll do is go ahead and sign in it's going to connect to your Azure account and once you're in you can go
ahead here and create a new project so the first one here is I'm just going to call this the Star Trek crew we're going
to use this to identify different Star Trek members we'll go down here and uh we haven't yet created a resource so
we'll go create new my custom Vision resource we'll drop this down we'll put
this in our Cog services uh we'll go stick with um Us West as much as we can here we have fo Ando fo is blocked up
for me so just choose so I think fo is the free tier but I don't get it and um once we're back here we'll go
down below and choose our standard and we're going to have a lot of options here so we have between classification
and object detection so classification is when you have an image and you just want to say what what is this image
right and so we have two modes where we can say let's apply multiple labels so let's say
there were two people in the photo or whether there was a dog and cat I think that's example that use a dog and a cat
or you just have a single class where it's like what is the one thing that is in this photo it can only be of one of
the particular categories this is the one we're going to do multic class and we have a bunch of different domains
here and if you want to you can go ahead and read about all the different domains and their best use case but we're going
to stick with A2 this is optimized for so that it's faster right and that's really good for our demo so we're going
to choose General A2 I'm going to go ahead and create this project and uh so now what we need to do
is start labeling our our our content so um what we'll do is I just want to go ahead and create the tags ahead of time
so we'll say Warf we'll have uh data and we'll have Crusher and now what we'll do is we'll
go ahead and upload those images so you know we uploaded the jupyter notebook but it was totally not necessary so here
is data because we're going to do it all through here and we'll just apply the data tag to them all at once which saves
us a lot of time I love that uh we'll upload now uh war and I don't want to upload them all
I have this one quick test image we're going to use to make sure that this works
correctly and I'm going to choose Warf and then we'll go ahead and add Beverly there she
is Beverly Crusher okay so we have all our our images in I don't know how this one got
in here but it's under worth it works out totally fine so uh what I want to do is uh go ahead and train this small
because they're all labeled so we have a ground truth and we'll let it go ahead and train so we'll go and press train
and we have two options quick training or Advanced Training Advanced Training where we can increase the time for
better accuracy but honestly uh we just want to do quick training so I'll go ahead and do quick training and it's
going to start its iterative process notice on the left hand side we have probability threshold the minimum
probability score for a prediction to be valid when calcul calculating precision and recall so we uh the thing is is that
if it doesn't at least meet that requirements it will quit out and if it gets above that then it might quit out
early just because it's good enough okay so training doesn't take too long it might take 5 to 10 minutes I can't
remember how long it takes but uh what I'll do is I'll see you back here in a moment okay all right so after waiting a
short little while here looks like our results are down we get 100% um match here so these are our evaluation metrics
to say whether uh the model was uh uh achieved its actual goal or not so we have Precision recall and I believe this
is average Precision uh and so it says that it did a really good job so that means that it should have no problem um
matching up an image so in the top right corner we have this button that's called quick test and this is going to give us
the opportunity to uh quickly test these so what we'll do is browse our files locally here and uh actually I'm going
to go to uh yeah we'll go here and we have Warf uh and so I have this quick image here we'll test and we'll see if
it actually matches up to bwf and it says 98.7% Warf that's pretty good I also have some additional images
here I just put into the repo to test against and we'll see what it matches up because I thought it'd be interesting to
do something that is not necessarily uh them but it's something pretty close to um you know it's pretty close to what
those are okay so we'll go to crew here and first we'll try Hugh okay and Hugh is a borg so he's
kind of like an Android and so we can see he mostly matches to data so that's pretty good uh we'll give another one go
Marto is a Klingon so he should be matched up to Warf very strong match to Warf that's pretty good and then palaski
she is a doctor and female so she should get matched up to Beverly Crusher and she does so this works out pretty darn
well uh and I hadn't even tried that so it's pretty exciting so now let's say we want to go ahead and well if if we
wanted to um make predictions we could do them in bulk here um I believe that you could do them in bulk but
anyway yeah I guess I always thought this was like I could have swore yeah if we didn't have these images before I
think that it actually has an upload option it's probably just the quick test so I'm a bit confused there um but
anyway so now that this is ready what we can do is go ahead and publish it uh so that it is publicly accessible so we'll
just say here a crew model okay and we'll drop that down say publish and once it's published now we
have this uh public out so this is an endpoint that we can go hit pragmatically uh I'm not going to do
that I mean we could use Postman to do that um but my point is is that we've basically uh figured it out for um
classification so now that we've done classification let's go back here to uh the vision here and let's now let's go
ahead and do object detection [Music] okay all right so we're still in custom
Vision let's go ahead and try out object detection so object detection is when you can identify a particular items in a
scene um and so this one's going to be combadge that's what we're going to call it because we're going to try to detect
combadge we have more domains here we're going to stick with the general A1 and we'll go ahead and create this
project here and so what we need to do is add a bunch of images I'm going to go ahead
and create our tag which is going to be called combadge uh you could look for multiple different kinds of labels but
then you need a lot of images so we're just going to keep it simple and have that there I'm going to go ahead and add
some images and we're going to go back um a couple steps here into our objects and here I have a bunch of photos and we
need exactly 15 to train so we got one two 3 4 5 6 7 8 9 10 11 12 13 14 15 16 and so I threw an additional image in
here this is the badge test so we'll leave that out and we'll see if that picks up really well and yeah we got
them all here and so we'll go ahead and upload those and we'll hit upload files
okay and we'll say done and we can now begin to label so we'll click into here and what I want to do if you hover over
it should start detecting things if it doesn't you can click and drag we'll click this one they're all com badges so
we're not going to tag anything else here okay okay so go here hover over is it going to give me the combadge no so
I'm just draag clicking and dragging to get it okay okay do we get this combadge
yes do we get this one yep so simple as that okay it doesn't always get it but
uh most cases it does okay didn't get that one so we'll just drag it
out okay it's not getting that one it's interesting like that one's
pretty clear but uh it's interesting what it picks out and what does what it does not grab eh so it's not getting
this one probably because the photo doesn't have enough contrast and this one has a lot hoping
that that gives us more data to work with here yeah I think the higher the contrast it's easier for it to uh
um detect those it's not getting that one it's not getting that one okay there
we go yes there are a lot I know I have some of these ones that are packed but
there's only like three photos that are like this yeah they have badges but they're
slightly different so we're going to leave those out oops I think it actually had that
one but we'll just tag it anyway and hopefully this will be worth the uh the effort
here there we go I think that was the last one okay great so we have all of our
tagged photos and what we can do is go ahead and train the model same option quick training Advanced Training we're
going to do a quick training here and notice that the options are slightly different we have probably threshold and
then we have overlap threshold so the minimum percentage of overlap between predicted bounding boxes and ground
truth boxes to be considered for correct prediction so I'll see you back here when it is done all right so after
waiting a little bit a while here it looks like um it's done it's trained and so Precision is at 75% so Precision the
number will tell you if a tag is predicted uh by your model How likely that it's likely to be so how likely did
it guess right then you have recall so the number will tell you out of the tags which should be predicted correctly what
percentage did your model correctly find so we have 100% uh and then you have mean average Precision this number will
tell you the overall object detector performance across all the tags okay so what we'll do is we'll go ahead and uh
do a quick test on this model and we'll see how it does I can't remember if I actually even ran this so it'll be
curious to see the first one here um it's not as clearly visible it's part of their uniform so I'm not expecting to
pick it up but we'll see what it does it picks up pretty much all of them with exception this one is definitely not a
comp badge but uh that's okay only show suggests obious the probabilities above the selected
threshold so if we increase it uh we'll just bring it down a bit so there it kind of improves it um if we
move it around back and forth okay so I imagine via the API we could choose that let's go look at our other sample image
here um I'm not seeing it uh where did I save it let me just
double check make sure that it's in the correct directory here okay yeah I saved it to the wrong place
just a moment um I will place
it just call that bad test two one second okay and so I'll just browse here again and so here we have another one
see if it picks up the badge right here there we go so looks like it worked so uh yeah I guess custom vision is uh
pretty easy to use and uh pretty darn good so what we'll do is close this off and make our way back to our Jupiter
labs to move on to um our our next uh lab here [Music]
okay all right so let's move on to the face service so just go ahead and double click there on the left hand side and
what we'll do is work our way from the top so the first thing we need to do is make sure that we have the computer
vision installed so the face service is part of the computer vision API and once that is done we'll go ahead and uh do
our Imports very similar to the last one but here we're using the face client we're still using the Cog cognitive
service credentials we will populate our keys we make the face client and authenticate and we're going to use the
same image we used um uh prior with our computer vision so the data one there and we'll go ahead and print out the
results and so we get an object back so it's not very clear what it is but here if we hit
show okay here it's data and it's identifying the face ID so going through this code so we're just saying open the
image we're going to uh set up our figure for plotting uh it's going to say well how many faces did it detect in the
photo and so here it says detected one face it will iterate through it and then we will create a bounding box around the
images we can do that because it returns back the face rectangle so we get a top left right Etc and uh we will draw that
Wrangle on top so we have magenta I could change it to like three if I wanted to uh I don't know what the other
colors are so I'm not even going to try but yeah there it is and then we annotate with the face ID that's the
unique identifier for the face and then we show the image okay so that's one and then if we wanted to get more detailed
information like attribute such as age emotion makeup or gender uh this resolution image wasn't large enough so
I had to find a different image and and do that so that's one thing you need to know as if it's not large enough it
won't process it so we're just loading data large very similar process but it
is uh the same thing detect with stream but now we're passing in um return face attributes and so here
we're saying the attributes we want uh and there's that list and we went through it in the lecture content and so
here we'll go ahead and run this and so we're getting more information so that magenta line is a bit hard to see I'm
just going to increase that to three okay still really hard to see but that's okay so approximate age 44 I
think the actor was a bit younger than that uh uh data technically is male presenting but he's an Android so he
doesn't necessarily have a gender I suppose he actually is wearing a lot of makeup but all it detects is it I guess
it's only particular on the lips and the eyes so it says he doesn't have makeup so maybe there's a color you know like
ey Shadow stuff maybe we would detect that in terms of personality I like how it's he's a a 002 Point per SB but he's
neutral right uh so just going through the code here very quickly so again it's the number of faces so it detected one
face uh and then we draw a bounding box around the face for the detected attributes it's uh return back in the
data here so we just say get the pH attributes turn it into a dictionary and then we can just uh get those values and
uh iterate over it so that's as complicated as it is um and so there we [Music]
go all right so we're on to uh our next cognitive service let's take a look at form recognizer all right and so form
recognizer uh it tries to identify um like forums and turns them into readable things and so they have one for uh
receipts in particular so at the top finally we're not using um the computer uh computer vision we actually have a
different one so this one's Azure AI form recognizer so we'll run that there but this one in particular isn't up to
date in terms of using it like um notice all the other ones they're using uh the cognitive service credential so for this
we actually had to use the Azure key uh credential which was annoying I tried to use the other one to be consistent um
but I I couldn't use it okay so what we'll do is run our keys like before we have a client very similar
process and this time we actually have a receipt and so we have begin recognize receipt so it's going to analyze the
receipt information and then it's what it's going to do is show us the image okay just so we have a reference to look
at now the image isn't actually yellow it's a white background I don't know why when it renders out here it does that
but that's just what happens and uh it even obscures the server name I I don't know why um but
anyway if we go down below um this is return results up here right so we got our results and so if we just print out
uh the results here we can see we get a recognized form back we get fields and some additional things and if we go into
the uh the fields itself we see there's a lot more information if you can make out like here it says Merchant phone
number form field label value and there's the number 512707 so for these things here like um
the receipts if we can just find the API quickly here it has predefined
Fields I'm not sure um yeah business card Etc um like if we just type in
merchant I'm just trying to see if there's a big old list here it's not really showing us a full list but these
are are predefined um things that are returned right so they've defined those uh maybe it's over here
there we go so these are the predefined ones that extracts out so we have uh receipt type Merchant name etc etc and
so if we go back to here you can see um I I have the field called Merchant name so we hit there it says Alm draft out
Cinema let's say we want to try to get that balance maybe we can try to figure out which one it is I never ran this
myself when I I made it so we'll see what it is but here it has total price what's interesting is that these this
has a space so it's kind of unusual ual you think it'd be together but let's see if that
works okay doesn't like that maybe that's just a typo on their part okay so we get none uh let's try
price see what it picks up nope nothing um we know that the phone number is there so we'll give the
phone number there we go so you know it's an okay service but uh you know uh you know
you're you're mileage will vary based on uh what you do there maybe we could try total because that makes more sense
right uh yeah there we go okay great so yeah it is pulling out the information um and so that's pretty much all you
need to know about that service there [Music] okay let's take a look at some of our
OCR capabilities here uh and I believe that's in computer vision so we'll go ahead and open that up at the top here
we'll install computer vision as we did before very similar to the other computer vision task but this time we
have a couple of ones here that'll explain that as we go through here we'll load our keys we'll do our credentials
we'll load the client okay and then we have this um function here called printed text so what this function is
going to do is it's going to uh print out the results of whatever text it processes okay so the idea is that we
are going to feed in an image and it's going to give us back out the text for the the image so we'll run this function
and I have two different images cuz I actually ran it on the first one and the results were terrible and so I got a a
second image and it was a bit better okay so we'll go ahead and run this it's going to show us the image okay and so
this is the photo and it was supposed to extract out Star Trek the Next Generation but because of the artifacts
and size of the image we get back uh not English okay and so you know maybe a high resolution image it would have um a
better a better time there um but that is what we got back okay so let's go take a look at our second image and see
how it did and this one I'm surprised it actually extracts out a lot more information you can see really has a
hard time with the Star Trek font but we get Deep Space 9 Nana Visitor tells all Life Death some errors here so it's not
perfect um but you know you can see that it does something here now there is the O this is like for OCR where we have
like for very simple images and text this is where we use the recognized printed text in stream but uh if we were
doing this for larger amounts of text and we want to do this uh want this analyzed asynchronously then we want to
use the read API and it's a little bit more involved um so what we'll do here is load a different image and this is a
script we'll look at the image here in a moment um but here we read in stream and we create these
operations okay and what it will do is it will asynchronous asynchronously send all the information over okay uh so I
think this is supposed to be results here minor typ and um we will go ahead and give that a
run okay and so here you can see it's extracting up the image if we want to uh uh see this image I thought I uh I
thought I showed this image here but I guess I don't yeah it says plot image here to show us the
image uh path it's up here it doesn't want to show us it's
funny because this one up here is showing us no problem right um well I can just show you the image
it's not a big deal but I'm not sure why it's not showing up here
today so if we go to our assets here I go to OCR uh I'm just going to open this
up it's opening up in Photoshop and so this is what it's transcribing okay so this is a thing this is like a guide to
Star Trek where they talk about like you know what what makes St Trek Star Trek so just looking here it's actually
pretty darn good okay but like read API is a lot more uh efficient because it can work uh
umly and so when you have a lot of text that's what you want to do okay um and like it's feeding in each individual
line right so that it can be more effective that way um so let's go look at some hand Rd and stuff so just in
case the image doesn't pop up we'll go ahead and open this one and so this is a a a handwritten note that uh William
Shatner wrote to a fan of Star Trek and it's basically incomprehensible I don't know if you can read that here but see
was very something he was something hospital and healthy was something he was something I can't even read it okay
so let's see what uh the machine thinks here and uh it says image path yeah it's called path let's just change that out
go ahead and run that and run that there and we'll go ahead and run it and here we get the image so uh poner us very
sick he was the hospital his Bey was Etc beat nobody lost his family knew Captain Halden so reads better than how I could
read it honestly like it is it's really hard right like if you looked at this like that looks like difficult
was Beady healthy I could see why it's guessing like that right dying that looks like dying to me you know what I
mean so you it's just poorly hand handwritten but I mean it's pretty good for what it is so uh yeah there you
[Music] go all right so let's take a look at another cognitive service here and this
one is text analysis and uh so what we'll do is install the Azure cognitive Services
language uh text analytics here so we go ahead and hit run all right and once that's uh installed uh this one is using
the cognitive Services credentials so it's a little bit more standard with our other ones here we'll go ahead and run
that there uh we'll make our credentials load our client and this one what we're going to do is try to determine
sentiment and understand why people like a particular movie or not so I've loaded a bunch of reviews um they are again I
can show you the data if it helps uh and so I'm just trying to find my right folder here and so if we go back look
our movie reviews here's like a a review someone wrote first Contact just works it works as a rousing chapter in the
Star Trek to less extent it works as a mainstream entertainment so different reviews for Star Trek first Contact
which was a a very popular movie back in the day um so what we'll do is we will load uh the reviews so
it's just iterating through the text files and showing us what the reviews are so here we can see all the ridden
text had a lot of trouble getting the last one to display but it does get loaded in and so here we're using the
the the um text analysis to show us uh key phrases because maybe that would give us an indicator and so that's the
object back but maybe that give us an indicator as to like what people are saying as important things so here we
see Borg ship Enterprise smaller ship escapes neutral zone travels contact damage uh co-writer Beautiful Mind
sophisticated science fiction best whales Leonard neoy okay uh wealth of unrealized potential uh filmmaker John
fr s okay so very interesting stuff as it here Borg ship again you've seen Borg ship a lot so that is kind of key
phrases let's go get uh C or customer sentiment or how people felt about it did they like it or not and so here we
just call sentiment and um what we'll do is if it's uh above five then it's positive and it's below five then it's a
negative review I think most people uh thought it was a very good film uh so this one says it's pretty low nine so
let's go take a look at that one uh it wasn't actually showing rendered there so maybe we'll have to open it up
manually see if that's actually accurate it's empty so there you go I guess we had a blank one in there um I must have
forgot to paste it in but that's okay uh that's a good indicator that uh you know that's what happens if you don't have it
so let's look at number one then which is uh well actually this one is nine this is 04 this one here is eight so
we'll open up eight when the Borg launched on Earth the Enterprise is sent to the neutral zone etc etc however a
smaller ship escapes travels the Enterprise follows back um meanwhile the survivors so like this is a synopsis it
doesn't say whether they like it or they don't but it was just 04 I I guess so there's nothing positive about it right
um if we look at one that was this one's pretty low which is no no it's not it's one so it seems like this person
probably really liked it or no I guess that's actually pretty low because it's one it's not nine Nine's very high let's
take a look at this one review number two uh if we go up here the doo has improved the story mon
turn the show but there's a wealth of unrealized potential so that's a fair one saying they maybe they don't like it
as much I don't know if they give it two stars right we could probably actually correlate it with the actual results
because I did get these off of IMDb and Rotten Tomatoes but uh yeah there you go that is Tex
[Music] analysis all right so now we're on to Q&A maker and so we're not going to need
to do anything pragmatically because Q&A maker is all about no code or low code to build out a questions and answers uh
uh bot service so what we'll do is go all the way up here and I want you to type in Q andm maker. a because as far
as I'm aware of it's not accessible through the portal sometimes you can find these things um again if we go to
the marketplace I'm just curious I'm going just take a look here really quickly uh
whenever it decides to log Us in here okay great so I'll go over to Marketplace and probably if we typed in
Q&A maybe we'd see something here Q&A yeah so we go here um give it a second
here seems like Azure is a little bit slow right now usually varies fast but uh you know
the service varies well it's not loading for me right now but that's okay because we're
not going to do it that way anyway um so uh again go to Q&A maker. and what I want you to do is go all way to the top
in the right corner and we'll hit sign in and what we'll be doing is connecting via our single sign on with our account
so it already knows I have an account there I'm going to give it a moment here and I'm going to go ahead and just give
it a second there we go so it says I don't have any
um knowledge bases which is true so let's go ahead and create ourselves a new knowledge base and here we have the
option between stable and preview I'm going to stick with stable because I don't know what's in preview I'm pretty
happy with uh that and so we need to connect uh Q&A Service uh uh Q&A service to our knowledge base and so back over
here in Azure actually I guess we do have to make one now that I remember we actually have to create a Q&A maker
service so I'll go down here and put this under my Cog services we'll say my um
Q&A Q&A service might complain about the name uh yep so I'll just put some numbers here
we'll pick uh free tier sounds good so I'll go free when I actually get the option that's what I will choose um down
below we'll choose free again usse sounds great to me uh it generates out the name it's the same name as here so
that's fine uh we don't need app insights but I'm going to leave it enabled because I think it changes it to
standard or s zero when you uh do not um have it enabled unusually and so we will create our Q&A
maker service give it a moment here and it says I remember it will say like even if you try it might have to
wait 10 minutes for it to create the service so even though even after it's provisioned um it'll take some time so
what we should do is prepare our doc because it can take in a variety different files and I just want to show
you here that uh the Q&A they have a whole paper here formatting the guidelines
and basically it's pretty smart about knowing where headings and answers is so for unstructured data we just have a
heading and we have some text so let's write some things in here that we can think of since we're all about
certification we should write some stuff here so how many adus certifications are there I believe right now there are uh
11 uh adus certifications okay and maybe if we use our headings
here this would probably be a good idea here y okay another one could be um how many F
fundamental Azure certifications are there and uh we'll give this a heading
we'll say um there are three Azure I think there's three there's other ones right like Power Platform stuff but just
being Azure specific there are three Azure uh fundamental certifications certification
so we have um the dp900 the AI 900 um the a900 I guess there's four there's the sc900
right so there are four okay we'll say which is the
hardest um Azure assoc Azure Association certification and uh what we'll say here
is I think I mean it's my my opinion is it's the Azure administrator had some background noise there that's why I was
a bit pausing there but the Azure administrator a 104 I would say that's the hardest uh which is
harder um the uh adabs or Azure certifications I would say uh Azure certifications are
harder because they uh check uh exact steps for implementation
where AWS focuses on Concepts okay so we have a bit of a um
knowledge base here so I'll save it and assuming that this is ready because we need a little bit time to put this
together we'll go back to q a get hit a refresh here give it a moment drop it down
choose our service and uh notice here that we have chitchat
extraction and only extraction we're going to do chitchat I will say uh my or this will be uh the reference name you
change any time this will be like uh uh certification Q&A and so here we want to populate so
we'll go to files here I'm going to go to my desktop and here it is I'll open
it we will choose professional tone go ahead and create that and so I'll see you back here moment all right so after
waiting a short little time here it loaded in our data so you can see that it it figured out which is the question
which is the answer and also has a bunch of default so here if somebody was asked something very s uh silly like can you
cry I'll say I don't have a body it has a lot of information pre-loaded for us which is really nice if we wanted to go
ahead and test this we could go and say um we'll go here and then we'll write in uh we say like
hello I say boring says good morning okay so we'll
say um how many uh certifications are there we didn't say AWS but let's just see what
happens and so it kind of inferred even though we didn't say AWS in particular so and notice that there's ads and Azure
so how many fundamental Azure certifications things like that and so it chose AWS so it's not like the
perfect service but it's pretty good I wonder what would happen if we um placed in uh one that's like Azure I don't know
how many Azure Sears there are we'll just say like there's 11 12 I can't never remember they're always adding
more but uh it I want to close this here there we go so let's just go add a new key pair here and we'll say how many
Azure [Music] certification are there I should have
said certifications I'll probably just set one moment so there there are 12 Azure
certifications who knows how many they have they could like 14 or something we could say like between 11 and
14 they just add them they just update them too frequently I can't keep track so uh we'll go here and we'll just say
certifications and we will save and retrain so we'll just wait here a moment great and so now we go ahead and
test this again so we'll say how many certifications are
there and see it's pulling the first answer if I say uh Azure if it's see if it gets the right one
here how many Azure certifications are there okay so you know uh maybe you'd have to say you'd have to have a generic
one for that match so if we go back here and we say how many
certifications are there you say uh you know like uh uh which certification uh uh which Ser
cloud service provider here we got ads
Azure uh prompt you can use Guides Through conversational flow prompts are used to link Q&A Pairs and can be
displayed um I haven't used this yet but I mean it sounds like something that's pretty good um because there is
multi-turn in this so the idea is that if you had to go through multiple steps you could absolutely do that um we try
this a little bit here uh follow prompt you can use the guide use convert prompts are used to link Q&A pairs
together texture button for suggested action oh okay so maybe we just do like AWS link to Q&A and then so search an
existing Q&A or create a new one um so it say like how many eight of us okay we're typing it
in context only this Falls up will not be understood out of the context flow sure because it should be within context
right and uh here we can do another one we say like um Azure we'll say how many
azure context only oops it uh got away from me there we'll save
that and uh what we'll do is save and train so we go back here and we'll say how
many uh certifications are there enter so we have to choose AWS and so there we go so we got something that
works pretty good there since I'm happy with it we can go ahead and go and publish that so we's say
publish and now that it's published we could use Postman or curl to uh trigger it but what I want to do is create a bot
because with Azure bot Services then we can actually utilize it um with other IND ations right it's a great way to uh
um use your Bot or to actually host your Bot so we'll go over here it'll link it over uh if you don't click it it doesn't
preload it in so it's kind of a pain if you lose it you have to go back there and click it again but uh let's just say
um certification q and day and we will look through here so all
going to go with free premium messages 10K 1K premium message units messages I'm kind of confused by the pricing but
F0 usually means free so that's what I'm going to go for that SDK or nodejs I'm going to use NOS not that we're going to
do anything there with it go ahead and create that and I don't think this takes too
long we'll see here and just go ahead and click on that there I'll just wait here a bit I'll see
you back here in a moment all right so after waiting I don't know about 5 minutes there it looks like our bot
service is deployed we'll go to that resour there uh you can download the bot source code actually never did this uh
so I don't know what it looks like so be curious to see this um just to see what the code is I assume that because we Cho
chose nodejs it would give us um that as the default there so download your c as you bought creating the source zip not
sure how long this takes might be regretting clicking on that but uh what we'll do is we'll go on
the left hand side here to channels because I just want to show uh here yeah I don't not didn't
download uh we'll try here in a second but um what we'll do is we'll go back po profile uh unspecified bot what are you
talking about yeah maybe it needs some time so you know maybe we'll just give
the bot a little bit of time here I'm not sure why it's giving us a hard time because this bot is definitely deployed
if we go over to our bot right bought Services it is here sometimes there's like latency you know with uh Azure oh
there we go okay see it works now fine right and so I want to show you that there's different channels and these are
just easy ways to integrate your Bot into different services so whether you wanted to use it with Alexa group me
Skype telepon twilio Skype business apparently they don't have that anymore because I got s teams now right uh keik
which I don't know people still use that slack we should had Discord telegram Facebook email um that's kind of cool
but teams teams is a really good one I use teams uh there's a direct line Channel I don't know what that means and
there's web chat which is just having like an ined code so if we go over we can go and test it over here just
testing our web chat and so it's the same thing as before but we just say things like uh um how many
certifications are there let Azure and get a clear answer back we'll go back up to our overview
let's try to see if we can download that code again I was kind of curious uh what that looks
like if it will download must be a lot of code eh
there we go so now we can hit download and so there is the code I'm going to go ahead and open that up uh so yeah I
guess when we chose JavaScript that made a lot more sense let's give it a little peek here I'm just going
to uh drop this on my desktop here so let make a new folder here and call this uh bot
code okay I know you can't see what I'm doing here but uh let's go here and d double click into here and then
just drag that code on in and then what we can do is open this up in VSS code I should have VSS code
running somewhere around here just going to go ahead and open that I'm off screen here I'll just show you my screen in a
moment say show code oops file open folder bot code
okay and uh we'll come all the way back here and so we got a lot of code here never looked at this before but you know
I'm a pretty good programmer so it's not too hard for me to understand um so looks like you got API
request things like that I guess it would just be like if you needed to integrate into your application then it
kind of shows you all the code there I'm just trying to see our dialogue choices nothing super
exciting okay you know when I go and make the um was it the AI or the AI 100 whatever the data scientist course is
I'm sure I'll be a lot more thorough here but I was just curious as to what that looks like now if we wanted to have
an easy integration uh we can get an M code for this so if we go back to our channels I
believe uh we can go and is it edit ah yes so here we have a code so what I'll do is go back to jupyter Labs
I'm just going to go make a new empty um notebook so we'll just go up here and say
notebook and this can be for our Q&A doesn't really matter what kernel uh we'll say Q and A maker just
to show like if you wanted a very very simple way of integrating your Bot um we would go back over
to wherever it is here ah here we are I'm going to go ahead and copy this iframe I think it's percentage
percentage HTML so it treats this cell as HTML and I don't have any HTML to render so
we will place that in there and notice we have to replace our secret key so I will go back here and I will show my key
and we will copy that and we will paste that key in here and then we'll run
this and I can type in here where am I just ask silly things uh who are you
how many Azure certifications are there well I wonder if I just leave the are there off let's
see if it figures it out okay cool so uh yeah I mean that's pretty much it with Q&A
maker um so yeah that's great so I think we're done here and we can move on to uh checking out uh leis or Luis learning
understanding to make a more uh robust bot okay
[Music] all right so we are on to our last cognitive service and this one is going
to be uh lwis or Louise depending on how you like to say it it's Luis which is language understanding so you type in
luis. a uh and that's going to bring us up to this external website still part of um Azure just has its own domain and
so here we'll choose our subscription and we have no author authoring source so I guess we'll have to go ahead and
create one ourselves so go down here and we'll choose my cognitive Services asure resource name so my o uh service or my
cognitive service create new cognitive service account but we already have one so I
don't want to make another one right it should show up here right are valid in the author authoring
region so it's possible that we're just in the incorrect region so we might end up creating two of these and that's
totally fine I don't care it's as long as we get this working here because we're going to delete everything at the
end anyway and so just say my Cog service 2 and uh we'll say West us because I
think that maybe we didn't choose one of these regions let's go double check uh if we go back to our
portal just the limitations of the service right so we'll go to my Cog Services here um I just want to go uh
cognitive services so just want to see where this is deployed and this is in um you West
us yes I don't know why it's not shown up there but whatever if that's what it wants we'll give it what it wants
okay shouldn't give us that much trouble but hey that's how it goes and so we have an author authoring
service I'm going to refresh here and see if it added a second one it didn't so all right
that's fine so we'll just say uh my sample bot um we'll use English as our culture
if nothing shows up here don't worry you can choose it later on I remember the first time I did this it didn't show up
and so now we have my Cog service my custom vision service we want Cog service
so um anyway it tells you about schema like how you make a schema animates talking about like bot action intent and
example utterance but we're just going to set up something very simple here so we're going to create our attent the one
that we always see is uh flight booking so I'll go here and do that and what we want to do is write an
undering so like uh book May flight to Toronto okay and so if someone were to type that in then the idea is it would
return back the intent this value and metadata around it and we could programmatically provide code right so
what we need is identity identities and we can actually just click here and uh make one here so enter named identity
we'll just call this location okay here we have an option machine learned and list if you flip
between it this is like imagine you have a ticket order and you have these values that can uh change or you just have a
value that always stays the same like list so that's our airport that makes sense we'll do
that if we go over to ENT entities we can see it here all right so uh nothing super
exciting there but what I want to show you is if we go ahead and um we should probably add fight booking should be uh
how about book flight flight booking fight booking okay so we'll go ahead and I know there's
only one but we'll go ahead and train our model because we don't need to know tons
right we cover a lot in the lecture content uh to build a complex spot is more for the uh associate level um but
now what we can do is go ahead and test this and we'll say book me a flight to Seattle okay and notice here it says
book flight we can go inspect it and we get some additional data so top scoring so it says How likely that was the
intent um okay so you get kind of an idea there there's additional things here it
doesn't really matter um we'll go back here and we will go ahead and publish our model so we can put it into a
production slot you can see we have sentiment analysis speech priming we don't care about either of those
things we can go and see where our endpoint is and so now we have uh an endpoint that we can work with um so
yeah I mean that's pretty much all you really need to learn about Lewis um but uh I think we're all done for cognitive
services so we're going to keep around our our notebook because um we're going to still use our jupyter notebook for
some other things things but what I want you to do is make your way over to um your resource groups because if
you've been pretty clean it's all within here we'll just take a look here so we have our
Q&A all of our stuff here I'm just making sure it's all there and so I'm just going to go ahead and delete this
Resource Group and that should wipe away everything okay for the cognitive Services
part all right so we're all good here and I'm just going to go off and I'll leave leave this open because it's
always a pain to get back to it and reopen it but let's make our way back to the home here in the Azure uh machine
Learning Studio and now we can actually explore building up machine learning [Music]
pipelines okay so we are on to the ml uh uh follow alongs here so we're going to learn how to build some pipelines so
first I think is the easiest would be Auto automated ml or also know as autom ml the idea here is it's going to just
um build out the entire pipeline for us so we don't have to do any thinking we just say what kind of model we want to
run and have it to make a prediction so what we'll do is a new automated ML and we're going to need a data set so I
don't have one but the nicest thing is they have these open data sets so if you click here you'll see there is a bunch
here and a lot of these you'll come across quite often not just on Azure but other places like this diabetes one I've
seen it like everywhere okay uh and so like if we just go click here maybe we can read a bit more here so diabetes
data set 422 samples with 10 features ideal for getting started with machine learning algorithms it's one of the
popular pyit learn toy data sets it's probably where I've seen it before though it's not showing up there uh you
scroll on down you can see the data uh you notice that it's available AZ your notebooks data bricks and Azure synapse
uh the thing is we have these values so age sex BMI BP and the Y is trying to make a prediction it's trying to say
what's the likelihood of you having diabetes or not and so it's not a Boolean value so it's not a binary
classifier it's kind of on a uh well I guess you would be doing binary classif classification say do you have di
diabetes or you can make a prediction to say what's the likelihood or this value if you gave another value in there but
um anyway you this is the predicting value a lot of times this is X so everything here is X and this is
considered y the actual prediction um so some sometimes it's why and sometimes it's actually named what it is uh but
that's just what it is here so we'll close that off and so we'll choose the diabetes set and it will be data set
one and so we'll worry about feedback later so we'll click on Sample uh diabetes we'll hit next and here it's
going to try to figure out uh what kind of model that we want we have to create a new experiment it's a container to run
the model in so we'll just say diabetes uh my diabetes it sounds a bit odd but that's what it is the target
column we want to predict um is the train to predict is the Y It's usually the Y um we don't have a compute cluster
so I'll go ahead and create a new compute we have dedicator or low priority technically we um it is low
priority but I just want this done low priority but don't G to compute nodes your job may be pre- emptied um I'm
going to stick with dedicated for the time being we're going to stick with CPU uh if we go with um this it does
take about an hour to run so when I ran this took about an hour so if you don't mind it's only going to cost you 15
cents but if you want this done a lot sooner I'm going to try to do something a little bit more powerful so I'm just
trying to decide here because if it only takes an hour uh I might run it on something more
powerful that's 90 cents that might be Overkill because it's not really deep learning uh it's just statistical
statistical stuff so try and large data set I wouldn't say it's large real time inference other latency sensitive
ones um how about why is this one I'm just looking
here because this one's 29 this one's more expensive but it has 32 GB of RAM this one is 28 oh 14 GB of RAM oh it's
storage so this one's our highest in the tier again you can choose this one you you just have to wait a a lot longer I
just want to see if it finishes a lot faster okay without having to go to the GPU level because I don't think GPU is
going to help too much here um the computer name is uh my diabetes machine minimum number nodes uh you want
to provision if you want dedicated nodes to set the count here uh maximum I guess I just want one node
right uh we will go ahead and oops uh complete name be2 characters long what doesn't is it too long okay
there we go we'll give it a moment here yeah it's going to spin up the
cluster so it does take a little bit time to start this so I'll see you back here when this is done
okay great so after a short little wait there it looks like uh our cluster is running if we double check it here we
can go to compute I believe that shows up under here under the compute cluster so there it is notice it's slightly
different this one shows you applications and this one is just size and Etc we can click in here see nodes
and run times we'll go make our way back here uh and we'll go ahead and hit next and notice that I think it actually will
select what it generally because it'll look at your prediction value maybe sample a bit of it and say okay you
probably want a regression thing so to predict a continuous numeric values so the thing is that if it was a label like
text or if it was just zero and one it probably would choose classification because it's um you saw our our y value
was like a number that was all over the place it thinks it's regression so I think that's a good indicator uh uh
there so let's go with regression you know but you might want it as a binary classifier but uh yeah
it's another story there so it's uh as soon as we created it just started it didn't give us the option to say hey I
want to start running it notice on this here it's going to do featurization so that means it's automatically going to
select out features for us which is what we wanted to do it set up to do regression uh we have some configuration
here so training time is 3 hours doesn't mean it's going to train for three hours but that's I guess it's timeout for it
um you could set a metric uh score threshold so it has to meet at least this to be successful if it's not going
to do it it probably would quit out early cross number Val or cross validations just make sure the data is
good you can see blocked algorithm so tensor flow DNN tensor flow L regression if it was using NN so deep learning
neural network I probably would have chose the GPU to see if it would go faster um look at the primary metric
it's normalized root square uh root mean Square AED sometimes on the exam they'll actually ask you like what's the prim
metric for this thing so it's good to uh take a look and see what they actually use for that I'll probably be sure to um
highlight that stuff in the actual lecture content um but this will take some time to run uh we have data guard
rails it will actually not populate I guess until We've ran it so so we'll just let it run and I'll see you back
here when it's done okay all right so after a very very very long wait our automl job is done it took 60 minutes so
using a larger instance didn't save me any time I don't know if maybe if I ran a GPU instance it would be a lot faster
I'd be very curious to try that out but not something for uh uh this certification course so we go into here
and yeah the cheaper instance was the same amount of time so it probably just needs gpus it really depends on the type
of models it's running so we have a bunch of different algorithms in here it ran uh about 42 different models I
thought last time I ran it I saw a lot more but you can see there's all kinds of models that it's running and then
it's going to choose the top candidate so it chose voting Ensemble so Ensemble is um uh we don't cover really in the
course because it's gets too much into ml but Ensemble is when you actually use two different weaker models and combine
the results in order to make a more uh uh powerful uh ml model okay um so here we'll get some explanation I tried this
before and I didn't get really good information so if we go here uh so like I don't have anything
under model performance so this tab requires array of predicted values from the model to be supplied we didn't
Supply any so we don't get any data Explorer so select a cohort of the data that all the data is is we have here um
so like here we were seeing age and I guess it's just giving us an indicator about the age information um
use the slider to show descending feature important select up to three cohorts to see the feature important SL
by side okay so I guess S5 and BM I don't know
what S5 is we'd have to look up the data set be BMI is your body mass index so that's a clear indicator as to what
affects whether you have diabetes or not so that makes sense age doesn't seem to be a huge factor which is kind of
interesting individual feature importance we can go here and just kind of like narrow in and say okay well why
is this outlier over here and they're like age 79 right so that's kind of interesting to see that information so
it does give you some uh explanation as to to you know why things are why they are um over here we have a little bit
more different data this is kind of interesting model performance uh I don't know what I'm
looking at but like here it's over mean squared so it's that uh mean squared calculation there again
okay so yeah it's something right uh but anyway the point is is that uh that we finally get metric so I guess we always
had to click there because that makes more sense um so yeah there's more values here sure data
transformation uh illustrates the data processing feature engine scaling techniques and machine learning
algorithm automl so you know if you were a real data scientist all this stuff would make sense to you um I think just
with time it'll it'll make sense but even at this point I I'm not sure and I don't care about the model right if
you're building something for real I'm sure uh the information becomes a lot more valuable so this model is done uh
and the idea is that we can deploy oops if we go back to the actual uh
models oh because we actually went into them e so we go back to the um autom ml here I think you can deploy any model
that you like so I think you go here and deploy this like if you prefer a different model you could deploy it um
if we go into Data guard rails we kind of skipped over that this is a way it does automatic featurization so it's
extracting up the feature so it how it handles the splitting how it handles missing features high card anality is
like if you have too much data it might have to do dimensionality reduction so that's just saying like hey if this is a
problem maybe we would do some pre-processing or stuff to make it easier to work with the data so if we're
happy with this we can go ahead and deploy it so let's say um deploy just say infer my
diabetes here we have AKs and E uh um Azure container instance let's do Azure kubernetes uh kubernetes service
cuz we did the other one here um say uh diabetes prodad maybe um AKs
diabetes oh compute name sorry um one of the inference ones okay so in order to uh deploy this we would have to
create our pipeline I'm not sure if I have enough in my quota here but let's go give it a go so I think what it's
wanting is one of these here uh I I think we'd want this wherever we are right I'm not
sure where we are If This Is Us East or uh West here let's go check
studio um Azure machine learning East
usest no I never did this when I was um I just use usually Azure container instance but I'm just curious
here say next my uh diabetes
prod we will we need to choose some nodes uh the number of nodes multiply by
the virtual machine's number of cors must be greater or equal to 12 okay no again if you're not confident
like if you're concerned about cost you can just again watch you don't have to do right um this is again a uh
fundamental certification it's not super important to get all the hands-on experience
yourself um but I'm just trying to explore this so we can see right because I I don't care about costs it's not a
big deal to me on my machine here uh so probably I don't have Sy pool must use a VM SKU with more
than two cores and four gigabytes well what did I choose did I not choose the right
one uh we'll try this again oh I chose three yeah that's
fair um uh what did it want 12 cores said before I
think invalid parameters more details because it already exists based
on that name a to it's given us all this trouble a this one we'll go ahead and delete you think
like it wouldn't matter like I wouldn't have to delete it out but that's fine this one failed now what's the
problem quota exceeded so I can't do it because I don't I'd have to go make a support request in reset so it's not a
real big deal um I guess what we could do is instead of doing it on AKs we could just deploy to container instance
if it'll let us um notice I don't have to fill anything additional in it'll just deploy I
think great uh and so I guess we'll let that deploy and I'll see you back here in a bit okay all right so I'm back here
checking on out on my or checking up on my automl here so we go over to Compu cute we go to inference clusters we
don't have anything under there if we go uh over to our experiments under our diabetes
here because we did choose to deploy the model right we clicked deploy so it should have created an ACI
instance let's make our way over to the portal the reason why it might not be sh up is because I'm just running out of
compute because again it's a quota thing um it's not a big deal for us to get a deploy it's not like we're going to do
anything with it but uh yeah so we can see that we have a container over here and it's
running so we must be able to uh see if we go to endpoints here ah here it is right I was under models that's my
problem uh so pipeline endpoints that would be something I I think that if we had deployed our designer I thought we
would have saw it under there but here we have our binary pipeline or our diabetes prod pipeline so if we wanted
to like test data you know we could pass stuff in here um I think if we wanted to kind of just like see this in action I'm
not sure if it's going to work but we'll give it a go so if we go into our sample diabetes data set and we just explore
some of the data we should be able to kind of Select out some values because I I don't know what these values mean so
let's just say like 36 oops 36 but we already know that BMI is the major factor here uh sex is
either one or two so we'll say two BMI will say 25.3 the BP will be
83 or whatever oops 83 here S1 160 S2 can be
99.6 uh three 45 45 and five [Music]
5.1 oh we only we're running out of metrics here uh 82 wonder why it doesn't give us all of
them oh I guess it does it's up to six okay so let's go ahead and test that see what we get and we got a result back
168 so uh that is uh autom ml all complete there for you um yeah so there you
[Music] go all right so let's take a look here at the uh visual designer because it's a
great way to get started very easily uh with uh if you don't know what you're doing and you want something a
little bit more advanced than automl and have some customization it's great to start with one of these samples let's go
ahead and expand it and see what we have here we have binary classification with custom python script uh TB parameters
for binary classification uh multiclass multi class classification so letter recognition
text classification all sorts of things usually binary classification classification is pretty easy I'm
looking for one that is pretty darn simple uh let's go take a look here so this says this sample shows how to
filter base features selection to selection features um binary classification so how
to predictors related to customer relationships using binary classes how to handle imbalance data sets using smot
and modules I'm not really worried about balancing uh customized python script to perform cost sensitive binary
classification tune parameters so you tune model parameters best models during the training process let's go with this
one this one seems okay to me um and so what you can see here is that it's using a sample data set I believe I think this
is a sample and if you wanted to see all of them you could literally drag them out here and do things with them uh I
haven't actually uh built one uh end to end yet for uh for this again I don't think it's like super important for uh
this level exam but uh this just shows you that there's a pre-built one if you've start to get the handle of ml you
know the full pipeline this isn't too confusing so at the beginning here we have our classification data and then
what it's going to do is say select columns in the data set so it says exclude column names work class
occupation native country so it's doing some pre-processing excluding that data might be interesting to go look at that
data set so if we go over to our data sets tab it should show up here I believe maybe because we haven't um uh
uh committed or submitted this we we can't see that data set yet but we'll look at it for a moment then we want to
clean our data so here it's saying clean all the columns so uh custom substitution
value see if we can see what it's substituting out uh it's not saying what so clean
missing data so I'm not sure what it's cleaning out there
but because I would suggest that it's using some kind of custom script um I'm not sure where it is but that's okay we
have split data pretty common to split your data so you would have a training and test data set uh it's usually really
good to randomize it so you want to randomize it then split it um and that's that's just so you get better results
then it has model hyperparameter tuning so the idea is that it's going to use ml to figure out the uh the best um
parameters for tuning over here we have the two class decision tree where it's going to do some work there it's going
to score our model and then it's going to evaluate our model and see if it's successful so this is all set up to go
so all we got to do is go to the top here there's a setting wheel here and we need to choose some type of compute so
I'm going to go here and we have this one here but I'm going to go create it's for my um my diabetes one but I'm going
to go ahead and make a new one and we're going to say um uh uh we recommend using a predefined
configuration to quickly set up compute training this one looks okay I don't know if it needs
two nodes but uh I guess we can do this one so we'll just say binary we'll just say binary
pipeline okay say save hopefully it's making a good suggestion and we'll have to wait
for that to spin up it's going to take a little bit of time okay so I'll see you back here in a moment all right so I got
message saying that that is ready so what we can do I think it was here my notebook instance no that's not it but I
I definitely saw a popup on my screen uh uh you might have saw it too you'd have to be paying close attention for that
but if you go over um it says that it's it's ready to go so what I'm going to do is make my way back over here we're
going to select our compute there is our binary pipeline I'm going to select that and there are some other options we're
not going to fill around with that we're going to go ahead and hit submit so we need a new experiment so I'm going to
just say um binary pipeline we'll hit submit okay and so this is now running
so after a little while here we're going to start seeing these go green so this is not started we'll give it a moment
here just so we can see some kind of animation and there it goes it's Off to the Races there's not much to do here
this is going to take a while I don't know I've have never ran this one in particular so I don't know if it's an
hour or 30 minutes so I'll see you back when it's done running U but yeah it's it's not that fun to watch but it's cool
that you get a visual uh illustration a so I'll see you back in a bit I just wanted to peek in here and take a look
at how it's progressing here and you can see it's still going and it's just uh cleaning the data it's still not done um
I'm not sure how long this has been running for if we go over to our experiments and we go into our I think
it's binary Pipeline and we look at the run time we're about 8 minutes in and it hasn't done a whole lot so it's still
cleaning the data I would have thought it be a little bit faster I'm kind of used to using like AWS and it goes um
Sage makers uh this doesn't usually take this long um but I mean it's nice that it's it's going here but uh yeah so
we're almost out of the pre-processing phase we'll be on to the uh the model tuning
okay all right so after waiting a little while looks like our pipeline is done uh so if we make our way over to
experiments and go to Binary pipeline we can see that it took 14 minutes and 22 seconds
uh we can go here and just see some uh additional information there's nothing really else to see we saw all the steps
already ran so you can see them all here uh okay and so let's say we wanted to there's nothing under metrics but um
enable metrics log data points compare these did within across runs we only did a single run so there's nothing to
compare so let's say we we're happy with this and we want to deploy this model well what I'm going to do is go back to
the designer uh click back here and so now in the top right corner we can create our inference pipeline so um I
can't remember if submits going to run it I don't want to run it again um I just want to go ahead and create
ourselves a realtime or batch pipeline we's say real time pipeline here and what this will do is it will
actually create a completely different pipeline so here's a completely new one uh but it's specifically designed to do
uh deployment okay so this is now one was for training the model this one is actually for uh uh taking in data and
doing inference okay so what we can do is we can go ahead and uh just submit this and so we'll put this under our
binary pipeline here we'll go ahead and hit submit and I believe that we need a
different kind of compute here I'm surprised that it's even running um no I guess it has a compute
there so it's going to run and once it uh finishes running then I believe that we we can go ahead head and um uh uh
deploy it okay so let's just wait for that to finish all right all right so after a little while there We've ran our
inference Pipeline and so uh it's definitely something that is ready for use the idea is that when we actually
it's going to go through this web service input to this web service output but uh not so important at this level uh
of certification let's see what it looks like to to go ahead and deploy it so we have we have the option between a
real-time endpoint and an existing endpoint uh we don't have an endpoint yet so
we'll just say uh binary pipeline okay and notice we have the option between oh it just it wants to
lowercase binary Pipeline and we have the option between Azure kubernetes service and add your
container instance um it's a lot easier to deploy I think to container instance so because and we'll be waiting forever
for kubernetes to start up so we're going to do container instance uh we have some options like SSL and things
like that not too worried about it so so we're just going to go ahead and hit deploy
okay and so that is going to go ahead and deploy that um so we'll wait for this real time inference if we go over
to our compute uh it should spin up so this is for a uh AKs so I don't know if it will
show up here I think only I've seen things under here but I think this will be for Azure kubernetes service and I
don't think we're going to see it show up under there uh however um we do not need to be running this anymore so we'll
go ahead and delete the binary pipeline because we're not uh we don't have it for any use right now and we might need
to free it up for something else okay so go ahead and delete it we don't need it and uh coming back to our
pipeline our designer here I'm just trying to see where we can keep track of it
um well I know it it's deploying SO waiting for Real Time endpoint so I'll see you back here when this is done okay
just takes a little bit of time all right so I think our pipeline is done if we make our way over to endpoints there
it is the binary pipeline if we wanted to go ahead there we could test the data um and so it actually already has
some pre-loaded data for us if we hit test it's nice that it fills it in E uh we get some results back okay so I
mean and then we see like scored label and income and score probability so things like that uh that is um useful so
it's giving back all all the results but I don't think it has yeah it doesn't have scored labels and scored
probabilities which is the value we want come to come back here so there are end points and that is the end of um our
Exploration with designer [Music] okay all right so let's take a look at
what it would be to actually train a job programmatically uh through the notebook so remember we saw these samples over
here and so we saw this image classification mnist and this is a very popular data set for doing uh computer
vision and these are really great if you want to really learn you should really go through these and just um uh uh read
through them because they're they're probably very very useful uh I've done a lot of this before so for me it's it's
just it's not too hard to figure out but I've actually never ran this one so let's run it together again we want to
be in uh Jupiter lab so you can go here and click it there or go to the compute if it's being a bit finicky and just
here we'll get a tab open here and we'll see how this goes so what I want to do and uh is just make sure we're back here
I'm going to click into this one and uh we have a few so there's part one and then we have the deploy stage so
let's look at training I don't know if we really need to deploy but we'll give it a read here so in this tutorial you
train ml model on a computer resource resources you'll be training and uh training and deployment workflow via the
Azure machine learning service in a notebook there's two parts to this this is using the mnus data set and scikit
learn and with Azure machine learning proba the SDK it's a popular data set with 70,000 grayscale images each image
is handwritten digits of 28 times by 28 times pixels representing numbers from 0 to 9 the goal is to create multiclass
classifier to define the digits in a given image that represents so we're going to learn a few things here but
let's just jump into it uh so the first thing is that we need to import our packages so here uh it does that map PL
plot lib inlines just make sure that when we print things that we visually see them we're going to need numpy and
then mat plod lib itself the Azure ml core uh and then we're going to import a workspace since we'll need one there and
uh then I guess it just checks the version making sure if we have the right version here okay so this is 1.28 z it's
pretty common even this an AWS they'll have like a script in here to update it in case it is out of date I'm surprised
it didn't include it in here but that's okay we'll scroll on down and by the way we're using python 3.6 Azure ml uh if
this is the future you know they might retire the old one you're using 3.8 but you know it should generally work if
it's in their sample data set I assume they try to maintain that okay so connect to a workspace so create a
workspace object from an existing workspace uh reads the file config.js so what we'll do is go run that I assume
it's kind of like a session and so here it says it's fig found our our workplace so really it's just it's not
creating a workspace it's just returning the existing one so that we have it as a variable here create an experiment so uh
that's pretty clear we saw experiments in the automl and the designer uh so we'll just hit run
there okay so we named it cor ML and we said experiment I wonder if it actually created one yet let's go over to
experiment and see if it's there so it is there cool that was fast I thought it would like print something out but it
didn't do anything there uh so creator attach an existing compute resource by using Azure machine compute a manage
service data scientist etc etc yada yada yada so create a a compute uh uh creation of a compute takes about five
minutes so let's see what it's trying to create so we have some environment variables that it wants to load in I'm
not sure how these are getting in here um I'm not sure where environment variables are set in
um uh Jupiter or even how they get feeded in but apparently they're somewhere but we have it doesn't matter
because these are defaulting so here it says CPU cluster uh zero and four it's going to
use a standard D2 V2 that is the cheapest one that we can run um I kind of want something a little bit more
powerful just for myself uh just cuz I want this to be done a lot sooner but again you know if you're don't have a
lot of money just stick with what's there okay so and this is CPU cluster so if we go here I just want to see what
her options are um
not sure why it's not showing us options here you don't have enough quota for the following VM sizes so it probably it's
because I'm running more than one VM right now yes I've s I've hit my
quota okay so like I probably have to request for more um so I think this is the
one I'm using what's the difference here this standard
dv2 vcpus it's the same one right so request quote to increase I don't know if this
is instant or not I'd have to make a support ticket oh that's going to take too long so the thing is is that uh
because the reason is is that I'm running the autom ML and the design and the uh designer in the background here
trying to create all the workshops or the uh uh the follow alongs at the same time but what I'll do is I'll just come
back and when I'm not running one of those other ones then I will uh I'll come back here and continue on but uh
we're just here at the step we want to create a a new uh compute okay all right so I'm back and I freed up uh one of my
compute instances if I go over here now I just have uh the one uh cluster instance for my uh automl but what we'll
do here is again just read through this so this will create a CPU cluster 0 to four nodes um standard D2 V2 I guess
we'll just stick with what what is here um just reading through here look look like it tries to find the compute Target
it's going to provision it it will create the cluster call Pool for minimum numbers of nodes for specific time so
wait for completion so we'll go ahead and hit play and so that's going to go and create us a new cluster so we're
just going to have to wait a little while here for it to create about 5 minutes and I'll see you back here in a
moment all right so uh the cluster started up if we go back over here we can see that it's confirmed I don't know
why it uh was so quick but uh it went pretty quick there so we're on the next section here explore the data so
download the mnist data set display some sample images so it's just talking about it being the open data set the code
retrieves in the file data set object which is a subass of data set file data set references a single or multiple
files of any format in your data store the class provides you with the ability to download or amount files to your
computer by creating a reference to the data source location Additionally you register the data set to your workspace
for easy retrieval during training there's a bit more how-tos but we'll give it good read here so we have the
open data set mnist it's kind of nice that they have that reference there uh so we have a data folder we make the
directory we are getting the data set we download it and then we are registering it so let's go ahead and run that not
sure how fast that is shouldn't take too long as it's running we'll go over here the left hand side refresh and we'll see
if it appears um uh not as of yet there it
is go into here maybe explore the data I'm not sure how would look like because these are all images right yeah so
they're in ubite gz so they're in compressed files we're not going to be able to see within them but they're
definitely there we know they're there so that that is now registered into our our data set uh display some sample
images so load the compressed into a files into numpy then use map plot lib plot 30 random images from the data set
from above not the step requires load data function it's included in the utils pie this file is included in the sample
folder we have it over here we just double click very simple file the load data and we'll go ahead and run
that and it's pretty pretty simple here uh so load data X train X test it are we setting up
our training and testing data here it kind of looks like it because it says train and test data that's when we
usually see that kind of split um and again it's doing a random split so that sounds pretty good to me
uh let's show some randomly chosen images yeah so I guess they do set up the training data here and then down
below we're actually showing the images so here's some random images train on a remote cluster so for this task you
submit the job to run on the remote training cluster to set up earlier submit your
job um create the directory create a training script create a script for run configuration submit the job so first we
will create our directory um and notice it created this directory
over here because I guess it's going to put the training file in there and so this
will actually write to a training file this makes uh quite a bit of sense so if we click into here it should now have a
training file it'll just give it a quick read see what's going on here so a lot of times when you create these training
files you have to do and this is the same if you're using AWS like when you're creating tra like or sagemaker um
you create a train file because it's part of Frameworks it's just how the Frameworks work but you'll have uh these
arguments uh so it could be like parameters to run for training um uh and there could be a whole sorts of ones
here here they're loading in the training and testing data so it's the same stuff we saw earlier when we were
just viewing the data um here it's doing a logistic regression it's using Li uh so linear
maybe linear learning model there it's doing multiclass on that there and so what
it's going to do is fit so fit is actually performing the training and then what it's going to do is make a
prediction on the test Set uh then it's going we're going to get accuracy so we're getting kind of a score so notice
that it's using accuracy uh as a evaluation metric I suppose right and then at the end we're going to dump the
data a lot of times like you have to save the model somewhere so they're outputting the actual weights of the
neural network and all other stuff it's a plk file I don't know what that is but if you're using like tensor flow you
would use tensor flow serving at the end of this a lot of times uh Frameworks will like Pi P torch or tensor flow or
mxnet they'll have a serving layer um but uh since we're just using S kit learn which is very simple it's just
going to dump out uh that file into our outputs this is going to probably run a container so this outputs isn't going to
necessarily be on um the outputs into here it's more like the outputs of the container and um
a lot of times the container will then place this somewhere so like it'll be saved on The Container but it'll be
passed out to the register or or something like that like model registry so anyway we ran this and so that
generated the file we don't want to keep on running this multiple times I probably would just overwrite the file
so it's not a big deal here it says notice how the script gets saved in the data model so here it's saying the data
uh data folder I guess we didn't look at that so we go top here um I didn't see this is data
folder was it wasn't really paying attention to where that was guess it looks like where more so
it's loading the data in so here it saves the data outut anything written to the strory is automatically uploaded to
your workspace so I guess that's just how it works so it probably will end up in here then um so util pii reference
the training script to load the data set correctly and copy the file over so um we will run this to copy the file
over so I'm guessing did it put it into here I'm just yeah so just put it in there because when it actually uh
packages it for the container it's going to bring that file over because it's a dependency
so configure the training job so create a script run config the directory that contains the script the compute Target
the training script training file Etc sometimes like in other Frameworks they'll just call them estimators but
here it's just called a script run config so uh I'm just trying to see what it's
doing so sidekit learn is the dependency okay sure we'll just hit run okay and then down below here we
have script run config so it looks like we're passing our arguments so we're saying this is
our data folder which is apparently here we're mounting it and then we're setting regularization to
0.5 sometimes you'll pass inde dependencies in here as well I guess these are technically are our parameters
that are getting configured up here at the top right but sometimes you'll have dependencies if you're in uh including
other files here uh and I guess that's up here right so see where it says environment and then we're saying
include the Azure ml defaults and the pyit learn and stuff like that and so then it gets passed in the EnV so that
makes sense to me we haven't ran that yet because we don't see any number here submit the job to the Clusters let's go
ahead and do that so it says it returns a preparing or
running State as soon as the job is completed so it's in a starting State monitor remote run so in total the
the first run takes 10 minutes but the second run uh is uh as long as the dependencies in Azure ml firment don't
change the same images reused and hence the start here start time is much faster here's what's happening while you wait
the image creation a Docker image is is created matching the python environment specified by the azl environment
the image is built and stored in the ACR the Azure container registry associated with your workspace let's go take a look
and see if that's the case because sometimes like resources aren't visible to you so I'm just curious do we
actually see it okay and yep there it is okay so they
did not lie um so associated with your workspace image creation uploading takes about 5
minutes this stage happens once for each python environment since the container's cach subsequent runs during image
creation logs are stem to the Run history you can monitor the image creation Pro process using these logs
wherever those are if you if the remote cluster requires more nodes to execute the Run than currently available
additional nodes are out added automatically scaling T typically takes about five minutes and I've seen this
before where if you're in your compute here uh sometimes it'll just say like scaling because it's just not
enough so uh running into the stage the necessary Scripts and files are sent to the compute Target then the data stores
are amounted copied the entry script is run so entry script is actually the train.py file while the job is running
SD out and the files is in the logs directory or stem to the Run history you can monitor the runs progress using
these logs the dot outputs directory of the run is copied over to the Run history in
your workspace so you can access these results you can check the progress of a running job in multiple ways this
tutorial uses the Jupiter widget so looks like we can uh run this watch the progress so maybe we'll run that and so
it's actually showing us the progress that's kind of cool I really like that so it's just a little widget
showing us all the things that it's doing let's go take a look and see what we can see under experiments and our run
pipeline because it was talking about things like outputs and things like that so over here in the outputs and logs I'm
just curious is if this is the same thing
I'm not sure if this uh is this Tails yeah it does tail it just moves so we can actually monitor it from here I
guess that's what it was talking about um so here we can see that it's setting up Docker it's actually building
a Docker image and then I'm not sure did it send it to I
mean it's on ACR already I think it looks like it's just still uh downloading extracting packages so maybe
it's actually running on the image now so we'll just wait there we pop back over here you know we can see probably
the same information is it identical yeah it is so we're 3 minutes in uh it's
probably not that fun to to watch it in real time and and talk about it so let's just wait until it's done I'll see you
back then okay all right so I'm uh about 17 minutes in here I'm not seeing any more uh movement here so it could be
that it is done it does say if you run this next step here it will wait for completion um
specify show output to true for verbose log so here actually did output a moment ago so maybe it actually was done um but
I just ran it twice so I'm not sure if that's going to cause me uh issues there so because I can't run the next step
unless I stop this um can I individually cancel this one here uh I think I can just
hit interrupt the kernel there there we go okay so I think that it's done okay because it's 18 minutes in and I don't
see any more logging in here it's just not very clear and also uh the logs we just have a lot of stuff going on here
like it's just so much so you know if we're keeping keeping Pace we probably would have saw all these created yeah so
another we just had a few more outputs there but uh I think that it's done okay it's just there's nothing
definitively saying like done do you know what I'm saying and then up here it doesn't say oh oh I guess it
does say that it's done all right so yeah I just never ran it with this tool so I just don't know so I guess it does
definitively say that I already ran this so we don't need to run that again I just feel like we'll get stuck there so
let's take a look at the metrics so regularization rate is 0.5 accuracy is nine so N9 is pretty good
the last step is training the script wrote in the output uh uh s SK learn I want to see if it's actually in our
environment here I don't think it is so outputs is somewhere it's in our workspace
somewhere but it's just not uh we just don't oh it's right here okay so it outputed the actual model right there um
and so you can see the associated files that are ran okay we'll run
it register the work model in space so you can work with other collaborators sure so if I click on that here and we
go back over to our models it is now registered over here okay and so we're done part one I don't
want to do all these other parts um training is enough as it is but let's just take a look at the deploy stage
okay so for prerequisites uh we're setting up a workspace we have our we are loading our
registered model okay we register it we have to import packages we are going to
um create scoring script deploy to an ACI model test the model if you want to do this you can go
through all the steps it does talk about a confusion Matrix and that is something that can show up on the exam is actually
talking about a confusion Matrix but we do cover that in lecture content so you generally understand what that is but um
you know I'm just I'm too tired I don't want to run through all this there's not a whole lot of value other than reading
reading through it yourself here um so I think we're all done here [Music]
okay okay one service we uh forgot to check out was Data labeling so let's go over there and give that a go so I'm
going to go ahead and create ourselves a new project I say my labeling project and we can say whether we want to
classify images or text um we have multiclass multi label bounding box uh segmentation let's go with multic
class I'll go back here for a second um multic class whoops I I don't know if we uh create
create data set but we could probably upload some local files uh let's say uh my St Trek Data
set it doesn't let us choose the image file type here be nice if these were images going to tell us what
here it's very finicky this input here uh file dis set references a single or multiple files in your public data store
or private public L okay so we'll go next uh if we can upload files directly that'd be nice oo upload a folder I like
that so what we'll do um is we do have some images in the free uh AI here under cognitive Services
assets uh we have um we'll go back here and we'll say I think objects would be the
easiest oh but we just want a folder right so yeah we'll just take objects yep we'll upload the 17
files uh yep we'll just let it stick to that path that seems fine to me we will go ahead and create
it and so now we have a data set there we'll go ahead and select that data set we'll say next your data set is
periodically checked for new data points any data points will be added as tasks it doesn't matter we're only doing this
for test uh enter the list of labels so we have um uh TNG DS9 uh
Voyager toss that's the types of Star Trek episodes um label
which um Star Trek series the image is from say
next I don't want enabled but you can have Auto uh enabled assistant labeler I'm going to say no we'll create the
project okay and I'll just wait for that to create and I'll see you back here in a moment okay all right so I'm back here
actually didn't have to wait long I think it instantly runs I just assumed like I was waiting for a state that says
completed but it's not something we have to do so uh we have 0 out of 17 progress we're going to go in here we're going to
go label some data we can view the instructions it's not showing up here but that's fine if we go to tasks we can
start labeling so what season is from or Series this is Voyager we'll hit submit this is Voyager we'll hit submit this is
toss we'll hit submit this is TNG this is TNG this is
DS9 DS9 Voyager
Voyager uh TNG DS9 you get the idea though and you got
some options here like change the contrast if someone can't see the photo or rotate it this is
Voyager Voyager uh TNG
DS9 uh Voyager Voyager and we're done so we'll go back
to our labeling job here we'll see we have the breakdown there uh and now our data set is
labeled we can export our data set CSV Coco as your ml data set I believe that means it'll go back into the data sets
over here which will make our Liv a little bit easier we go back to data labeling okay so you just granted people
access to the studio they'd be able to just go in here and and jump into that job okay uh if we go over to the data
set I believe we should have a labeled version of it now so my labeling project so I believe that is the uh the labeled
stuff here right yeah so it's labeled so there you go we're all done aure machine learning
uh and so all that's left is to do some cleanup [Music]
okay so we're all done with Azure machine learning uh if we want to we can go to our compute and just uh kill the
services we have here now if we go to the resource Group and delete everything it'll it'll take all these things down
anyway but I'm just going to go a bit paranoid so I'm going to just manually do this
okay hit delete okay and so we'll go back to portal. azure.com
and uh I'm going to go to my resource groups and everything is contained it should be all contained within my studio
just be sure to check these other ones for that and we can see all the stuff that we spun up we'll go ahead and hit
delete Resource Group um I don't know if it includes like because I don't see like container registry right so I know
like it puts stuff there I guess it does it says container registry so that's pretty much
everything right and that'll take down everything so and if you're paranoid all you can do is go
to all resources and double check over here because if there's anything running it will show up here okay um but that's
pretty much it and so just delete and we're all done
The AI-900 exam covers five main domains: AI workloads and considerations (15-20%), fundamental machine learning principles on Azure (20-25%), computer vision features (15-20%), natural language processing (NLP) features (15-20%), and generative AI workloads (15-20%). It focuses on understanding capabilities and concepts rather than complex problem-solving.
Beginners should plan for about 20-30 hours of study, intermediates with some Azure or ML experience around 8-10 hours, and experienced cloud professionals roughly 5 hours. A recommended study plan is 30 to 60 minutes daily over two weeks, combining theoretical lectures with hands-on labs for better retention.
Key Azure AI services include Azure Cognitive Services such as Computer Vision, Form Recognizer, Text Analytics, Speech Services, Q&A Maker, and Language Understanding Intelligent Service (LUIS). Familiarity with Azure Machine Learning Studio for model building and deployment, Automated ML, and Azure OpenAI Service for generative AI is also important.
The course explains Artificial Intelligence as systems mimicking human cognition, Machine Learning as algorithms that improve from data without explicit programming, and Deep Learning as neural networks with multiple hidden layers designed for complex data processing, providing foundational knowledge for Azure AI applications.
It covers Microsoft's six Responsible AI principles: Fairness, Reliability & Safety, Privacy & Security, Inclusiveness, Transparency, and Accountability, along with practical examples and frameworks that guide ethical AI development, ensuring learners understand best practices for responsible AI usage.
The course includes labs and exercises such as setting up compute resources and notebooks, using Azure Machine Learning Studio for training and deploying models, working with Azure cognitive services SDKs, creating custom vision projects, and integrating Q&A Maker and Azure Bot Service for conversational AI applications.
Focus on both conceptual understanding and practical skills by using lectures, hands-on labs, and practice exams. Manage your 60 minutes effectively, expect diverse question types like multiple choice and drag-and-drop, and remember there's no penalty for wrong answers, so attempt all questions. Prioritize learning service capabilities and responsible AI concepts for success.
Heads up!
This summary and transcript were automatically generated using AI with the Free YouTube Transcript Summary Tool by LunaNotes.
Generate a summary for freeRelated Summaries
Complete Microsoft Azure Developer Associate (AZ-204) Study Guide
This comprehensive guide covers the Microsoft Azure Developer Associate AZ-204 certification exam, providing insights into Azure functions, storage options, app services, database management, security, monitoring, and API management. It includes practical examples, key concepts, and hands-on tips to help you prepare effectively and achieve certification success.
Ultimate Guide to Azure DevOps Certification Course: Pass the Exam with Confidence
Join Andrew Brown's free DevOps certification course and learn everything you need to know to pass the Azure A400 certification exam!
Unlocking Azure DevOps with the A400 Certification: A Comprehensive Guide
Master Azure DevOps with our A400 Certification guide. Learn key topics, strategies, and tips for success.
Comprehensive AWS and Azure Cloud Computing Course Summary and Key Projects
Explore an extensive cloud computing course covering AWS and Azure platforms, services, security, and architecture. Learn through practical projects and interview preparations to build a strong foundation and advance your cloud career.
A Step-by-Step Roadmap to Mastering AI: From Beginner to Confident User
This video provides a comprehensive roadmap for anyone looking to start their AI journey, emphasizing the importance of understanding core concepts before diving into tools. It offers practical tips on building an AI learning system, developing critical thinking skills, and strategically selecting AI tools to enhance productivity.
Most Viewed Summaries
Kolonyalismo at Imperyalismo: Ang Kasaysayan ng Pagsakop sa Pilipinas
Tuklasin ang kasaysayan ng kolonyalismo at imperyalismo sa Pilipinas sa pamamagitan ni Ferdinand Magellan.
A Comprehensive Guide to Using Stable Diffusion Forge UI
Explore the Stable Diffusion Forge UI, customizable settings, models, and more to enhance your image generation experience.
Mastering Inpainting with Stable Diffusion: Fix Mistakes and Enhance Your Images
Learn to fix mistakes and enhance images with Stable Diffusion's inpainting features effectively.
Pamamaraan at Patakarang Kolonyal ng mga Espanyol sa Pilipinas
Tuklasin ang mga pamamaraan at patakaran ng mga Espanyol sa Pilipinas, at ang epekto nito sa mga Pilipino.
Pamaraan at Patakarang Kolonyal ng mga Espanyol sa Pilipinas
Tuklasin ang mga pamamaraan at patakarang kolonyal ng mga Espanyol sa Pilipinas at ang mga epekto nito sa mga Pilipino.

