great so I think we can start since the
meeting is recorded So if everyone uh
jump jumps in later can can watch the
recording
so hi everyone and welcome to these
um Cloud skill challenge study session
around a create classification models
with Azure machine learning designer
so today I'm thrilled to be here with
John uh John do my introduce briefly
yourself
uh thank you Carlotta hello everyone
Welcome to our Workshop today I hope
that you are all excited for it I am
John Aziz a gold Microsoft learn student
ambassador and I will be here with uh
Carlota to like do the Practical part
about this module of the cloud skills
challenge thank you for having me
perfect thanks John so for those who
don't know me I'm kellota
based in Italy and focus it on AI
machine learning Technologies and about
the use in education
um so
um these Cloud skill challenge study
session is based on a learn module a
dedicated learn module I sent to you uh
the link to this module uh in the chat
in a way that you can follow along the
model if you want or just have a look at
the module later at your own pace
um
so before starting I would also like to
remember to remember you uh the code of
conduct and guidelines of our student
Masters community so please during this
meeting be respectful and inclusive and
be friendly open and with coming and
respectful of other each other
differences
if you want to learn more about the code
of content you can use this link in the
deck again.ms slash s-a-c-o-c
and now we are
um we are ready to to start our session
so as we mentioned it we are going to
focus on classification models and Azure
ml uh today so first of all we are going
to um identify uh the kind of
um of scenarios in which you should
choose to use a classification model
we're going to introduce Azure machine
learning and Azure machine designer
we're going to understand uh which are
the steps to follow to create a
classification model in Azure mesh
learning and then John will
um
lead an amazing demo about training and
Publishing a classification model in
Azure ml designer
so let's start from the beginning let's
start from identifying classification
machine learning scenarios
so first of all what is classification
classification is a form of machine
learning that is used to predict which
category or class an item belongs to for
example we might want to develop a
classifier able to identify if an
Incoming Email should be filtered or not
according to the style the center the
length of the email Etc in this case the
characteristics of the email are the
features
and the label is a classification of
either a zero or one representing a Spam
or non-spam for the including email so
this is an example of a binary
classifier if you want to assign
multiple categories to the incoming
email like work letters love letters
complaints or other categories in this
case a binary classifier is not longer
enough and we should develop a
multi-class classifier so classification
is an example of what is called
supervised machine learning
in which you train a model using data
that includes both the features and
known values for label
so that the model learns to fit the
feature combinations to the label then
after training has been completed you
can use the train model to predict
labels for new items for for which the
label is unknown
but let's see some examples of scenarios
for classification machine learning
models so we already mentioned an
example of a solution in which we would
need a classifier but let's explore
other scenarios for classification in
other Industries for example you can use
a classification model for a health
clinic scenario and use clinical data to
predict whether patient will become sick
or not
uh you can use
um
oh sorry so when I became muted it's a
long time or you can use you can use uh
some models for classification for
example you can use you're saying this
uh so I I was I was
this one like you you have been muted
for uh one second okay okay perfect
perfect uh yeah I was talking sorry for
that so I was talking about the possible
you can use a classification model like
have Clinic scenario Financial scenario
or other third one is business type of
scenario you can use characteristics or
small business to predict if a new
Venture will will succeed or not for
example and these are all types of
binary classification
uh but today we are also going to talk
about Azure machine learning so let's
see
um what is azure Mash learning so
training and deploying an effective
machine learning model involves a lot of
work much of it time consuming and
resource intensive so Azure machine
learning is a cloud-based service that
helps simplify some of the tasks it
takes to prepare data train a model and
also deploy it as a predictive service
so it helps that the scientists increase
their efficiency by automating many of
the time consuming tasks Associated to
creating and training a model
and it enables them also to use
cloud-based compute resources that scale
effectively to handle large volumes of
data while incurring costs only when
actually used
to use Azure machine learning you
fasting fast you need to create a
workspace resource in your Azure
subscription and you can then use these
workspace to manage data compute
resources code models and other
artifacts after you have created an
Azure machine learning workspace you can
develop Solutions with the Azure machine
learning service either with developer
tools or the Azure machine Learning
Studio web portal
in particular International Learning
Studio is a web portal for machine
learning Solutions in Azure and it
includes a wide range of features and
capabilities that help data scientists
prepare data train models publish
Predictive Services and monitor also
their usage
so to begin using the web portal you
need to assign the workspace you created
in the Azure portal to the Azure machine
Learning Studio
as its core Azure Mash learning is a
service for training and managing
machine learning models for which you
need compute resources on which to run
the training process
compute targets are um one of the main
basic concept of azure Mash learning
they are cloud-based resources on which
you can run model training and AD
exploration processes
so initial machine Learning Studio you
can manage the compute targets for your
data science activities and there are
four kinds of of compute targets you can
create we have the computer instances
which are vital machine set up for
running machine learning code during
development so they are not designed for
production
then we have compute clusters which are
a set of virtual machines that can scale
up automatically based on traffic
we have inference clusters which are
similar to compute clusters but they are
designed for deployment so they are a
deployment targets for Predictive
Services that use train models
and finally we have attached compute
which are any compute Target that you
manage yourself outside of azramel like
for example virtual machines or Azure
databricks clusters
so we talked about Azure machine
learning but we also mentioned it
mentioned Azure machine learning
designer what is azure machine learning
designer so initial machine Learning
Studio there are several ways to author
classification machine learning models
one way is to use a visual interface and
this visual interface is called designer
and you can use it to train test and
also deploy machine learning models and
the drag and drop interface makes use of
clearly defined inputs and outputs that
can be shared reused and also Version
Control
and using the designer you can identify
the building blocks or components needed
for your model place and connect them on
your canvas and run a machine learning
job
so
um each designer project so each project
in the designer is known as a pipeline
and in the design we have a left panel
for navigation and a canvas on your
right hand side in which you build your
pipeline visually so pipelines let you
organize manage and reuse complex
machine learning workflows across
projects and users
a pipeline starts with the data set from
which you want to train the model
because all begins with data when
talking about data science and machine
learning and each time you run a
pipeline the configuration of the
pipeline and its results are stored in
your workspace as a pipeline job
so the second main concept of azure
machine learning is a component so going
hierarchically from the pipeline we can
say that each building block of a
pipeline is called a component
learning component encapsulate one step
in a machine learning pipeline so it's a
reusable piece of code with inputs and
outputs something very similar to a
function in any programming language
and in a pipeline project you can access
data assets and components from the left
panels
asset Library tab as you can see
um here in the screenshot in the deck
so you can create data assets on using
an adoc page called Data Page and a data
set is a reference to a data source
location
so this data source location could be a
local file a data store a web file or
even an age group open that set
and these data assets will appear along
with standard sample data set in the
designers asset Library
um
and another basic concept of azure ml is
azure machine learning jobs
so basically when you submit a pipeline
you create a job which will run all the
steps in your pipeline so a job execute
a task against a specified compute
Target
jobs enable systematic tracking for your
machine learning experimentation in
Azure ml
and once a job is created azramel
maintains a run record uh for for the
job
um but let's move to the classification
steps so
um let's introduce uh how to create a
classification model in Azure ml but you
will see it in more details in a
handsome demo that John will will guide
through in a few minutes
so you can think of the steps to train
and evaluate a classification machine
learning model as four main steps so
first of all you need to prepare your
data so you need to identify the
features and the label in your data set
you need to pre-process so you need to
clean and transform the data as needed
then the second step of course is
training the model
and for training the model you need to
split the data into two groups a
training and a validation set
then you train a machine learning model
using the training data set and you test
the machine learning model for
performance using the validation data
set
the third step is performance evaluation
um which means comparing how close the
model's predictions are to the known
labels and these lead us to compute some
evaluation performance metrics
and then finally
um so these three steps are not
um not performed uh every time in a
linear manner it's more an iterative
process but once you obtain you achieve
a a performance with which you are
satisfied so you are ready to let's say
go into production and you can deploy
your train model as a predictive service
into a real-time uh to a real-time
endpoint and to do so you need to
convert the training pipeline into a
real-time inference Pipeline and then
you can deploy the model as an
application on a server or device so
that others can consume this model
so let's start with the first step which
is prepaid data reward data can contain
many different issues that can affect
the utility of the data and our
interpretation of the results so also
the machine learning model that you
train using this data for example real
world data can be affected by a bed
recording or a bad measurement and it
can also contain missing values for some
parameters and Azure machine learning
designer has several pre-built
components that can be used to prepaid
data for training these components
enable you to clean data normalize
features join tables and and more
let's come to uh training so to train a
classification model you need a data set
that includes historical features so the
characteristics of the entity for which
one to make a prediction and known label
values the label is the class indicator
we want to train a model to predict it
and it's common practice to train a
model using a subset of the data while
holding back some data with which to
test the train model and this enables
you to compare the labels that the model
predicts with the actual known labels in
the original data set
this operation can be performed in the
designer using the split data component
as shown by the screenshot here in the
in the deck
there's also another component that you
should use which is the score model
component to generate the predicted
class label value using the validation
data as input so once you connect all
these components
um the component specifying the the
model we are going to use the split data
component the trained model component
and the score model component you want
to run an a new experiment in the
initial map which will use the data set
on the canvas to train and score a model
after training a model it is important
we say to evaluate its performance to
understand how bad how how good sorry
our model is performing
and there are many performance metrics
and methodologies for evaluating how
well a model makes predictions the
component to use to perform evaluation
in Azure ml designer is called as
intuitive as it is evaluate model
once the job of training and evaluation
of the model is completed you can review
evaluation metrics on the completed job
Page by right clicking on the component
in the evaluation results you can also
find the so-called confusion Matrix that
you can see here in the right side of of
this deck
a confusion Matrix shows cases where
both the predicted and actual values
were one uh the so-called true positives
at the top left and also cases where
both the predicted and the actual values
were zero the so-called true negatives
at the bottom right while the other
cells show cases where the predicting
and actual values differ
called false positive and false
negatives and this is an example of a
confusion Matrix for a binary classifier
why for a multi-class classification
model the same approach is used to
tabulate each possible combination of
actual and predictive value counts so
for example a model with three possible
classes would result in three times
three Matrix
the confusion Matrix is also useful for
the metrics that can be derived from it
like accuracy recall or precision
um we we say that the last step is
deploying the train model to a real-time
endpoint as a predictive service and in
order to automate your model into
service that makes continuous
predictions you need first of all to
create any and and then deploy an
inference pipeline the process of
converting the training pipeline into a
real-time inference pipeline removes
training components and adds web service
inputs and outputs to handle requests
and the inference pipeline performs they
seem that the transformation as the
first pipeline but for new data then it
uses the train model to infer or predict
label values based on its feature
um so I think I've talked a lot for now
I would like to let John show us
something in practice uh with uh with
the Hands-On demo so please John go
ahead sharing your screen and guide us
through this demo of creating a
classification with the Azure machine
learning designer
uh thank you so much Carlotta for this
interesting explanation of the Azure ml
designer and now
um I'm going to start with you in the
Practical demo part so uh if you want to
follow along go to the link that Carlota
sent in the chat so like you can do
the demo or the Practical part with me
I'm just going to share my screen
and
go here so uh
where am I right now I'm inside the
Microsoft learn documentation this is
the exercise part of this module and we
will start by setting two things which
are a prequisite for us to work inside
this module which are the users group
and the Azure machine learning workspace
and something extra which is the compute
cluster that calendar Target about so I
just want to make sure that you all have
a resource Group created inside your
auditor inside your Microsoft Azure
platform so this is my research group
inside this is this Resource Group I
have created an Azure machine learning
workspace so I'm just going to access
the workspace that I have created
already from this link I am going to
open it which is the studio web URL and
I will follow the steps so what is this
this is your machine learning workspace
or machine Learning Studio you can do a
lot of things here but we are going to
focus mainly on the designer and the
data and the compute so another
prequisite here as Carlotta told you and
we need some resources to power up the
the classification the processes that
will happen
so we have created this Computing
cluster
and we have like Set uh some presets for
it so
where can you find this preset you go
here under the create compute you'll
find everything that you need to do so
the size is the Standard ds11 Version 2
and it's a CPU not GPU because we don't
know the GPU and we don't need a GPU and
a like it is ready for us to use
the next thing which we will look into
is the designer how can you access the
designer
you can either click on this icon or
click on the navigation menu and click
on the designer for me
um
now I am inside my designer
what we are going to do now is the
pipeline that Carlotta told you about
and from where can I know these steps if
you follow along in the learn module you
will find everything that I'm doing
right now in details uh with screenshots
of course so I'm going to create a new
pipeline and I can do so by clicking on
this plus button
it's going to redirect me to the
designer authoring the pipeline uh where
I can drag and drop data and components
that the Carlota told you the difference
between
and here I am going to do some changes
to the settings I am going to connect
this with my compute cluster that I
created previously so I can utilize it
from here I'm going to choose this
compute cluster demo that I have showed
you before in the Clusters here
and I am going to change the name to
something more meaningful instead of
byline and the date of today I'm going
to name it diabetes
and
let's just check this training
let's say training 0.1 or okay
and I am going to close this tab and in
order to have a bigger place to work
inside because this is where we will
work where everything will happen so I
will click on close from here
and I will go to the data and I will
create a new data set
how can I create a new data set there is
multiple options here you can find from
local files from data store from web
files from open data set but I'm going
to choose from web files as this is the
way we're going to create our data
from here the information of my data set
I'm going to get them from the Microsoft
learn module so if we go to the step
that says create a data set
under it it illustrates that you can
access the data from inside the asset
library and inside your asset liability
you'll find the data and find the
component and I'm going to select
this link because this is where my data
is stored if you open this link you will
find this is this is a CSV file I think
yeah and you can like all the data are
here
all right now let's get back
um
[Music]
and you are going to do something
meaningful but because I have already
created it before twice so I'm gonna
like add a number to the name
uh the data set is tabular and there is
the file but this is a table so we're
going to choose the table
[Music]
for data set time
now we will click on next that's gonna
review or uh display for you the content
of this file that you have
like imported to this workspace
and for these settings these are like
related to our filed format
so this is a delimited file and it's not
plain text it's not a Json the delimiter
is comma as like we have seen that they
those
so I'm choosing
errors because the only the first five
[Music]
for example okay uh if you have any
doubts if you have any problems please
don't hesitate to all right through me
in the chat
and like what what is blocking you and
me and Carlota will try to help you and
like whenever possible
and now this is the new preview for my
data set I can see that I have an ID I
have patient ID I have pregnancies I
have the age of the people
have the body mass together I think
and they have diabetical or not as a
zero and one zero indicates a negative
the person doesn't have diabetes and one
indicates a positive that this person
has diabetes okay
now I'm going to click on next here I am
defining my schema all the data types
inside my columns the column names which
columns to include which to exclude and
here we will include everything except
the path of the bath color and we are
going to review the data types of each
column so let's review this first one
this is numbers numbers then it's the
integer and this is
um like decimal
dotted
decimal number so we are going to choose
this data type
and for this one
it says diabetic and it's a zero under
one and we are going to make it as
integerables
now we are going to click on next and
move to reviewing everything this is
everything that we have defined together
I will click on create
and
now the first step has ended we have
gotten our data ready
now what now we're going to utilize the
designer
um Power we're going to drag and drop
our data set to create the pipeline
so I have like click on it and drag it
to this space it's gonna appear to you
and we can inspect it by right click and
choose preview data
to see what we have created together
from here you can see everything that we
have like seen previously but in more
details and we are just going to close
this now what now we are gonna do the
processing that Carlota like mentioned
these are some instructions about the
data about how you can loot them how you
can open them but we are going to move
to the transformation or the processing
so as Carlotta told you like any data
for us to work on we have to do some
processing to it
to make it easy easier for the model to
be trained and easier to work with so uh
we're gonna do the normalization and
normalization meaning is uh
to scale our data either down or up but
we're going to scale them down
and like we are going to decrease and
relatively decrease
the the values all the values to work
with lower numbers and if we are working
with larger numbers it's going to take
more time if we're working with smaller
numbers it's going to take less time to
calculate them and that's it so
where can I find the normalized data I
can find it inside my component
so I will choose the component and
search for normalized data
I will drag and drop it as usual and I
will connect between these two things
by clicking on this spot this like
Circle and
drag and drop until the next circuit
now we are going to Define our
normalization method
so I'm going to double click on the
normalized data
it's going to open the settings for the
normalization
as better transformation method which is
a mathematical way
that is going to scale our data
according to
we're going to choose min max and for
this one we are going to choose use 0
for constant column we are going to
choose true
and we are going to Define which columns
to normalize so we are not going to
normalize the whole data set we are
going to choose a subset from the data
set to normalize so we're going to
choose everything except for the patient
ID and the diabetic because the patient
ID is a number but it's a categorical
data it describes a vision it's not a
number that I can sum I can say patient
ID number one plus patient ID number two
no this is a patient and another
location it's not a number that I can do
mathematical operations on so I'm not
going to choose it so we will choose
everything as I said except for the
diabetic and the patient might I will
click on Save
and it's not showing me a warning again
everything is good
now I can click on submit
and review my normalization output
um
so uh if you click on submit here
and you will like choose create new and
set the name that is mentioned here
inside the notebook so it it tells you
to like create a job and name it name
the experiment Ms learn diabetes
training because you will continue
working on and building component later
I have it already like created I am the
like we can review it together so uh let
me just open this in another tab I think
I have it
here
okay
so these are all the jobs that I have
read them
all the jobs there let's do this over
these are all the jobs that I have
submitted previously
and I think this one is the
normalization job so let's see the
output of it
as you can see it says uh check mark yes
which means that it worked and we can
preview it how can I do that right click
on it choose preview data
and as you can see all the data are
scaled down
so everything is between zero
and uh one I think
so like everything is good for us now we
can move forward to the next step
which is to create the whole pipeline so
uh Carlota told you that
we're going to use a classification
model to create this data set so uh let
me just drag and drop everything
to get runtime and we're doing
about about everything by
so
as a result we are going to explain
yeah so I'm going to give this split
data I'm going to take the
transformation data to split data and
connect it like that
I'm going to get a three model
components because I want to train my
model
and I'm going to put it right here
okay
like let's just move it down there okay
and we are going to use a classification
model
a two class
logistic regression model
so I'm going to give this algorithm to
enable my model to work
this is the untrained model this is
here
the left the left
the left like Circle I'm going to
connect it to the data set and the right
one we are going to connect it to
evaluate model
evaluate model so let's search for
evaluate model here
so because we want to do what we want to
evaluate our model and see how it it has
been doing it is it good is it bad
um sorry like
this is
down there
after the school
so we have to get the score model first
so let's get it
and this will take the trained model and
the data set
to score our model and see if it's
performing good or bad
and
um
after that like we have finished
everything now we are going to do the
what
the presets for everything
as a starter we will be splitting our
data so
how are we going to do this according to
what to the split rules so I'm going to
double click on and choose split rows
and the percentage is
70 percent for the and 30 percent of the
data for
the valuation or for the scoring okay
I'm going to make it a randomization so
I'm going to split data randomly and the
seat is uh
132 23 I think yeah
and I think that's it
the split says why this holes and that's
good
now for the next one which is the train
model we are going to connect it as
mentioned here
and like we have done that and then why
am I having here like let's double click
on it yeah it has like it needs the
label column that I am trying to predict
so from here I'm going to choose
diabetic I'm going to save
I'm going to close this one
so it says here
the diabetic label the model it will
predict the zero and one because this is
a binary classification algorithm so
it's going to predict either this or
that
and
um
I think that's everything to run the the
pipeline
so everything is done everything is good
for this one we're just gonna leave it
like for now because this is the next
step
um this will like be put instead of the
score model but then it's
delete it for now
okay
now we have to submit the job in order
to see the output of it so I can click
on submit and choose the previous job
which is the one that I have showed you
before
and then let's review its output
together here
so if I go to the jobs
if I go to Ms learn maybe it is training
I think it's the one that lasted the
longest this one here
so here I can see
the job output what happened inside
the model as you can see
so the normalization we have like seen
before the split data I can preview it
the result one or the result two as it
splits the data to 70 here and three
thirty percent here
um I can see the score model which is
like something that we need
to review
um inside the scroll model uh like from
here
we can see that
let's get back here
like this is the data that the model has
been scored and this is a scoring output
so it says code label true and if he is
not diabetic so this is
um
around prediction let's say
for this one it's true and true and this
is like a good like what do you say
prediction and the probabilities of this
score
which means the certainty of our model
of that this is really true it's 80 for
this one is 75
so these are some cool metrics that we
can review to understand how our model
is performing it's performing good for
now
let's check our evaluation model
so this is the extra one that I told you
about instead of the like
score model only we are going to add
what evaluate model
after it so here
we're going to go to our asset library
and we are going to choose the evaluate
model
and we are going to put it here and we
are going to connect it and we are going
to submit the job using the same name of
the job that we used previously
let's review it uh also so after it
finishes you will find it here so I have
already done it before this is how I'm
able to see the output
so let's see
what what is the output of this
evaluation process
here it mentioned to you that there are
some metrics
like the confusion Matrix which Carlotta
told you about there is the accuracy the
Precision the recall and F1 School
every Matrix gives us some insight about
our model it helps us to understand it
more more and
like understand if it's overfitting if
it's good if it's bad and really really
like understand how it's working
now I'm just waiting for the job to load
until it loads
um
we can continue to
to work on our
model so I will go to my designer I'm
just going to confirm this
and I'm going to continue working on it
from
where we have stopped where have we
stopped
we have stopped on the evaluate model so
I'm going to choose this one
and it says here
select experiment create inference
pipeline so
I am going to go to the jobs
I'm going to select my experiment
I hope this works
okay salute finally now we have our
evaluate model output
let's previews evaluation results
and uh
cool come on
finally now we can create our inference
pipeline so
I think it says that
um
select the experiment then select Ms
learn so
I am just going to select it
and finally now we can the ROC curve we
can see it that the true positive rate
and the force was integrate the false
positive rate is increasing with time
and also the true positive rate true
positive is something that it predicted
that it is uh positive it has diabetes
and it's really a it's really true it
the person really has diabetes okay and
for the false positive it predicted that
someone has diabetes and someone doesn't
has it this is what true position and
false positive means this is The Recoil
curve so we can like review the metrics
of our model this is the lift curve I
can change the threshold of my confusion
Matrix here
and this could look don't want to add
anything about the the the graphs and
you can do so
um
yeah so just wanted to if you go yeah I
just wanted to comment comment for the
RSC curve uh that actually from this
graph the metric which uh usually we're
going to compute is the end area under
under the curve and this coefficient or
metric
um it's a confusion
um is a value that could span from from
zero to one and the the highest is
um
this the highest is the the score so the
the closest one
um so the the highest is the amount of
area under this curve
um the the the highest performance uh we
we've got from from our model and
another thing is what John is
um playing with so this threshold for
the logistic
regression is the threshold used by the
model
um to
um
to predict uh if the category is zero or
one so if the probability the
probability score is above the threshold
then the category will be predicted as
one while if the the probability is
below the threshold in this case for
example 0.5 the category is predicted as
as zero so that's why it's very
important to um to choose the the
threshold because the performance really
can vary
um
with this threshold value
uh thank you uh so much uh kellota and
as I mentioned now we are going to like
create our inference pipeline so we are
going to select the latest one which I
already have it opened here this is the
one that we were reviewing together this
is where we have stopped and we're going
to create an inference pipeline we are
going to choose a real-time inference
pipeline okay
um from where I can find this here as it
says real-time inference pipeline
so it's gonna add some things to my
workspace it's going to add the web
service inboard it's going to have the
web service output because we will be
creating it as a web service to access
it from the internet
uh what are we going to do we're going
to remove this diabetes data okay
and we are going to get a component
called Web
input and what's up let me check
it's enter data manually
we have we already have the with input
present
so we are going to get the entire data
manually
and we're going to collect it to connect
it as it was connected before like that
and also I am not going to directly take
the web service sorry escort model to
the web service output like that
I'm going to delete this
and I'm going to execute a python script
before
I display my result
so
this will be connected like okay but
so
the other way around
and from here I am going to connect this
with that and there is some data uh that
we will be getting from the node or from
the the examination here and this is the
data that will be entered like to our
website manually okay this is instead of
the data that we have been getting from
our data set that we created so I'm just
going to double click on it and choose
CSV and I will choose it has headers
and I will take or copy this content and
put it there okay
so let's do it
I think I have to click on edit code now
I can click on Save and I can close it
another thing which is the python script
that we will be executing
um yeah we are going to remove this also
we don't need the evaluate model anymore
so we are going to remove
script that I will be executing okay
I can find it here
um
yeah
this is the python script that we will
execute and it says to you that this
code selects only the patient's ID
that's correct label the school
probability and return returns them to
the web service output so we don't want
to return all the columns as we have
seen previously
uh the determines everything
so
we want to return certain stuff the
stuff that we will use inside our
endpoint so I'm just going to select
everything and delete it and
paste the code that I have gotten from
the uh
the Microsoft learn Docs
now I can click on Save and I can close
this
let me check something I don't think
it's saved it's saved but the display is
wrong okay
and now I think everything is good to go
I'm just gonna double check everything
so uh yeah we are gonna change the name
of this uh
Pipeline and we are gonna call it
predict
diabetes okay
now let's close it and
I think that we are good to go so
um
okay I think everything is good for us
I just want to make sure of something is
the data is correct the data is uh yeah
it's correct
okay now I can run the pipeline let's
submit
select an existing Pipeline and we're
going to choose the MS layer and
diabetes training which is the pipeline
that we have been working on
from the beginning of this module
I don't think that this is going to take
much time so we have submitted the job
and it's running
until the job ends we are going to set
everything
and for deploying a service
in order to deploy a service okay
um
I have to have the job ready so
until it's ready or you can deploy it so
let's go to the job the job details from
here okay
and until it finishes
Carlotta do you think that we can have
the questions and then we can get back
to the job I'm deploying it
yeah yeah yeah so yeah yeah guys if you
have any questions
uh on on what you just uh just saw here
or into introductions feel free this is
a good moment we can uh we can discuss
now while we wait for this job to to
finish
uh and the
can can
we have the energy check one or like
what do you think uh yeah we can also go
to the knowledge check
um
yeah okay so let me share my screen
yeah so if you have not any questions
for us we can maybe propose some
questions to to you that you can
um
uh to check our knowledge so far and you
can uh maybe answer to these questions
uh via chat
um so we have do you see my screen can
you see my screen
yes
um so John I think I will read this
question loud and ask it to you okay so
are you ready to transfer
yes I am
so
um you're using Azure machine learning
designer to create a training pipeline
for a binary classification model so
what what we were doing in our demo
right and you have added a data set
containing features and labels uh a true
class decision Forest module so we used
a logistic regression model our
um in our example here we're using A2
class decision force model
and of course a trained model model you
plan now to use score model and evaluate
model modules to test the train model
with the subset of the data set that
wasn't used for training
but what are we missing so what's
another model you should add and we have
three options we have join data we have
split data or we have select columns in
in that set
so
um while John thinks about the answer uh
go ahead and
um
answer yourself so give us your your
guess
put in the chat or just come off mute
and announcer
a b yes
yeah what do you think is the correct
answer for this one I need something to
uh like I have to score my model and I
have to evaluate it so I I like I need
something to enable me to do these two
things
I think it's something you showed us in
in your pipeline right John
of course I did
uh we have no guests yeah
can someone like someone want to guess
uh we have a b yeah
uh maybe
so uh in order to do this in order to do
this I mentioned the
the module that is going to help me to
to divide my data into two things 70 for
the training and thirty percent for the
evaluation so what did I use I used
split data because this is what is going
to split my data randomly into training
data and validation data so the correct
answer is B and good job eek thank you
for participating
next question please
yes
answer so thanks John
uh for uh explaining us the the correct
one
and we want to go with question two
yeah so uh I'm going to ask you now
karnata you use Azure machine learning
designer to create a training pipeline
for your classification model
what must you do before you deploy this
model as a service you have to do
something before you deploy it what do
you think is the correct answer
is it a b or c
share your thoughts without in touch
with us in the chat and
um and I'm also going to give you some
like minutes to think of it before I
like tell you about
yeah so let me go through the possible
answers right so we have a uh create an
inference pipeline from the training
pipeline
uh B we have ADD and evaluate model
module to the training Pipeline and then
three we have uh clone the training
Pipeline with a different name
so what do you think is the correct
answer a b or c
uh also this time I think it's something
we mentioned both in the decks and in
the demo right
yes it is
it's something that I have done like two
like five minutes ago
it's real time real time what's
um
yeah so think about you need to deploy
uh the model as a service so uh if I'm
going to deploy model
um I cannot like evaluate the model
after deploying it right because I
cannot go into production if I'm not
sure I'm not satisfied over my model and
I'm not sure that my model is performing
well
so that's why I would go with
um
I would like exclude B from from my from
my answer
uh while
um thinking about C uh I don't see you I
didn't see you John cloning uh the
training Pipeline with a different name
uh so I I don't think this is the the
right answer
um while I've seen you creating an
inference pipeline uh yeah from the
training Pipeline and you just converted
it using uh a one-click button right
yeah that's correct so uh this is the
right answer
uh good job so I created an inference
real-time Pipeline and it has done it
like it finished it finished the job is
finished so uh we can now deploy
ment
yeah
exactly like on time
I like it finished like two seconds
three three four seconds ago
so uh
until like um
this is my job review so
uh like this is the job details that I
have already submitted it's just opening
and once it opens
um
like I don't know why it's so heavy
today it's not like that usually
yeah it's probably because you are also
showing your your screen on teams
okay so that's the bandwidth of your
connection is exactly do something here
because yeah finally
I can switch to my mobile internet if it
did it again so I will click on deploy
it's that simple I'll just click on
deploy and
I am going to deploy a new real-time
endpoint
so what I'm going to name it I'm
description and the compute type
everything is already mentioned for me
here so I'm just gonna copy and paste it
because we like we have we are running
out of time
so it's all Azure container instance
which is a containerization service also
both are for containerization but this
gives you something and this gives you
something else for the advanced options
it doesn't say for us to do anything so
we are just gonna click on deploy
and now we can test our endpoint from
the endpoints that we can find here so
it's in progress if I go here
under the assets I can find something
called endpoints and I can find the
real-time ones and the batch endpoints
and we have created a real-time endpoint
so we are going to find it under this uh
title so if I like click on it I should
be able to test it once it's ready
it's still like loading but this is the
input and this is the output that we
will get back so if I click on test and
from here I will input some data to the
endpoint
which are the patient information The
Columns that we have already seen in our
data set the patient ID the pregnancies
and of course of course I'm not gonna
enter the label that I'm trying to
predict so I'm not going to give him if
the patient is diabetic or not this end
point is to tell me this is the end
point or the URL is going to give me
back this information whether someone
has diabetes or he doesn't so if I input
these this data I'm just going to copy
it and go to my endpoint and click on
test I'm gonna give the result pack
which are the three columns that we have
defined inside our python script the
patient ID the diabetic prediction and
the probability the certainty of whether
someone is diabetic or not based on the
uh based on the prediction so that's it
and like uh I think that this is really
simple step to do you can do it on your
own you can test it
and I think that I have finished so
thank you
uh yes we are running out of time I I
just wanted to uh thank you John for for
this demo for going through all these
steps to
um create train a classification model
and also deploy it as a predictive
service and I encourage you all to go
back to the learn module
um and uh like depend all these topics
at your at your own pace and also maybe
uh do this demo on your own on your
subscription on your Azure for student
subscription
um and I would also like to recall that
this is part of a series of study
sessions of cloud skill challenge study
sessions
um so you will have more in the in the
in the following days and this is for
you to prepare let's say to to help you
in taking the a cloud skills challenge
which collect
a very interesting learn module that you
can use to scale up on various topics
and some of them are focused on AI and
ml so if you are interested in these
topics you can select these these learn
modules
um so let me also copy
um the link the short link to the
challenge in the chat uh remember that
you have time until the 13th of
September to take the challenge and also
remember that in October on the 7th of
October you have the you can join the
student the the student developer Summit
which is uh which will be a virtual or
in for some for some cases and hybrid
event so stay tuned because you will
have some surprises in the following
days and if you want to learn more about
this event you can check the Microsoft
Imaging cap Twitter page and stay tuned
so thank you everyone for uh for joining
this session today and thank you very
much Sean for co-hosting with with this
session with me it was a pleasure
thank you so much Carlotta for having me
with you today and thank you like for
giving me this opportunity to be with
you here
great I hope that uh yeah I hope that we
work again in the future sure I I hope
so as well
um so
bye bye speak to you soon
bye