A simple probabilistic way to Logistic regression….

Source: https://unsplash.com/photos/505eectW54k

When I started to learn classification techniques, I searched many books and many blogs for the probabilistic approach and why we need to take some particular functions. Because though it is simple concept, many of us do not understand the play behind the scene.

In data science, we often deal with classification tasks. For classification tasks, a simple model is LOGISTIC REGRESSION. Everybody use this model one or the other time. If we are using python, sklearn library will provide implementation of this model and it is more or less 3 to 4…


An amazing way to transform given basis into orthonormal basis

Source: https://unsplash.com/photos/MD6E2Sv__iA

In Linear Algebra, We usually deal with vector spaces and its subspaces. It is easier to perform any operations, if we have orthogonal basis of subspace. Standard basis are orthogonal and orthonormal. Hence it will be easy when we work with standard basis. So what do we do if we have basis of subspace which are not orthogonal? There comes Gram-Schmidt Orthogonalization process. Let us walkthrough the process.

A subspace, here inner product space, has linearly independent set of vectors in the basis. The number of vectors in the basis is dimension of the subspace. In orthogonalization process we take…

Praveen Hegde

MSc Statistics student, Machine Learning Learner

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store