You can think of vectors representing many dimensions of related information. For instance, Netflix might store all the ratings a user gives to movies in a vector. This is clearly a vector of very large dimensions (in the millions) and very sparse as the user might have rated only a few movies.
Similarly, Amazon might store the items purchased by a user in a vector, with each slot or dimension representing a unique product and the value of the slot, the number of such items the user bought. One task that is frequently done in these settings is to find similarities between users. And, we can use dot-product between vectors to do just that.
As you know, the dot-product is proportional to the length of two vectors and to the angle between them. In fact, the dot-product between two vectors, normalized by their lengths is called as the cosine distance and is frequently used in recommendation engines.
u <- c(0.5, 0.5)
v <- c(3, -4)
u%*%v
## [,1]
## [1,] -0.5
crossprod(u,v) #is "slightly faster" than %*%
## [,1]
## [1,] -0.5
\[sqrt(u_{1}^{\,2} + u_{2}^{\;2})\]
u_len <- sqrt(0.5^2 + 0.5^2)
v_len <- sqrt(3^2 + (-4)^2)
u_len
## [1] 0.7071068
v_len
## [1] 5
\[3u - 2v \]
(3*u) - (2*v)
## [1] -4.5 9.5
\[cos\theta = \frac{u \cdot v}{||u||v||} \]
acos( u%*%v / ( u_len * v_len ) )
## [,1]
## [1,] 1.712693
Set up a system of equations with 3 variables and 3 constraints and solve for x. Please write a function in R that will take two variables (matrix A & constraint vector b) and solve using elimination. Your function should produce the right answer for the system of equations for any 3-variable, 3-equation system. You don’t have to worry about degenerate cases and can safely assume that the function will only be tested with a system of equations that has a solution. Please note that you do have to worry about zero pivots, though. Please note that you should not use the built-in function solve to solve this system or use matrix inverses. The approach that you should employ is to construct an Upper Triangular Matrix and then back-substitute to get the solution. Alternatively, you can augment the matrix A with vector b and jointly apply the Gauss Jordan elimination procedure.
function_solve <- function(matrix1, vector1){
#join the matrix and vector together
augmented_matrix <- cbind(matrix1, vector1)
augmented_matrix[2,] <- augmented_matrix[2,] - ( augmented_matrix[2,1] / augmented_matrix[1,1] ) * augmented_matrix[1,]
augmented_matrix[3,] <- augmented_matrix[3,] - ( augmented_matrix[3,1] / augmented_matrix[1,1] ) * augmented_matrix[1,]
#zero pivot
if(augmented_matrix[2,2] == 0){
if(augmented_matrix[3,2] == 0){
return(x)
}else{
temp <- augmented_matrix[2,]
augmented_matrix[2,] <- augmented_matrix[3,]
augmented_matrix[3,] <- temp
}
}
augmented_matrix[3, ] <- augmented_matrix[3, ] - ( augmented_matrix[3,2] / augmented_matrix[2,2] ) * augmented_matrix[2,]
#calculations
r <- nrow(augmented_matrix)
c <- ncol(augmented_matrix)
calc <- c(1:r)
calc[3] <- augmented_matrix[r,c]/augmented_matrix[r,c-1]
calc[2] <- ( augmented_matrix[2,4] - augmented_matrix[2,3] * calc[3]) / augmented_matrix[2,2]
calc[1] <- ( augmented_matrix[1,4] - augmented_matrix[1,3] * calc[3] - augmented_matrix[1,2] * calc[2]) / augmented_matrix[1,1]
return(calc)
}
matrix1 <- matrix(c(1,2,-1,1,-1,-2,3,5,4), nrow=3)
vector1 <- c(1,2,6)
function_solve(matrix1, vector1)
## [1] -1.5454545 -0.3181818 0.9545455
#compare to the built in function
solve(matrix1, vector1)
## [1] -1.5454545 -0.3181818 0.9545455
https://stat.ethz.ch/R-manual/R-devel/library/base/html/Trig.html
https://stat.ethz.ch/R-manual/R-devel/library/base/html/crossprod.html
https://www.khanacademy.org/math/precalculus/precalc-matrices
https://stackoverflow.com/questions/16044377/how-to-do-gaussian-elimination-in-r-do-not-use-solve
https://martin-thoma.com/solving-linear-equations-with-gaussian-elimination/