Free Databricks Databricks-Certified-Professional-Data-Scientist Questions
-
-
Page: 1/28
Total 138 Questions
Question No 1
Feature Hashing approach is "SGD - based classifiers avoid the need to predetermine vector size by simply picking a reasonable size and shoehorning the training data into vectors of that size" now with large vectors or with multiple locations per feature in Feature hashing?
Question No 2
What are the advantages of the Hashing Features?
Question No 3
Question - 3: In machine learning, feature hashing, also known as the hashing trick (by analogy to the kernel trick), is a fast and space - efficient way of vectorizing features (such as the words in a language), i.e., turning arbitrary features into indices in a vector or matrix. It works by applying a hash function to the features and using their hash values modulo the number of features as indices directly, rather than looking the indices up in an associative array. So what is the primary reason of the hashing trick for building classifiers?
Question No 4
Suppose A, B , and C are events. The probability of A given B , relative to P(|C), is the same as the probability of A given B and C (relative to P ). That is,
Question No 5
What is the considerable difference between L1 and L2 regularization?
-
Page: 1/28
Total 138 Questions
-