-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathnaive-bayes.py
59 lines (48 loc) · 1.87 KB
/
naive-bayes.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
import numpy as np
class GaussianNaiveBayes:
def __init__(self):
self.classes = None
self.mean = None
self.var = None
self.priors = None
def fit(self, X, y):
self.classes = np.unique(y)
n_features = X.shape[1]
n_classes = len(self.classes)
# Initialize mean, var, and priors
self.mean = np.zeros((n_classes, n_features), dtype=np.float64)
self.var = np.zeros((n_classes, n_features), dtype=np.float64)
self.priors = np.zeros(n_classes, dtype=np.float64)
# Calculate mean, var, and priors for each class
for idx, c in enumerate(self.classes):
X_c = X[y == c]
self.mean[idx, :] = X_c.mean(axis=0)
self.var[idx, :] = X_c.var(axis=0)
self.priors[idx] = X_c.shape[0] / float(X.shape[0])
def predict(self, X):
y_pred = [self._predict(x) for x in X]
return np.array(y_pred)
def _predict(self, x):
posteriors = []
# Calculate posterior probability for each class
for idx, c in enumerate(self.classes):
prior = np.log(self.priors[idx])
class_conditional = np.sum(np.log(self._pdf(idx, x)))
posterior = prior + class_conditional
posteriors.append(posterior)
# Return the class with the highest posterior probability
return self.classes[np.argmax(posteriors)]
def _pdf(self, class_idx, x):
mean = self.mean[class_idx]
var = self.var[class_idx]
numerator = np.exp(- (x - mean) ** 2 / (2 * var))
denominator = np.sqrt(2 * np.pi * var)
return numerator / denominator
# Example usage:
# X, y would be predefined datasets
# For example:
# X = np.array([[1, 2], [2, 3], [3, 4], [4, 5], [5, 6]])
# y = np.array([0, 0, 0, 1, 1])
# model = GaussianNaiveBayes()
# model.fit(X, y)
# predictions = model.predict(X)