Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Binary file modified .DS_Store
Binary file not shown.
8 changes: 8 additions & 0 deletions .idea/.gitignore

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

8 changes: 8 additions & 0 deletions .idea/OpenBaseLab-Edu.iml

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

6 changes: 6 additions & 0 deletions .idea/inspectionProfiles/profiles_settings.xml

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

4 changes: 4 additions & 0 deletions .idea/misc.xml

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

8 changes: 8 additions & 0 deletions .idea/modules.xml

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

6 changes: 6 additions & 0 deletions .idea/vcs.xml

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

55 changes: 55 additions & 0 deletions BaseML/AdaBoost.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,55 @@
import pandas as pd
import numpy as np
import os
from sklearn.metrics import accuracy_score, mean_squared_error
from sklearn.ensemble import AdaBoostClassifier


class AdaBoost:
def __init__(self
):
self.cwd = os.path.dirname(os.getcwd()) # 获取当前文件的绝对路径
self.file_dirname = os.path.dirname(os.path.abspath(__file__))
self.model = AdaBoostClassifier(n_estimators=100, random_state=0)
self.dataset_path = ' '
self.test_size = ' '
self.test_set = ' '

def train(self, seed=0, data_type='csv'):
np.random.seed(seed)
if data_type == 'csv':
dataset = pd.read_csv(self.dataset_path, sep=',', header=None).values
elif data_type == 'pandas':
dataset = self.load_pd()
elif data_type == 'list':
dataset = self.load_list()
np.random.shuffle(dataset)

data, label = dataset[:, :-1], dataset[:, -1]
train_index = int((1 - self.test_size) * len(dataset))
train_data, train_label = data[:train_index, ], label[:train_index]
self.test_set = {
'data': data[train_index:, ],
'label': label[train_index:]
}
self.model.fit(train_data, train_label)

def inference(self, mode='cls'):
pred = self.model.predict(self.test_set['data'])
if mode == 'cls':
acc = accuracy_score(self.test_set['label'], pred)
print('准确率为:{}%'.format(acc * 100))
elif mode == 'reg':
loss = mean_squared_error(self.test_set['label'], pred)
print('Loss: {}'.format(loss))

def load_dataset(self, path, test_size=0.2):
self.dataset_path = path
self.test_size = test_size

def load_pd(self):
pass

def load_list(self):
pass

2 changes: 0 additions & 2 deletions BaseML/CART.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,9 +4,7 @@

class CART:
def __init__(self,
backbone='KNNClassifier',
):
self.backbone = backbone
# 获取外部运行py的绝对路径
self.cwd = os.path.dirname(os.getcwd())
# 获取当前文件的绝对路径
Expand Down
55 changes: 55 additions & 0 deletions BaseML/GaussianNB.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,55 @@
import pandas as pd
import numpy as np
import os

from sklearn.metrics import accuracy_score, mean_squared_error
from sklearn.naive_bayes import GaussianNB


class GaussianNB:
def __init__(self
):
self.cwd = os.path.dirname(os.getcwd()) # 获取当前文件的绝对路径
self.file_dirname = os.path.dirname(os.path.abspath(__file__))
self.model = GaussianNB()
self.dataset_path = ' '
self.test_size = ' '
self.test_set = ' '

def train(self, seed=0, data_type='csv'):
np.random.seed(seed)
if data_type == 'csv':
dataset = pd.read_csv(self.dataset_path, sep=',', header=None).values
elif data_type == 'pandas':
dataset = self.load_pd()
elif data_type == 'list':
dataset = self.load_list()
np.random.shuffle(dataset)

data, label = dataset[:, :-1], dataset[:, -1]
train_index = int((1 - self.test_size) * len(dataset))
train_data, train_label = data[:train_index, ], label[:train_index]
self.test_set = {
'data': data[train_index:, ],
'label': label[train_index:]
}
self.model.fit(train_data, train_label)

def inference(self, mode='cls'):
pred = self.model.predict(self.test_set['data'])
if mode == 'cls':
acc = accuracy_score(self.test_set['label'], pred)
print('准确率为:{}%'.format(acc * 100))
elif mode == 'reg':
loss = mean_squared_error(self.test_set['label'], pred)
print('Loss: {}'.format(loss))

def load_dataset(self, path, test_size=0.2):
self.dataset_path = path
self.test_size = test_size

def load_pd(self):
pass

def load_list(self):
pass
2 changes: 0 additions & 2 deletions BaseML/KNN.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,10 +5,8 @@

class KNN:
def __init__(self,
backbone='KNN',
n_neighbors=10,
):
self.backbone = backbone
# 获取外部运行py的绝对路径
self.cwd = os.path.dirname(os.getcwd())
# 获取当前文件的绝对路径
Expand Down
38 changes: 17 additions & 21 deletions LR.py → BaseML/LR.py
Original file line number Diff line number Diff line change
@@ -1,22 +1,19 @@
from turtle import back
import pandas as pd
import numpy as np
import os
import cv2
import os
from sklearn.metrics import accuracy_score ,mean_squared_error, r2_score
from sklearn.metrics import accuracy_score, mean_squared_error, r2_score
from sklearn import linear_model

class LR:
def __init__ (self,
backbone='LR'
):
self.backbone = backbone #获取外部运行py的绝对路径
self.cwd = os.path.dirname(os.getcwd()) #获取当前文件的绝对路径
self.file_dirname = os.path.dirname(os.path.abspath(__file__))
def __init__(self,):
self.cwd = os.path.dirname(os.getcwd()) #获取当前文件的绝对路径
self.file_dirname = os.path.dirname(os.path.abspath(__file__))
self.model = linear_model.LinearRegression()
self.dataset_path = ' '
self.test_size = ' '

def train(self,seed=0,data_type='csv'):
def train(self, seed=0, data_type='csv'):
np.random.seed(seed)
if data_type == 'csv':
dataset = pd.read_csv(self.dataset_path,sep=',',header=None).values
Expand All @@ -26,27 +23,26 @@ def train(self,seed=0,data_type='csv'):
dataset = self.load_list()
np.random.shuffle(dataset)

data,label = dataset[:,:-1],dataset[:,-1]
train_index = int((1-self.test_size)*len(dataset))
train_data,train_label = data[:train_index,],label[:train_index]
data, label = dataset[:,:-1],dataset[:,-1]
train_index = int((1-self.test_size)*len(dataset))
train_data, train_label = data[:train_index,],label[:train_index]
self.test_set = {
'data':data[train_index:,],
'label':label[train_index:]
'data': data[train_index:,],
'label': label[train_index:]
}
self.model.fit(train_data,train_label)

def inference(self,mode = 'cls'):
def inference(self, mode='cls'):
pred = self.model.predict(self.test_set['data'])
loss = mean_squared_error(self.test_set['label'],pred)
print('Loss: {}'.format(loss))


def load_dataset(self,path,test_size=0.2):
self.dataset_path = path
self.test_size=test_size
self.test_size = test_size

def load_pd():
def load_pd(self):
pass

def load_list():
pass
def load_list(self):
pass
2 changes: 0 additions & 2 deletions BaseML/PCA.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,10 +4,8 @@

class PCA:
def __init__(self,
backbone='KNNClassifier',
n_components='mle',
):
self.backbone = backbone
# 获取外部运行py的绝对路径
self.cwd = os.path.dirname(os.getcwd())
# 获取当前文件的绝对路径
Expand Down
4 changes: 1 addition & 3 deletions BaseML/Perceptron.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,10 +3,8 @@


class Perceptron:
def __init__(self,
backbone='KNNClassifier',
def __init__(self
):
self.backbone = backbone
# 获取外部运行py的绝对路径
self.cwd = os.path.dirname(os.getcwd())
# 获取当前文件的绝对路径
Expand Down
54 changes: 54 additions & 0 deletions BaseML/SVM.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,54 @@
import pandas as pd
import numpy as np
import os
from sklearn.metrics import accuracy_score, mean_squared_error
from sklearn.svm import SVC


class SVM:
def __init__(self,
):
self.cwd = os.path.dirname(os.getcwd()) # 获取当前文件的绝对路径
self.file_dirname = os.path.dirname(os.path.abspath(__file__))
self.model = SVC()
self.dataset_path = ' '
self.test_size = ' '
self.test_set = ' '

def train(self, seed=0, data_type='csv'):
np.random.seed(seed)
if data_type == 'csv':
dataset = pd.read_csv(self.dataset_path, sep=',', header=None).values
elif data_type == 'pandas':
dataset = self.load_pd()
elif data_type == 'list':
dataset = self.load_list()
np.random.shuffle(dataset)

data, label = dataset[:, :-1], dataset[:, -1]
train_index = int((1 - self.test_size) * len(dataset))
train_data, train_label = data[:train_index, ], label[:train_index]
self.test_set = {
'data': data[train_index:, ],
'label': label[train_index:]
}
self.model.fit(train_data, train_label)

def inference(self, mode='cls'):
pred = self.model.predict(self.test_set['data'])
if mode == 'cls':
acc = accuracy_score(self.test_set['label'], pred)
print('准确率为:{}%'.format(acc * 100))
elif mode == 'reg':
loss = mean_squared_error(self.test_set['label'], pred)
print('Loss: {}'.format(loss))

def load_dataset(self, path, test_size=0.2):
self.dataset_path = path
self.test_size = test_size

def load_pd(self):
pass

def load_list(self):
pass
Loading