예제 #1
0
def main():
    l, h, e, trainSetFile, testSetFile = getArgs()
    trainDataSet = arff.load(open(trainSetFile, 'r'))
    testDataSet = arff.load(open(testSetFile, 'r'))
    labels = trainDataSet['attributes'][-1][1]

    fullTrain, fullTest = dataprocess.process(trainDataSet, testDataSet,
                                              labels)

    ann = NeuralNet(h, 1, fullTrain)
    trainInfo = ann.train(fullTrain, e, l)
    for item in trainInfo:
        print('Epoch: {0}\tCross-entropy error: {1}\tCorrectly classified instances: {2}\tMisclassified instances: {3}'.format\
        (item[0], item[1], item[2], item[3]))

    testInfo, correct, wrong = ann.test(fullTest, labels)
    for item in testInfo:
        print('Activation of output unit: {0}\tPredicted class: {1}\tCorrect class: {2}'.format\
        (item[0], item[1], item[2]))
    print('Correctly classified instances: {0}\tMisclassified instances: {1}'.
          format(correct, wrong))
예제 #2
0
  def get(self):
    args = parser.parse_args(strict=True)

    ageMin = 0
    if (args['ageMin'] is not None):
      ageMin = args['ageMin']
        
    ageMax = 100
    if (args['ageMax'] is not None):
      ageMin = args['ageMax']

    gap = 0
    if (args['gap'] is not None):
      gap = args['gap']

    overlap = 0
    if (args['overlap'] is not None):
      overlap = args['overlap']

    fileName = 'data/Ace_Beta_Diur_1SamplingRate.txt'
    if (args['fileName'] is not None):
      fileName = args['fileName']


    drugList = ['1', '3', '5']
    if (args['drugList'] is not None):
      drugList = args['drugList']
      drugList = "".join(drugList)

    patients, drugStats = dataprocess.process(fileName = fileName, drugList = drugList, overlap = overlap, gap = gap, ageMin = ageMin, ageMax = ageMax)

    patientsObj = dataprocess.patientsJSON(patients)
    drugsObj = dataprocess.drugsJSON(drugStats, drugList)

    return { "patients": patientsObj,
             "drugs" : drugsObj }
예제 #3
0
import numpy as np
import cv2 as cv
from collections import deque
import baselinemodel as bm
import dataprocess as dp

from keras.datasets import mnist

(X_train, y_train), (X_test, y_test) = mnist.load_data()
x, y, input, output = dp.process(X_train, y_train)

model = bm.baseline_model(input, output)
model.fit(x, y)


def predict(k):
    l = np.zeros((1, 28, 28), dtype=np.uint8)
    l[0] = k

    v = dp.inputProcess(l)

    i = model.predict_class(v)
    print(i)


def image_resize(image, width=None, height=None, inter=cv.INTER_AREA):
    # initialize the dimensions of the image to be resized and
    # grab the image size
    dim = None
    (h, w) = image.shape[:2]
예제 #4
0
total = 0
resultnum = 0
#print(clname[0][-1])
for j in range(len(clname)):
    a = np.array(data)[:, j]
    a = a.reshape(1, -1)
    # print("a.shape before")
    # print(a.shape)
    a = dp.PCA.transform(a)
    # print("a.shape after")
    # print(a.shape)
    #print(a)
    for i in range(len(list_)):
        if (i != int(clname[j][11])):
            name = list_[i]
            line, number = dp.process(name)
            if (len(clname[j]) > 13):
                #print(clname[j])
                #print(idf.Archimedes(a,line,number,int(0)))
                if (idf.Archimedes(a, line, number, int(clname[j][13:]))):
                    resultnum += 1
            else:
                #print(0)
                #print(clname[j])
                # print(idf.Archimedes(a,line,number,int(0)))
                if (idf.Archimedes(a, line, number, int(0))):
                    resultnum += 1
            total += 1
print(total)
print("identificationattack结果是:")
print(resultnum / total)
예제 #5
0
import os
import pandas as pd
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.multiclass import OneVsRestClassifier
from sklearn.linear_model import LogisticRegression
from sklearn.pipeline import Pipeline
import json

# Load Data
train_set = pd.read_csv('data/hackathon_loreal_train_set.csv')
test_set = pd.read_csv('data/hackathon_loreal_test_set.csv')

if os.path.isfile('data/data_process.csv'):
    train_set = pd.read_csv('data/data_process.csv', sep=";")
else:
    process(train_set, "train_process")

X_train = train_set.text
y_train = train_set[['skincare', 'hair', 'make-up', 'other']]
categories = ['skincare', 'hair', 'make-up', 'other']
#X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.30, random_state=42)

# Define a pipeline combining a text feature extractor with multi lable classifier
LogReg_pipeline = Pipeline([
    ('tfidf', TfidfVectorizer()),
    ('clf', OneVsRestClassifier(LogisticRegression(solver='sag'), n_jobs=1)),
])

if os.path.isfile('data/data_process.csv'):
    test_set = pd.read_csv('data/test_process.csv', sep=";")
else: