コード例 #1
0
    def get(self, params):
        if self.get_argument("fetch", True):
            scope = ['https://spreadsheets.google.com/feeds']
            credentials = ServiceAccountCredentials.from_json_keyfile_name('client_secret.json',scope)
            word_repository = WordRepository()
            dict_utils = DictUtils()

            try:
                gc = gspread.authorize(credentials)
                sh = gc.open_by_key("1QeU3AoSghCAvYBD8kauC8oEZ0wA6j_gPj1pkvIY4MPU")

                for ws_count, worksheets in enumerate(sh.worksheets()):

                    word_list = worksheets.col_values(1)
                    type_list = worksheets.col_values(2)
                    meaning_list = worksheets.col_values(3)

                    if ws_count < 26:
                        word_tuple_list = []
                        for i, (word, word_type, meaning) in enumerate(zip(word_list, type_list, meaning_list)):

                            if i > 0:
                                if word != '':

                                    word_tuple_list.append(dict(id=str(uuid.uuid1().hex), word=word,
                                                                type=word_type,
                                                                meaning_zg=meaning,
                                                                meaning_uni=meaning))
                            else:
                                pass
                        word_repository.bulk_insert(word_tuple_list, dict_utils.get_model(str(worksheets.title).lower()))

                    else:
                        self.respond({}, NO_CONTENT_ERROR, code=SUCCESS)

            except Exception as ex:
                print ex.message
                self.respond({}, ex.message, code=SERVER_ERROR)
コード例 #2
0
    parser.add_argument('--drop_out', type=float, default=0.5)
    parser.add_argument('--m', type=float, default=0.3)
    parser.add_argument('--p', type=float, default=0.55)
    parser.add_argument('--flag', default="PER")
    parser.add_argument('--dataset', default="conll2003")
    parser.add_argument('--lr', type=float, default=1e-4)
    parser.add_argument('--batch_size', type=int, default=300)

    parser.add_argument('--model', default="")
    parser.add_argument('--iter', type=int, default=1)

    args = parser.parse_args()

    dp = DataPrepare(args.dataset)
    mutils = AdaptivePUUtils(dp)
    dutils = DictUtils()

    trainSet, validSet, testSet, prior = mutils.load_new_dataset(
        args.flag, args.dataset, args.iter, args.p)
    print(prior)
    trainSize = len(trainSet)
    validSize = len(validSet)
    testSize = len(testSet)
    print(("train set size: {}, valid set size: {}, test set size: {}").format(
        trainSize, validSize, testSize))

    charcnn = CharCNN(dp.char2Idx)
    wordnet = WordNet(dp.wordEmbeddings, dp.word2Idx)
    casenet = CaseNet(dp.caseEmbeddings, dp.case2Idx)
    featurenet = FeatureNet()
    pulstmcnn = AdaPULSTMCNN2(dp, charcnn, wordnet, casenet, featurenet, 150,