Ejemplo n.º 1
0
 def find(self, species_name):
     session = Session()
     try:
         species = session.query(DbSpecies).filter(
             DbSpecies.name == species_name).one()
     except:
         species = None
     finally:
         session.close()
     return species
Ejemplo n.º 2
0
 def get_species(self):
     session = Session()
     try:
         species = session.query(DbSpecies).filter(
             DbSpecies.species_id == self.species_id).one()
     except:
         species = None
     finally:
         session.close()
     return species
Ejemplo n.º 3
0
 def find_by_id(self, type_id):
     session = Session()
     try:
         habitat_type = session.query(HabitatType).filter(
             HabitatType.type_id == type_id).one()
     except:
         habitat_type = None
     finally:
         session.close()
     return habitat_type
Ejemplo n.º 4
0
 def find(self, species_name):
     session = Session()
     try:
         species = session.query(DbSpecies).filter(
             DbSpecies.name == species_name
             ).one()
     except:
         species = None
     finally:
         session.close()
     return species
Ejemplo n.º 5
0
 def get_species(self):
     session = Session()
     try:
         species = session.query(DbSpecies).filter(
             DbSpecies.species_id == self.species_id
             ).one()
     except:
         species = None
     finally:
         session.close()
     return species
Ejemplo n.º 6
0
Archivo: n2k.py Proyecto: leaffan/geo
 def find_by_id(self, type_id):
     session = Session()
     try:
         habitat_type = session.query(HabitatType).filter(
             HabitatType.type_id == type_id
             ).one()
     except:
         habitat_type = None
     finally:
         session.close()
     return habitat_type
Ejemplo n.º 7
0
Archivo: n2k.py Proyecto: leaffan/geo
 def get_typical_species(self):
     session = Session()
     try:
         habitat_species_links = session.query(HabitatTypesSpecies).filter(
             HabitatTypesSpecies.type_id == self.type_id
         ).all()
         species_ids = [lnk.species_id for lnk in habitat_species_links]
         species = session.query(Species).filter(
             Species.species_id.in_(species_ids)
             ).all()
     except:
         species = None
     finally:
         session.close()
     return species
Ejemplo n.º 8
0
Archivo: n2k.py Proyecto: leaffan/geo
 def get_associated_habitat_types(self):
     session = Session()
     try:
         species_habitat_links = session.query(HabitatTypesSpecies).filter(
             HabitatTypesSpecies.species_id == self.species_id
         ).all()
         type_ids = [lnk.type_id for lnk in species_habitat_links]
         habitat_types = session.query(HabitatType).filter(
             HabitatType.type_id.in_(type_ids)
             ).all()
     except:
         habitat_types = None
     finally:
         session.close()
     return habitat_types
Ejemplo n.º 9
0
 def find_by_name(self, name):
     session = Session()
     try:
         species = session.query(Species).filter(Species.name == name).one()
     except:
         # finding species
         spf = SpeciesFinder(name)
         sp_name, sp_url, sp_info = spf.find_species()
         # creating new species
         species = Species(sp_name, sp_url, sp_info)
         # adding species to database
         session.merge(species)
         session.commit()
     finally:
         session.close()
     return species
Ejemplo n.º 10
0
Archivo: n2k.py Proyecto: leaffan/geo
 def find_by_name(self, name):
     session = Session()
     try:
         species = session.query(Species).filter(
             Species.name == name
             ).one()
     except:
         # finding species
         spf = SpeciesFinder(name)
         sp_name, sp_url, sp_info = spf.find_species()
         # creating new species
         species = Species(sp_name, sp_url, sp_info)
         # adding species to database
         session.merge(species)
         session.commit()
     finally:
         session.close()
     return species
Ejemplo n.º 11
0
 def get_associated_habitat_types(self):
     session = Session()
     try:
         species_habitat_links = session.query(HabitatTypesSpecies).filter(
             HabitatTypesSpecies.species_id == self.species_id).all()
         type_ids = [lnk.type_id for lnk in species_habitat_links]
         habitat_types = session.query(HabitatType).filter(
             HabitatType.type_id.in_(type_ids)).all()
     except:
         habitat_types = None
     finally:
         session.close()
     return habitat_types
Ejemplo n.º 12
0
 def get_typical_species(self):
     session = Session()
     try:
         habitat_species_links = session.query(HabitatTypesSpecies).filter(
             HabitatTypesSpecies.type_id == self.type_id).all()
         species_ids = [lnk.species_id for lnk in habitat_species_links]
         species = session.query(Species).filter(
             Species.species_id.in_(species_ids)).all()
     except:
         species = None
     finally:
         session.close()
     return species
Ejemplo n.º 13
0
    #    spf.set_search_name(str(sp.name))
    #    sp_name, sp_url, sp_info = spf.find_species()
    #    new_species = Species(sp_name, sp_url, sp_info)
    #    session.merge(new_species)
    #    session.commit()
    #    print "Done"

    from sqlalchemy import and_

    tgt = r"D:\work\ms.monina\wp4\florkart_2012_occurrences_eunis_characteristic_species_n2k_areas_germany.txt"

    habitat_type = '9120'

    output = list()

    session = Session()

    habitat_types = session.query(HabitatType).all()

    for ht in habitat_types[:]:

        habitat_type = ht.type_id

        # retrieving all sites with the current habitat type
        sites_with_habitat_type = session.query(SitesHabitatTypes).filter(
            SitesHabitatTypes.type_id == habitat_type).all()

        # bailing out if no sites were found
        if not len(sites_with_habitat_type):
            continue
Ejemplo n.º 14
0
from common import Session

from first import Domain_ as FirstDomain
from second import Domain_ as SecondDomain


if __name__ == '__main__':
    session = Session()
    print session.query(SecondDomain).one()
    print session.query(FirstDomain).one()
Ejemplo n.º 15
0
 def __init__(self, env):
     """
     :param env:
     """
     self.session = Session.Session()
     self.get_session = self.session.get_session(env)
Ejemplo n.º 16
0
class DbSiteOnQuadTile(Base):
    __tablename__ = 'sites_on_quad_tiles'
    __autoload__ = True
    __table_args__ = {'autoload_with': Engine}

    def __init__(self):
        pass


if __name__ == '__main__':

    from operator import attrgetter

    habitat_type = '9120'

    session = Session()

    # retrieving all sites with the current habitat type
    sites_with_habitat_type = session.query(DbSitesHabitatTypes).filter(
        DbSitesHabitatTypes.type_id == habitat_type).all()
    # retrieving typical species for current habitat type
    typical_species = session.query(DbHabitatTypesSpecies).filter(
        DbHabitatTypesSpecies.type_id == habitat_type).all()

    # reducing sites to site ids
    f = attrgetter('site_id')
    site_ids = [f(site) for site in sites_with_habitat_type]
    # reducing species to species ids
    f = attrgetter('species_id')
    species_ids = [f(typical_sp) for typical_sp in typical_species]
Ejemplo n.º 17
0
Archivo: n2k.py Proyecto: leaffan/geo
    #    spf.set_search_name(str(sp.name))
    #    sp_name, sp_url, sp_info = spf.find_species()
    #    new_species = Species(sp_name, sp_url, sp_info)
    #    session.merge(new_species)
    #    session.commit()
    #    print "Done"

    from sqlalchemy import and_
    
    tgt = r"D:\work\ms.monina\wp4\florkart_2012_occurrences_eunis_characteristic_species_n2k_areas_germany.txt"
    
    habitat_type = '9120'
    
    output = list()

    session = Session()
    
    habitat_types = session.query(HabitatType).all()
    
    for ht in habitat_types[:]:
    
        habitat_type = ht.type_id
    
        # retrieving all sites with the current habitat type
        sites_with_habitat_type = session.query(SitesHabitatTypes).filter(
            SitesHabitatTypes.type_id == habitat_type).all()
        
        # bailing out if no sites were found
        if not len(sites_with_habitat_type):
            continue
Ejemplo n.º 18
0
def main():

    net = Model(n_feature)
    optimizer = torch.optim.Adam(net.parameters(), weight_decay=weight_decay)

    session = Session(net, optimizer)
    device = session.device
    clock = session.clock
    logger = session.logger

    net.to(device)
    logger.info(net)

    # prepare data
    data_loader_train, data_loader_test,  data_train, data_test = load_dataset(data_set=data_set)

    cost = torch.nn.CrossEntropyLoss()

    while True:
        clock.tock()
        if clock.epoch > n_epochs:
            break

        logger.info("Epoch {}/{}".format(clock.epoch, n_epochs))
        logger.info("-" * 10)

        train_loss_arcface, train_loss_ce, train_correct = 0.0, 0.0, 0.0
        train_correct_cls, train_num_cls = [0] * 10, [0] * 10

        for idx, data in enumerate(data_loader_train):
            X_train, y_train = data
            X_train, y_train = Variable(X_train).to(device), Variable(y_train).to(device)

            net.train()
            outputs = net(X_train)

            _, pred = torch.max(outputs.data, 1)

            for i in range(10):
                index = y_train == i
                pred_i = pred[index]
                label_i = y_train[index].data
                train_num_cls[i] += len(pred_i)
                train_correct_cls[i] += torch.sum(pred_i == label_i).item()

            optimizer.zero_grad()
            outputs_2 = am_softmax(outputs, y_train, scale, margin)
            loss_arcface = cost(outputs_2, y_train)
            loss_ce = cost(scale * outputs, y_train)

            loss_arcface.backward()
            optimizer.step()

            step_correct = torch.sum(pred == y_train.data).item()
            train_loss_arcface += loss_arcface.item()
            train_loss_ce += loss_ce.item()
            train_correct += step_correct

            if idx % 10 == 0: # update for every 10 step
                session.update_curv_state('train_step', step_correct/len(y_train), loss_arcface.item())

            if idx % 100 == 0: # print train info
                logger.info("step: {}, train arcface loss: {:.4f}, ce loss: {:.4f}, train acc: {:.4f}".format(idx, loss_arcface.item(), loss_ce.item(), step_correct / len(y_train)))
            clock.tick()

        test_loss, test_correct = 0.0, 0.0
        test_correct_cls, test_num_cls = [0] * 10, [0] * 10

        for data in data_loader_test:
            X_test, y_test = data
            X_test, y_test = Variable(X_test).to(device), Variable(y_test).to(device)
            net.eval()
            outputs = net(X_test)
            _, pred = torch.max(outputs.data, 1)

            for i in range(10):
                idx = y_test == i
                pred_i = pred[idx]
                label_i = y_test[idx].data
                test_num_cls[i] += len(pred_i)
                test_correct_cls[i] += torch.sum(pred_i == label_i).item()

            test_correct += torch.sum(pred == y_test.data).item()
            test_loss += cost(scale*outputs, y_test).item()

        train_acc = train_correct / len(data_train)
        train_loss_arcface = 64 * train_loss_arcface / len(data_train)
        train_acc_cls = np.array(train_correct_cls) / np.array(train_num_cls)
        assert np.sum(np.array(train_num_cls)) == len(data_train)
        assert np.sum(np.array(train_correct_cls)) == train_correct

        test_acc = test_correct /len(data_test)
        test_loss = 64 * test_loss / len(data_test)
        test_acc_cls = np.array(test_correct_cls) / np.array(test_num_cls)
        assert np.sum(np.array(test_num_cls)) == len(data_test)
        assert np.sum(np.array(test_correct_cls)) == test_correct


        session.update_best_state(test_acc)
        session.update_curv_state('train_epoch', train_acc, train_loss_arcface, train_acc_cls)
        session.update_curv_state('val_epoch', test_acc, test_loss, test_acc_cls)

        logger.info("Loss is:{:.4f}, Train Accuracy is:{:.2f}%, Test Accuracy is:{:.2f}%, {}".format(
            train_loss_arcface, 100 * train_acc, 100 * test_acc, session.best_state))
        logger.info(', '.join([ '{:.4f}'.format(x) for x in train_acc_cls]))
        logger.info(', '.join(['{:.4f}'.format(x) for x in test_acc_cls]))


        if clock.epoch in [5, 20, 50, 100]:
            session.save_checkpoint('epoch-{}'.format(clock.epoch))
        session.save_checkpoint('latest')

    session.end()

    print('drawing curve')
    draw_curve(session.curv_stat, session.log_curv_dir, data_set)
    if n_feature == 2:
        print('drawing featue')
        draw_feature(os.path.join(session.log_model_dir, 'best-accuracy'), 1, data_set)
Ejemplo n.º 19
0
class DbSiteOnQuadTile(Base):
    __tablename__ = 'sites_on_quad_tiles'
    __autoload__ = True
    __table_args__ = {'autoload_with': Engine}

    def __init__(self):
        pass
    
if __name__ == '__main__':
    
    from operator import attrgetter
    
    habitat_type = '9120'

    session = Session()
    
    # retrieving all sites with the current habitat type
    sites_with_habitat_type = session.query(DbSitesHabitatTypes).filter(
        DbSitesHabitatTypes.type_id == habitat_type).all()
    # retrieving typical species for current habitat type
    typical_species = session.query(DbHabitatTypesSpecies).filter(
        DbHabitatTypesSpecies.type_id == habitat_type).all()

    # reducing sites to site ids
    f = attrgetter('site_id')
    site_ids = [f(site) for site in sites_with_habitat_type]
    # reducing species to species ids
    f = attrgetter('species_id')
    species_ids = [f(typical_sp) for typical_sp in typical_species]