Beispiel #1
0
def browse_table(conn_params, get_data={}, post_data={}):
    # initializations
    sub_q_data = {'tbl': get_data.get('tbl'),'db':get_data.get('db')}
    pg_index = get_data.get('pg', 1) # page 1 is the starting point
    sub_q_data['limit'] = getattr(settings, 'TT_MAX_ROW_COUNT', 30) # 30 is the default row limit
    # page 1 starts with offset 0 and so on
    sub_q_data['offset'] = int(sub_q_data['limit']) * ( int(pg_index) - 1)
    for item in ('schm', 'sort_key', 'sort_dir',):
        if get_data.has_key(item): sub_q_data[item] = get_data.get(item)
    # retrieve and run queries
    keys = rpr_query(conn_params, 'primary_keys', sub_q_data)
    count = sa.full_query(conn_params, 
        sql.generate_query('count_rows', conn_params['dialect'], sub_q_data)[0],
        )['rows']
    # get the table's row listing
    r = sa.full_query(conn_params,
        sql.generate_query('browse_table', conn_params['dialect'], sub_q_data)[0]
        )
    # format and return data
    if type(r) == dict:
        r.update({
            'total_count': count[0][0],
            'pg': pg_index,
            'limit':sub_q_data['limit'],
            'keys': keys
        })
        return r
    else:
        return fns.http_500(r)
Beispiel #2
0
def get_constraints(conn_params, query_type, get_data={}, form_data={}):
    # first face the simple guy
    if conn_params['dialect'] == 'mysql':
        r = sa.full_query(conn_params,
            sql.generate_query(query_type, conn_params['dialect'], get_data)[0])
        return r

    # this sequence of code would be repeated often
    def get_cols_with_post_as_dict(get_data):
        # query for column name associated with its ordinal position. 
        sql_stmt = sql.generate_query('column_assoc', conn_params['dialect'], get_data)[0]
        cols_with_pos = sa.full_query(conn_params, sql_stmt)
        # transform to dict for quick and easy indexing and retrieval
        cols_with_pos = dict(cols_with_pos['rows'])
        return cols_with_pos


    # get raw constrainsts description
    sql_stmt = sql.generate_query('constraints', conn_params['dialect'], get_data)[0]
    con_desc = sa.full_query(conn_params, sql_stmt,);
    # char to constraint type mapping as written in PostgreSQL documentation
    char_constraint_map = {
        'c': 'CHECK',
        'f': 'FOREIGN KEY',
        'p': 'PRIMARY KEY',
        'u': 'UNIQUE',
    }

    columns = ['type', 'columns', 'description', 'name',]
    rows = []
    cols_with_pos = get_cols_with_post_as_dict(get_data)

    for row in con_desc['rows']:
        # contype 0, conname 1, conkey 2, confkey 3, relname 4, consrc 5
        contype, conname, conkey, confkey, relname, consrc = row
        # any contype not in the mapping above is not yet supported
        if not char_constraint_map.has_key(contype): continue
        desc = '' # starts empty
        constrained_columns = ", ".join( [cols_with_pos[i] for i in conkey] )
        type_ = char_constraint_map[contype]
        # if this query is a foreign key translate the column names
        if contype == 'c':
            desc = consrc
        if contype == 'f':
            if relname != get_data.get('tbl'): # we need to get another columns with pos
                get_data['tbl'] = relname
                fr_cols_with_pos = get_cols_with_post_as_dict(get_data)
            else: fr_cols_with_pos = cols_with_pos

            referred_columns = [ fr_cols_with_pos[i] for i in confkey]
            desc = relname + ":" + ", ".join(referred_columns)

        # done
        rows.append([ type_, constrained_columns, desc, conname])

    return {'columns': columns, 'rows': rows, 'count': len(con_desc['rows'])}
Beispiel #3
0
def run_tbl_operations(conn_params, form_type, get_data={}, post_data={}):
    # generate / derive queries to be run
    queries, msg = [], ''
    if form_type == 'tbl_edit_form':
        queries, msg = sql.alter_table(conn_params['dialect'], get_data, post_data)
    elif form_type == 'tbl_vacuum_form':
        queries = sql.pg_vacuum_stmt(get_data, post_data)
    else:
        # implemented queries: analyze table and reindex table
        # only the pg dialect
        queries = sql.generate_query(form_type, conn_params['dialect'], get_data)

    # execute queries and return formatted status messages
    if form_type == 'tbl_edit_form':
        ret = sa.short_query(conn_params, queries)
    else: ret = sa.execute_outside_transaction(conn_params, queries)

    if ret['status'] == 'success':
        if form_type == 'tbl_edit_form': ret['msg'] = msg + ' change succesfull'
        else:
            msg = '{command} operation ran succesfully'
            if form_type == 'tbl_vacuum_form': com = 'VACUUM'
            elif form_type == 'analyze_table': com = 'ANALYZE'
            elif form_type == 'reindex_table': com = 'REINDEX'
            ret['msg'] = msg.format(command=com)

    ret['msg'] = '<p class="query-msg {status}">{msg}</p>'.format(
            status = 'error' if ret['status'] == 'error' else 'success',
            msg = ret['msg']
        )
    return ret
Beispiel #4
0
 def get_cols_with_post_as_dict(get_data):
     # query for column name associated with its ordinal position. 
     sql_stmt = sql.generate_query('column_assoc', conn_params['dialect'], get_data)[0]
     cols_with_pos = sa.full_query(conn_params, sql_stmt)
     # transform to dict for quick and easy indexing and retrieval
     cols_with_pos = dict(cols_with_pos['rows'])
     return cols_with_pos
Beispiel #5
0
def get_dependencies(conn_params, get_data={}): # might later be extended for objects besides tables
    '''
    logic from pgadmin3
    '''
    # get the total listing of the dependent types
    conn_params['db'] = get_data['db'] if get_data.has_key('db') else conn_params['db']
    q_1 = sql.generate_query('pgadmin_deps', conn_params['dialect'], get_data)[0]
    totl_deps = sa.full_query(conn_params, q_1)
    # columns in totl_deps are u'deptype', u'classid', u'relkind', u'adbin', u'adsrc', u'type', 
    #                           u'ownertable', u'refname', u'nspname'
    # raise Exception(totl_deps)
    columns = ['type', 'name', 'restriction',]
    
    tbl_data_rows = []
    for row in totl_deps['rows']:
        refname, typestr, depstr = '', '', ''
        # get the type of this object described in this row
        # type is the sixth column of this query
        type_ = row[5]
        if type_[0] in ('c', 's', 't'):
            continue # ununderstood types: handled internally
        elif type_[0] == 'i': typestr = 'index'
        elif type_[0] == 'S': typestr = 'sequence'
        elif type_[0] == 'v': typestr = 'view'
        elif type_[0] == 'x': typestr = 'exttable'
        elif type_[0] == 'p': typestr = 'function'
        elif type_[0] == 'n': typestr = 'schema'
        elif type_[0] == 'y': typestr = 'type'
        elif type_[0] == 'T': typestr = 'trigger'
        elif type_[0] == 'l': typestr = 'language'
        elif type_[0] == 'R':
            pass
        elif type_[0] == 'C':
            if type_[1] == 'c': typestr = 'check'
            elif type_[1] == 'f': 
                refname = row[6] +'.'
                typestr = 'foreign key'
            elif type_[1] == 'u': typestr = 'unique'
            elif type_[1] == 'p': typestr = 'primary key'
            elif type_[1] == 'x': typestr = 'exclude'
        elif type_[0] == 'A':
            if row[3] != None and row[3].startswith("{FUNCEXPR"):
                typestr = 'function'
                refname = row[4] # adbin
        # complete refname
        # appends the name of the foreign key if the type of object is a foreign key
        # function has the refname already set
        refname = refname + row[7] # refname
        # deptype is the first column of this query
        deptype = row[0]
        if deptype == 'i': depstr = 'internal'
        elif deptype == 'a': depstr = 'auto'
        elif deptype == 'n': depstr = 'normal'
        elif deptype == 'p': depstr = 'pin'

        tbl_data_rows.append([typestr, refname, depstr])

    return {'count': totl_deps['count'], 'columns': columns, 
        'rows': tbl_data_rows
    }
Beispiel #6
0
def browse_table(conn_params, get_data={}, post_data={}):
    # initializations
    sub_q_data = {'tbl': get_data.get('tbl'),'db':get_data.get('db')}
    sub_q_data['offset'] = get_data.get('offset') if get_data.has_key('offset') else 0
    sub_q_data['limit'] = get_data.get('limit') if get_data.has_key('limit') else getattr(settings, 'TT_MAX_ROW_COUNT', 100)
    for item in ('schm', 'sort_key', 'sort_dir',):
        if get_data.has_key(item): sub_q_data[item] = get_data.get(item)
    # retrieve and run queries
    keys = rpr_query(conn_params, 'primary_keys', sub_q_data)
    count = sa.full_query(conn_params, 
        sql.generate_query('count_rows', conn_params['dialect'], sub_q_data)[0],
        )['rows']
    # get the table's row listing
    r = sa.full_query(conn_params,
        sql.generate_query('browse_table', conn_params['dialect'], sub_q_data)[0]
        )
    # format and return data
    if type(r) == dict:
        r.update({'total_count': count[0][0], 'offset': sub_q_data['offset'],
            'limit':sub_q_data['limit'], 'keys': keys})
        return r
    else:
        return fns.http_500(r)
Beispiel #7
0
def rpr_query(conn_params, query_type, get_data={}, post_data={}):
    '''
    Run queries that have to be generated on the fly. Most queries depends on get_data, 
    while some few depends on post_data

    get_data and post_data are gotten from request.GET and request.POST or form.cleaned_data

    some functions in this file should have been (and previously have been) in this function. but 
    the contents of a function are not indexed by IDEs and that makes debugging difficult
    '''
    # common queries that returns success state as a dict only
    no_return_queries = ('create_user', 'drop_user', 'create_db','create_table',
        'drop_table', 'empty_table', 'delete_row', 'create_column', 'drop_column',
        'drop_db', 'drop_sequence', 'reset_sequence', 'drop_constraint', )
    
    psycopg2_queries = ('drop_db', )

    if query_type in no_return_queries:
        conn_params['db'] = get_data['db'] if get_data.has_key('db') else conn_params['db']
        query_data = {}
        query_data.update(get_data, **post_data)
        queries = sql.generate_query( query_type, conn_params['dialect'],query_data)

        if conn_params['dialect'] == 'postgresql' and query_type in psycopg2_queries:
            # this queries needs to be run outside a transaction block
            # SA execute functions runs all its queries inside a transaction block
            result = sa.execute_outside_transaction(conn_params, queries)
        else: result = sa.short_query(conn_params, queries)
        return HttpResponse( json.dumps(result) )
    
    # specific queries with implementations similar to both dialects
        
    elif query_type in ('indexes', 'primary_keys', 'foreign_key_relation'):
        if conn_params['dialect'] == 'postgresql' and query_type == 'indexes':
            return get_constraints(conn_params, query_type, get_data)

        r = sa.full_query(conn_params,
            sql.generate_query(query_type, conn_params['dialect'], get_data)[0])
        return r

    elif query_type in ('get_single_row',):
        sub_q_data = {'tbl': get_data['tbl'],'db':get_data['db']}
        if get_data.has_key('schm'):
            sub_q_data['schm'] = get_data['schm']
        # generate where statement
        sub_q_data['where'] = ""
        for ind in range(len(post_data)):
            sub_q_data['where'] += post_data.keys()[ind].strip() + "=" 
            val = post_data.values()[ind].strip()
            sub_q_data['where'] += fns.quote(val)
            if ind != len(post_data) - 1: sub_q_data['where'] += ' AND '
        # retrieve and run queries
        conn_params['db'] = get_data['db']
        # assert False
        q = sql.generate_query(query_type, conn_params['dialect'], sub_q_data)
        r =  sa.full_query(conn_params, q[0])
        return r
        

    elif query_type in ('table_rpr', 'table_structure', 'raw_table_structure', 'seqs_rpr'):
        sub_q_data = {'db': get_data['db'],}
        if get_data.has_key('tbl'):
            sub_q_data['tbl'] = get_data['tbl']
        if get_data.has_key('schm'):
            sub_q_data['schm'] = get_data['schm']
        # make query
        if conn_params['dialect'] == 'postgresql' and query_type == 'raw_table_structure':
            q = 'table_structure'
        else: q = query_type

        r = sa.full_query(conn_params,
            sql.generate_query(q, conn_params['dialect'], sub_q_data)[0] )
        # further needed processing
        if conn_params['dialect'] == 'postgresql' and query_type.count('table_structure'):
            rwz = []
            for tuple_row in r['rows']:
                row = list(tuple_row)
                data_type_str = row[1]
                _l = [ data_type_str ]
                datetime_precision = row[7]
                if data_type_str in ('bit', 'bit varying', 'character varying', 'character') and type(row[4]) is int:
                    _l.append( '({0})'.format(row[4]) )
                # elif data_type_str in ('numeric', 'decimal') and type(row[5]) is int or type(row[6]) is int:
                elif data_type_str in ('numeric', 'decimal'):
                    numeric_precision, numeric_scale = row[5], row[6]
                    _l.append( '(%d,%d)' % (numeric_precision, numeric_scale) )
                # interval types should need some revising
                elif data_type_str in ('interval', ):
                    _l.append( '(%d)' % datetime_precision )
                # time and timestamps have a somewhat different declarative syntax
                elif datetime_precision is int and (data_type_str.startswith('time') or data_type_str.startswith('timestamp')):
                    _in = 9 if data_type_str.startswith('timestamp') else 4 # the length of time and timestamp respectively
                    _l = [ data_type_str[:_in], '(%d)' % datetime_precision , data_type_str[_in:]]
                # append the current row to rwz
                if query_type == 'table_structure': rwz.append([row[0], "".join(_l), row[2], row[3] ])
                elif query_type == 'raw_table_structure': 
                    row.append("".join(_l))
                    rwz.append(row)
            # change r['rows']
            r['rows'] = rwz
            # change r['columns']
            if query_type == 'table_structure':
                r['columns'] = [ r['columns'][0], r['columns'][1], r['columns'][2], r['columns'][3] ]
            elif query_type == 'raw_table_structure': r['columns'].append('column_type')

        return r
        
    # queries with dissimilar implementations 
    elif conn_params['dialect'] == 'mysql':
        
        if query_type == 'describe_databases':
            conn_params['db'] = 'INFORMATION_SCHEMA';
            query = sql.stored_query(query_type, conn_params['dialect'])
            return sa.full_query(conn_params, query)
        
        else:
            return fns.http_500('query not yet implemented!')
    else:
        return fns.http_500('dialect not supported!')
Beispiel #8
0
def rpr_query(conn_params, query_type, get_data={}, post_data={}):
    '''
    Run queries that have to be generated on the fly. Most queries depends on get_data, 
    while some few depends on post_data

    get_data and post_data are gotten from request.GET and request.POST or form.cleaned_data
    '''
    # common queries that returns success state as a dict only
    no_return_queries = ('create_user', 'drop_user', 'create_db','create_table',
        'drop_table', 'empty_table', 'delete_row', 'create_column', 'delete_column',)
    
    if query_type in no_return_queries:
        conn_params['db'] = get_data['db'] if get_data.has_key('db') else conn_params['db']
        query_data = {}
        query_data.update(get_data, **post_data)
        q = sql.generate_query( query_type, conn_params['dialect'],query_data)
        result = sql.short_query(conn_params, q)
        return HttpResponse( json.dumps(result) )
    
    # specific queries with implementations similar to both dialects
    elif query_type == 'user_rpr':
        if conn_params['dialect'] == 'mysql':
            conn_params['db'] = 'mysql'
        r = sql.full_query(conn_params, 
            sql.stored_query(get_data['query'],conn_params['dialect']) )
        if type(r) == dict:
            r
        else:
            return fns.http_500(r)
        
    elif query_type in ('indexes', 'primary_keys', 'foreign_key_relation'):
        
        if conn_params['dialect'] == 'postgresql': conn_params['db'] = get_data['db']
        r = sql.full_query(conn_params,
            sql.generate_query(query_type, conn_params['dialect'], get_data)[0])
        return r
        
    elif query_type in ('get_single_row',):
        sub_q_data = {'tbl': get_data['tbl'],'db':get_data['db']}
        if get_data.has_key('schm'):
            sub_q_data['schm'] = get_data['schm']
        # generate where statement
        sub_q_data['where'] = ""
        for ind in range(len(post_data)):
            sub_q_data['where'] += post_data.keys()[ind].strip() + "=" 
            sub_q_data['where'] += post_data.values()[ind].strip()
            if ind != len(post_data) - 1: sub_q_data['where'] += ' AND '
        # retrieve and run queries
        conn_params['db'] = get_data['db']
        # assert False
        q = sql.generate_query(query_type, conn_params['dialect'], sub_q_data)
        r =  sql.full_query(conn_params, q[0])
        return r
        

    elif query_type in ('table_rpr', 'table_structure', 'raw_table_structure'):
        conn_params['db'] = get_data['db']
        sub_q_data = {'db': get_data['db'],}
        if get_data.has_key('tbl'):
            sub_q_data['tbl'] = get_data['tbl']
        if get_data.has_key('schm'):
            sub_q_data['schm'] = get_data['schm']
        # make query
        if conn_params['dialect'] == 'postgresql' and query_type == 'raw_table_structure':
            q = 'table_structure'
        else: q = query_type

        r = sql.full_query(conn_params,
            sql.generate_query(q, conn_params['dialect'], sub_q_data)[0] )
        # further needed processing
        if conn_params['dialect'] == 'postgresql' and query_type.count('table_structure'):
            rwz = []
            for tuple_row in r['rows']:
                row = list(tuple_row)
                _l = [ row[1] ]
                if row[1] in ('bit', 'bit varying', 'character varying', 'character') and type(row[4]) is int:
                    _l.append( '({0})'.format(row[4]) )
                elif row[1] in ('numeric', 'decimal') and type(row[5]) is int or type(row[6]) is int:
                    _l.append( '({0},{1})'.format(row[5], row[6]) )
                elif row[1] in ('interval', 'time with time zone', 'time without time zone',
                    'timestamp with time zone', 'timestamp without time zone') and type(row[7]) is int:
                    _l.append( '({0})'.format(row[7]) )
                # append the current row to rwz
                if query_type == 'table_structure': rwz.append([row[0], "".join(_l), row[2], row[3] ])
                elif query_type == 'raw_table_structure': 
                    row.append("".join(_l))
                    rwz.append(row)
            # change r['rows']
            r['rows'] = rwz
            # change r['columns']
            if query_type == 'table_structure':
                r['columns'] = [ r['columns'][0], r['columns'][1], r['columns'][2], r['columns'][3] ]
            elif query_type == 'raw_table_structure': r['columns'].append('column_type')

        return r
        
    elif query_type == 'browse_table':
        # initializations        
        sub_q_data = {'tbl': get_data['tbl'],'db':get_data['db']}
        sub_q_data['offset'] = get_data['offset'] if get_data.has_key('offset') else 0
        sub_q_data['limit'] = get_data['limit'] if get_data.has_key('limit') else getattr(settings, 'TT_MAX_ROW_COUNT', 100)
        for item in ['schm', 'sort_key', 'sort_dir']:
            if get_data.has_key(item): sub_q_data[item] = get_data[item]
        # retrieve and run queries
        conn_params['db'] = get_data['db']
        keys = rpr_query(conn_params, 'primary_keys', sub_q_data)
        count = sql.full_query(conn_params, 
            sql.generate_query('count_rows', conn_params['dialect'], sub_q_data)[0],
            )['rows']
        r = sql.full_query(conn_params,
            sql.generate_query(query_type, conn_params['dialect'], sub_q_data)[0]
            )
        # format and return data
        if type(r) == dict:
            r.update({'total_count': count[0][0], 'offset': sub_q_data['offset'],
                      'limit':sub_q_data['limit'], 'keys': keys})
            return r
        else:
            return fns.http_500(r)
        
    # queries that just asks formats and return result
    elif query_type in ('existing_tables',):
        query_data = {'db':get_data['db'],}
        if get_data.has_key('tbl'): query_data['tbl'] = get_data['tbl']
        if conn_params['dialect'] == 'postgresql':
            query_data['schm'] = get_data['schm']
            conn_params['db'] = query_data['db']
            
        q = sql.generate_query(query_type, conn_params['dialect'], query_data)
        r =  sql.full_query(conn_params,
            q[0])
        return r['rows']

        
    # queries with dissimilar implementations
    elif conn_params['dialect'] == 'postgresql':
            return fns.http_500('query ({query_type}) not implemented!'.format(query_type=query_type))
            
    elif conn_params['dialect'] == 'mysql':
        
        if query_type == 'describe_databases':
            conn_params['db'] = 'INFORMATION_SCHEMA';
            query = sql.stored_query(query_type, conn_params['dialect'])
            return sql.full_query(conn_params, query)
        
        else:
            return fns.http_500('query not yet implemented!')
    else:
        return fns.http_500('dialect not supported!')