28

For some reason, I want to dump a table from a database (sqlite3) in the form of a csv file. I'm using a python script with elixir (based on sqlalchemy) to modify the database. I was wondering if there is any way to dump the table I use to csv.

I've seen sqlalchemy serializer but it doesn't seem to be what I want. Am I doing it wrong? Should I call the sqlite3 python module after closing my sqlalchemy session to dump to a file instead? Or should I use something homemade?

Ben
  • 51,770
  • 36
  • 127
  • 149
tmoisan
  • 1,212
  • 2
  • 13
  • 27

9 Answers9

42

Modifying Peter Hansen's answer here a bit, to use SQLAlchemy instead of raw db access

import csv
outfile = open('mydump.csv', 'wb')
outcsv = csv.writer(outfile)
records = session.query(MyModel).all()
[outcsv.writerow([getattr(curr, column.name) for column in MyTable.__mapper__.columns]) for curr in records]
# or maybe use outcsv.writerows(records)

outfile.close()
Zitrax
  • 19,036
  • 20
  • 88
  • 110
RyanWilcox
  • 13,890
  • 1
  • 36
  • 60
  • 1
    Just a small remark: `outcsv.writerows(records)` will result in a `Error: sequence expected` – miku May 23 '11 at 23:34
  • 4
    Also you can get all columns at once by using the `__mapper__` [attribute](http://www.sqlalchemy.org/docs/orm/extensions/declarative.html#synopsis), like so: `[ outcsv.writerow([ getattr(curr, column.name) for column in MyModel.__mapper__.columns ]) for curr in records ]` – miku May 23 '11 at 23:46
  • 4
    To add an initial header row to describe the columns, use: `outcsv.writerow([column.name for column in MyModel.__mapper__.columns]) ` – bschwagg Sep 10 '15 at 04:47
  • 3
    note that for python3, the file write should be done in text mode, not binary: outfile = open('mydump.csv', 'w') – christok Mar 07 '21 at 19:15
30

There are numerous ways to achieve this, including a simple os.system() call to the sqlite3 utility if you have that installed, but here's roughly what I'd do from Python:

import sqlite3
import csv

con = sqlite3.connect('mydatabase.db')
outfile = open('mydump.csv', 'wb')
outcsv = csv.writer(outfile)

cursor = con.execute('select * from mytable')

# dump column titles (optional)
outcsv.writerow(x[0] for x in cursor.description)
# dump rows
outcsv.writerows(cursor.fetchall())

outfile.close()
Peter Hansen
  • 21,046
  • 5
  • 50
  • 72
19

I adapted the above examples to my sqlalchemy based code like this:

import csv
import sqlalchemy as sqAl

metadata = sqAl.MetaData()
engine = sqAl.create_engine('sqlite:///%s' % 'data.db')
metadata.bind = engine

mytable = sqAl.Table('sometable', metadata, autoload=True)
db_connection = engine.connect()

select = sqAl.sql.select([mytable])
result = db_connection.execute(select)

fh = open('data.csv', 'wb')
outcsv = csv.writer(fh)

outcsv.writerow(result.keys())
outcsv.writerows(result)

fh.close

This works for me with sqlalchemy 0.7.9. I suppose that this would work with all sqlalchemy table and result objects.

TNT
  • 3,392
  • 1
  • 24
  • 27
8

I know this is old, but i just had this problem and this is how i solved it

import pandas as pd
from sqlalchemy import create_engine

basedir = os.path.abspath(os.path.dirname(__file__))
sql_engine = create_engine(os.path.join('sqlite:///' + os.path.join(basedir, 'single_file_app.db')), echo=False)
results = pd.read_sql_query('select * from users',sql_engine)
results.to_csv(os.path.join(basedir, 'mydump2.csv'),index=False,sep=";")
shrewmouse
  • 5,338
  • 3
  • 38
  • 43
Manu
  • 81
  • 1
  • 1
6
with open('dump.csv', 'wb') as f:
    out = csv.writer(f)
    out.writerow(['id', 'description'])

    for item in session.query(Queue).all():
        out.writerow([item.id, item.description])

I found this to be useful if you don't mind hand-crafting your column labels.

Benjamin W.
  • 46,058
  • 19
  • 106
  • 116
michael g
  • 603
  • 7
  • 14
1
import csv

f = open('ratings.csv', 'w')
out = csv.writer(f)
out.writerow(['id', 'user_id', 'movie_id', 'rating'])

for item in db.query.all():
    out.writerow([item.username, item.username, item.movie_name, item.rating])
f.close()
1

I spent a lot of time searching for a solution to this problem and finally created something like this:

from sqlalchemy import inspect

with open(file_to_write, 'w') as file:
    out_csv = csv.writer(file, lineterminator='\n')

    columns = [column.name for column in inspect(Movies).columns][1:]
    out_csv.writerow(columns)

    session_3 = session_maker()

    extract_query = [getattr(Movies, col) for col in columns]
    for mov in session_3.query(*extract_query):
        out_csv.writerow(mov)

    session_3.close()

It creates a CSV file with column names and a dump of the entire "movies" table without "id" primary column.

shrewmouse
  • 5,338
  • 3
  • 38
  • 43
0

In a modular way: an example using slqalchemy with automap and mysql.

database.py:

from sqlalchemy.ext.automap import automap_base
from sqlalchemy.orm import Session
from sqlalchemy import create_engine

Base = automap_base()

engine = create_engine('mysql://user:pass@localhost:3306/database_name', echo=True)

Base.prepare(engine, reflect=True)

# Map the tables
State = Base.classes.states

session = Session(engine, autoflush=False)

export_to_csv.py:

from databases import *
import csv

def export():

    q = session.query(State)

    file = './data/states.csv'

    with open(file, 'w') as csvfile:
        outcsv = csv.writer(csvfile, delimiter=',',quotechar='"', quoting = csv.QUOTE_MINIMAL)

        header = State.__table__.columns.keys()

        outcsv.writerow(header)     

        for record in q.all():
            outcsv.writerow([getattr(record, c) for c in header ])

if __name__ == "__main__":
    export()

Results:

name,abv,country,is_state,is_lower48,slug,latitude,longitude,population,area Alaska,AK,US,y,n,alaska,61.370716,-152.404419,710231,571951.25 Alabama,AL,US,y,y,alabama,32.806671,-86.79113,4779736,50744.0 Arkansas,AR,US,y,y,arkansas,34.969704,-92.373123,2915918,52068.17 Arizona,AZ,US,y,y,arizona,33.729759,-111.431221,6392017,113634.57 California,CA,US,y,y,california,36.116203,-119.681564,37253956,155939.52 Colorado,CO,US,y,y,colorado,39.059811,-105.311104,5029196,103717.53 Connecticut,CT,US,y,y,connecticut,41.597782,-72.755371,3574097,4844.8 District of Columbia,DC,US,n,n,district-of-columbia,38.897438,-77.026817,601723,68.34 Delaware,DE,US,y,y,delaware,39.318523,-75.507141,897934,1953.56 Florida,FL,US,y,y,florida,27.766279,-81.686783,18801310,53926.82 Georgia,GA,US,y,y,georgia,33.040619,-83.643074,9687653,57906.14

Andre Araujo
  • 2,348
  • 2
  • 27
  • 41
0

A simple way to do it using pandas + sqlalchemy

import os
import pandas as pd
from sqlalchemy import create_engine, select
from sqlalchemy import MetaData, Table
from pathlib import Path   

def convert_to_csv(tablename, filename):
    engine = create_engine('sqlite:///your-file.sqlite')
    connection = engine.connect()
    
    metadata = MetaData()
    table = Table(tablename, metadata, autoload_with=engine)
    stmt = select(table)
    results = connection.execute(stmt).fetchall() # .fetchmany(size=10)

    filepath = Path(filename)  
    filepath.parent.mkdir(parents=True, exist_ok=True)  

    df = pd.DataFrame(results)
    df.to_csv(filepath, index=False)

    print(f'\n data has exported successfully into {os.getcwd()}/{filepath}\n')


convert_to_csv('your-table-name', '../your-path/your-new-file.csv')
Ricardo Canelas
  • 2,280
  • 26
  • 21