aboutsummaryrefslogtreecommitdiff
path: root/vendor/github.com/rubenv/sql-migrate/doc.go
blob: eb4ed8575a4edbba01f6a2259c598880a930faec (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
/*

SQL Schema migration tool for Go.

Key features:

	* Usable as a CLI tool or as a library
	* Supports SQLite, PostgreSQL, MySQL, MSSQL and Oracle databases (through gorp)
	* Can embed migrations into your application
	* Migrations are defined with SQL for full flexibility
	* Atomic migrations
	* Up/down migrations to allow rollback
	* Supports multiple database types in one project

Installation

To install the library and command line program, use the following:

	go get -v github.com/rubenv/sql-migrate/...

Command-line tool

The main command is called sql-migrate.

	$ sql-migrate --help
	usage: sql-migrate [--version] [--help] <command> [<args>]

	Available commands are:
		down      Undo a database migration
		new       Create a new migration
		redo      Reapply the last migration
		status    Show migration status
		up        Migrates the database to the most recent version available

Each command requires a configuration file (which defaults to dbconfig.yml, but can be specified with the -config flag). This config file should specify one or more environments:

	development:
		dialect: sqlite3
		datasource: test.db
		dir: migrations/sqlite3

	production:
		dialect: postgres
		datasource: dbname=myapp sslmode=disable
		dir: migrations/postgres
		table: migrations

The `table` setting is optional and will default to `gorp_migrations`.

The environment that will be used can be specified with the -env flag (defaults to development).

Use the --help flag in combination with any of the commands to get an overview of its usage:

	$ sql-migrate up --help
	Usage: sql-migrate up [options] ...

	  Migrates the database to the most recent version available.

	Options:

	  -config=config.yml   Configuration file to use.
	  -env="development"   Environment.
	  -limit=0             Limit the number of migrations (0 = unlimited).
	  -dryrun              Don't apply migrations, just print them.

The up command applies all available migrations. By contrast, down will only apply one migration by default. This behavior can be changed for both by using the -limit parameter.

The redo command will unapply the last migration and reapply it. This is useful during development, when you're writing migrations.

Use the status command to see the state of the applied migrations:

	$ sql-migrate status
	+---------------+-----------------------------------------+
	|   MIGRATION   |                 APPLIED                 |
	+---------------+-----------------------------------------+
	| 1_initial.sql | 2014-09-13 08:19:06.788354925 +0000 UTC |
	| 2_record.sql  | no                                      |
	+---------------+-----------------------------------------+

MySQL Caveat

If you are using MySQL, you must append ?parseTime=true to the datasource configuration. For example:

	production:
		dialect: mysql
		datasource: root@/dbname?parseTime=true
		dir: migrations/mysql
		table: migrations

See https://github.com/go-sql-driver/mysql#parsetime for more information.

Library

Import sql-migrate into your application:

	import "github.com/rubenv/sql-migrate"

Set up a source of migrations, this can be from memory, from a set of files or from bindata (more on that later):

	// Hardcoded strings in memory:
	migrations := &migrate.MemoryMigrationSource{
		Migrations: []*migrate.Migration{
			&migrate.Migration{
				Id:   "123",
				Up:   []string{"CREATE TABLE people (id int)"},
				Down: []string{"DROP TABLE people"},
			},
		},
	}

	// OR: Read migrations from a folder:
	migrations := &migrate.FileMigrationSource{
		Dir: "db/migrations",
	}

	// OR: Use migrations from bindata:
	migrations := &migrate.AssetMigrationSource{
		Asset:    Asset,
		AssetDir: AssetDir,
		Dir:      "migrations",
	}

Then use the Exec function to upgrade your database:

	db, err := sql.Open("sqlite3", filename)
	if err != nil {
		// Handle errors!
	}

	n, err := migrate.Exec(db, "sqlite3", migrations, migrate.Up)
	if err != nil {
		// Handle errors!
	}
	fmt.Printf("Applied %d migrations!\n", n)

Note that n can be greater than 0 even if there is an error: any migration that succeeded will remain applied even if a later one fails.

The full set of capabilities can be found in the API docs below.

Writing migrations

Migrations are defined in SQL files, which contain a set of SQL statements. Special comments are used to distinguish up and down migrations.

	-- +migrate Up
	-- SQL in section 'Up' is executed when this migration is applied
	CREATE TABLE people (id int);


	-- +migrate Down
	-- SQL section 'Down' is executed when this migration is rolled back
	DROP TABLE people;

You can put multiple statements in each block, as long as you end them with a semicolon (;).

If you have complex statements which contain semicolons, use StatementBegin and StatementEnd to indicate boundaries:

	-- +migrate Up
	CREATE TABLE people (id int);

	-- +migrate StatementBegin
	CREATE OR REPLACE FUNCTION do_something()
	returns void AS $$
	DECLARE
	  create_query text;
	BEGIN
	  -- Do something here
	END;
	$$
	language plpgsql;
	-- +migrate StatementEnd

	-- +migrate Down
	DROP FUNCTION do_something();
	DROP TABLE people;

The order in which migrations are applied is defined through the filename: sql-migrate will sort migrations based on their name. It's recommended to use an increasing version number or a timestamp as the first part of the filename.

Normally each migration is run within a transaction in order to guarantee that it is fully atomic. However some SQL commands (for example creating an index concurrently in PostgreSQL) cannot be executed inside a transaction. In order to execute such a command in a migration, the migration can be run using the notransaction option:

	-- +migrate Up notransaction
	CREATE UNIQUE INDEX people_unique_id_idx CONCURRENTLY ON people (id);

	-- +migrate Down
	DROP INDEX people_unique_id_idx;

Embedding migrations with packr

If you like your Go applications self-contained (that is: a single binary): use packr (https://github.com/gobuffalo/packr) to embed the migration files.

Just write your migration files as usual, as a set of SQL files in a folder.

Use the PackrMigrationSource in your application to find the migrations:

	migrations := &migrate.PackrMigrationSource{
		Box: packr.NewBox("./migrations"),
	}

If you already have a box and would like to use a subdirectory:

	migrations := &migrate.PackrMigrationSource{
		Box: myBox,
		Dir: "./migrations",
	}

Embedding migrations with bindata

As an alternative, but slightly less maintained, you can use bindata (https://github.com/shuLhan/go-bindata) to embed the migration files.

Just write your migration files as usual, as a set of SQL files in a folder.

Then use bindata to generate a .go file with the migrations embedded:

	go-bindata -pkg myapp -o bindata.go db/migrations/

The resulting bindata.go file will contain your migrations. Remember to regenerate your bindata.go file whenever you add/modify a migration (go generate will help here, once it arrives).

Use the AssetMigrationSource in your application to find the migrations:

	migrations := &migrate.AssetMigrationSource{
		Asset:    Asset,
		AssetDir: AssetDir,
		Dir:      "db/migrations",
	}

Both Asset and AssetDir are functions provided by bindata.

Then proceed as usual.

Extending

Adding a new migration source means implementing MigrationSource.

	type MigrationSource interface {
		FindMigrations() ([]*Migration, error)
	}

The resulting slice of migrations will be executed in the given order, so it should usually be sorted by the Id field.
*/
package migrate