Skip to content

Latest commit

 

History

History
358 lines (266 loc) · 12.2 KB

README.md

File metadata and controls

358 lines (266 loc) · 12.2 KB

datarouter-mysql

datarouter-mysql is an implementation of datarouter-storage that allows you to connect to a MySQL database.

Installation with Maven

<dependency>
	<groupId>io.datarouter</groupId>
	<artifactId>datarouter-mysql</artifactId>
	<version>0.0.126</version>
</dependency>

Installation with Datarouter

You can install this module by adding its plugin to the WebappBuilder.

.addPlugin(new DatarouterMysqlPluginBuilder()
		...
		.build()

Usage

Primary Key

This class represents the primary key of the MySQL table. Datarouter will define a PRIMARY KEY with the columns defined in this class.

package io.datarouter.example.docs.dataroutermysql;

import java.util.List;

import io.datarouter.model.field.Field;
import io.datarouter.model.field.imp.StringField;
import io.datarouter.model.field.imp.StringFieldKey;
import io.datarouter.model.key.primary.base.BaseRegularPrimaryKey;

public class MysqlExampleDatabeanKey extends BaseRegularPrimaryKey<MysqlExampleDatabeanKey>{

	private String id;

	public static class FieldKeys{
		public static final StringFieldKey id = new StringFieldKey("id");
	}

	public MysqlExampleDatabeanKey(){
	}

	public MysqlExampleDatabeanKey(String id){
		this.id = id;
	}

	@Override
	public List<Field<?>> getFields(){ // the list of columns of the PRIMARY KEY
		return List.of(new StringField(FieldKeys.id, id));
	}

}

You can define an AUTO_INCREMENT key column by specifying a FieldGeneratorType with value MANAGED on the field key.

public static final LongFieldKey id = new LongFieldKey("id")
		.withFieldGeneratorType(FieldGeneratorType.MANAGED);

When using composite primary keys, the ordering of the getFields method matters. The columns in the primary key will follow the same order.

Databean

This class represents a MySQL table. Each instance will be a row of that table. Besides its primary key, this one defines an INT column called someInt.

package io.datarouter.example.docs.dataroutermysql;

import java.util.List;
import java.util.function.Supplier;

import io.datarouter.model.databean.BaseDatabean;
import io.datarouter.model.field.Field;
import io.datarouter.model.field.imp.comparable.IntegerField;
import io.datarouter.model.field.imp.comparable.IntegerFieldKey;
import io.datarouter.model.serialize.fielder.BaseDatabeanFielder;

public class MysqlExampleDatabean extends BaseDatabean<MysqlExampleDatabeanKey,MysqlExampleDatabean>{

	private Integer someInt;

	private static class FieldKeys{
		private static final IntegerFieldKey someInt = new IntegerFieldKey("someInt");
	}

	public MysqlExampleDatabean(){
		super(new MysqlExampleDatabeanKey()); // it is required to initialize the key field of a databean
	}

	public MysqlExampleDatabean(String id, Integer someInt){
		super(new MysqlExampleDatabeanKey(id));
		this.someInt = someInt;
	}

	public static class MysqlExampleDatabeanFielder
	extends BaseDatabeanFielder<MysqlExampleDatabeanKey,MysqlExampleDatabean>{

		public MysqlExampleDatabeanFielder(){
			super(MysqlExampleDatabeanKey::new);
		}

		@Override
		public List<Field<?>> getNonKeyFields(MysqlExampleDatabean databean){
			return List.of(new IntegerField(FieldKeys.someInt, databean.someInt));
		}

	}

	@Override
	public Supplier<MysqlExampleDatabeanKey> getKeySupplier(){
		return MysqlExampleDatabeanKey::new;
	}

	public Integer getSomeInt(){
		return someInt;
	}

}

Datarouter will generate the following CREATE TABLE statement for such a databean.

create table testDatabase.ExampleDatabean (
 someInt int(11),
 id varchar(255) not null,
 primary key (id)
)engine INNODB, character set utf8mb4, collate utf8mb4_bin, comment 'created by test-server [2018-11-26T11:38:14.538-08:00]', row_format Dynamic;

Dao

Now that we have a databean class, we can create a node that will allow us to perform database queries. In the following router, we define a simple node. The node is configured to use a database client called mysqlClient. We will use a configuration file to tell datarouter how to connect this client to the database.

package io.datarouter.example.docs.dataroutermysql;

import io.datarouter.example.docs.dataroutermysql.MysqlExampleDatabean.MysqlExampleDatabeanFielder;
import io.datarouter.storage.Datarouter;
import io.datarouter.storage.client.ClientId;
import io.datarouter.storage.dao.BaseDao;
import io.datarouter.storage.node.factory.NodeFactory;
import io.datarouter.storage.node.op.combo.SortedMapStorage;
import jakarta.inject.Inject;
import jakarta.inject.Singleton;

@Singleton
public class MysqlExampleDao extends BaseDao{

	private static final ClientId MYSQL_CLIENT = ClientId.writer("mysqlClient", true);

	public final SortedMapStorage<MysqlExampleDatabeanKey,MysqlExampleDatabean> node;

	@Inject
	public MysqlExampleDao(Datarouter datarouter, NodeFactory nodeFactory){
		super(datarouter);
		node = nodeFactory.create(MYSQL_CLIENT, MysqlExampleDatabean::new, MysqlExampleDatabeanFielder::new)
				.buildAndRegister();
	}

}

Dependency injection configuration

For this example, we will be using Guice to inject dependencies. It's also possible to use other dependency injectors like Spring.

package io.datarouter.example.docs.dataroutermysql;

import java.util.Collections;

import io.datarouter.client.mysql.field.codec.factory.MysqlFieldCodecFactory;
import io.datarouter.client.mysql.field.codec.factory.StandardMysqlFieldCodecFactory;
import io.datarouter.inject.guice.BaseGuiceModule;
import io.datarouter.storage.config.guice.DatarouterStorageGuiceModule;
import io.datarouter.storage.dao.DaoClasses;

public class MysqlExampleGuiceModule extends BaseGuiceModule{

	@Override
	protected void configure(){
		// install the bindings of datarouter-storage
		install(new DatarouterStorageGuiceModule());
		// bind the standard codec factory - you can create your own if you want to define your own field types
		bindDefaultInstance(MysqlFieldCodecFactory.class, new StandardMysqlFieldCodecFactory(Collections.emptyMap()));
		// datarouter will use the application's name to look for configuration files
		// we register all the daos of our application here
		bind(DaoClasses.class).toInstance(new DaoClasses(MysqlExampleDao.class));
	}

}

Configuration files

Client configuration

There are two ways to configure the client options.

  1. Configuration in a datarouter-properties file.

The following configuration file tells datarouter that the mysqlClient client defined in the dao is a MySQL client that needs to connect to localhost:3306/testDatabase. This file needs to be located in /etc/datarouter/config/datarouter-test-app.properties, which is defined in the app's DatarouterProperties.java.

client.mysqlClient.type=mysql
client.mysqlClient.url=localhost:3306/testDatabase
client.mysqlClient.user=user
client.mysqlClient.password=password
  1. Configuration in the code

You can define the client options in the code using the MysqlClientOptionsBuilder and add the ClientOptionsBuilder to the app's WebappBuilder.

Properties properties = new MysqlClientOptionsBuilder(clientId)
		.setClientOptionType(clientId)
		.withUrl("localhost:3306/testDatabase")
		.withUser("user")
		.withPassword("password")
		.build();

Schema update configuration

Datarouter can create databases, tables and keep the schema up-to-date with what is defined in the code. There are two ways to configure the schema update options.

  1. Configuration in a schema-update.properties file.

To activate it, you will have to add this file at /etc/datarouter/config/schema-update.properties.

schemaUpdate.enable=true
schemaUpdate.execute.createDatabases=true
schemaUpdate.execute.createTables=true
schemaUpdate.execute.addColumns=true
schemaUpdate.execute.deleteColumns=true
schemaUpdate.execute.modifyColumns=true
schemaUpdate.execute.addIndexes=true
schemaUpdate.execute.dropIndexes=true
schemaUpdate.execute.modifyEngine=true
schemaUpdate.execute.modifyCharacterSetOrCollation=true
schemaUpdate.execute.modifyRowFormat=true
  1. Configuration in the code

You can define the schema update options in the code using the SchemaUpdateOptionsBuilder and add the implementation of SchemaUpdateOptionsFactory to the app's WebappBuilder.

Properties properties = new SchemaUpdateOptionsBuilder(true)
		.enableSchemaUpdateExecuteCreateDatabases()
		.enableSchemaUpdateExecuteCreateTables()
		.enableSchemaUpdateExecuteAddColumns()
		.enableSchemaUpdateExecuteDeleteColumns()
		.enableSchemaUpdateExecuteModifyColumns()
		.enableSchemaUpdateExecuteAddIndexes()
		.enableSchemaUpdateExecuteDropIndexes()
		.enableSchemaUpdateExecuteModifyEngine()
		.enableSchemaUpdateExecuteModifyCharacterSetOrCollation()
		.enableSchemaUpdateExecuteModifyRowFormat()
		.build();

On production environments, it is recommended to use schemaUpdate.print instead of schemaUpdate.execute. The ALTER TABLE statements will be logged and emailed instead of executed. The email will contains instruction on how to update the table schema in two format: sql query and pt-online-schema-update command. For some specific schema changes (like primary key change) two alternative instructions will be included: one to for production environments with additional steps for preserving backward compatibility and a second for other environments.

Individual clients could be exempt from schema updates by adding the schemaUpdate.execute.ignoreClients=client1,client2,... property or in the code with

new SchemaUpdateOptionsBuilder(true)
		...
		.withSchemaUpdateExecuteIgnoreClients(stringListOfClientsToIgnore)

Individual tables could be exempt from schema updates by adding the schemaUpdate.execute.ignoreTables=table1,table2,... property or in the code with

new SchemaUpdateOptionsBuilder(true)
		...
		.withSchemaUpdateExecuteIgnoreTables(stringListOfTablesToIgnore)

Application code

We have everything we need to start writing application code and database queries. The following main method will start the framework, write a databean to the MySQL table, and then read it.

package io.datarouter.example.docs.dataroutermysql;

import java.util.List;
import java.util.concurrent.ThreadLocalRandom;

import com.google.inject.Guice;
import com.google.inject.Injector;

import io.datarouter.storage.config.Config;
import io.datarouter.storage.config.PutMethod;
import io.datarouter.util.Require;
import io.datarouter.util.tuple.Range;

public class MysqlExampleMain{

	public static void main(String[] args){
		// create the Injector with our test module
		Injector injector = Guice.createInjector(List.of(new MysqlExampleGuiceModule()));
		// get an instance of our dao with the injector
		MysqlExampleDao dao = injector.getInstance(MysqlExampleDao.class);
		// instantiate a databean
		Integer someInt = ThreadLocalRandom.current().nextInt();
		MysqlExampleDatabean databean = new MysqlExampleDatabean("foo", someInt);
		// write the databean to the database, will issue an INSERT ... ON DUPLICATE KEY UPDATE by default
		dao.node.put(databean);
		// other put behaviors are available with PutMethod, this one will issue an INSERT IGNORE
		dao.node.put(databean, new Config().setPutMethod(PutMethod.INSERT_IGNORE));
		// read the databean using the same primary key
		MysqlExampleDatabean roundTripped = dao.node.get(new MysqlExampleDatabeanKey("foo"));
		// check that we were able to read the someInt column
		Require.isTrue(roundTripped.getSomeInt() == someInt);
		// databeans are equal if their keys are equal, they also sort by primary key
		Require.isTrue(roundTripped.equals(databean));
		// let's put another databean, with a different key
		Integer anotherInt = ThreadLocalRandom.current().nextInt();
		MysqlExampleDatabean anotherDatabean = new MysqlExampleDatabean("bar", anotherInt);
		dao.node.put(anotherDatabean);
		// you can fetch the rows given a range of primary keys, here, we fetch everything
		long sum = dao.node.scan(Range.everything()).streamInts(MysqlExampleDatabean::getSomeInt).sum();
		Require.isTrue(sum == someInt + anotherInt);
	}

}

Local Testing

To build this module locally, add mysql.properties to /etc/datarouter/test and schema-update.properties to /etc/datarouter/config.

License

This library is licensed under the Apache License, Version 2.0 - see LICENSE for details.