Catalog


Preface

It's been a month since I joined the new company , Finished the work at hand , A few days ago, I finally had time to study the code of the company's old projects . In the process of studying code , I found that... Was used in the project Spring Aop To realize the separation of reading and writing of database , In line with their love of learning ( I don't believe it myself ...) The character of the , Decided to write an example project to realize spring aop The effect of separation of reading and writing .

The deployment environment

database :MySql

Number of Libraries :2 individual , A master from

About mysql The master-slave environment of has been described in the article before deployment , No more details here , Reference resources 《 Hand to hand , How to be in windows System building mysql Master slave replication environment 》

Start the project

First , without doubt , Start building a SpringBoot engineering , And then in pom The following dependencies are introduced into the file :

<dependencies>
        <dependency>
            <groupId>com.alibaba</groupId>
            <artifactId>druid-spring-boot-starter</artifactId>
            <version>1.1.10</version>
        </dependency>
        <dependency>
            <groupId>org.mybatis.spring.boot</groupId>
            <artifactId>mybatis-spring-boot-starter</artifactId>
            <version>1.3.2</version>
        </dependency>
        <dependency>
            <groupId>tk.mybatis</groupId>
            <artifactId>mapper-spring-boot-starter</artifactId>
            <version>2.1.5</version>
        </dependency>
        <dependency>
            <groupId>mysql</groupId>
            <artifactId>mysql-connector-java</artifactId>
            <version>8.0.16</version>
        </dependency>
        <!--  Dynamic data sources   The required depend on  ### start-->
        <dependency>
            <groupId>org.springframework.boot</groupId>
            <artifactId>spring-boot-starter-jdbc</artifactId>
            <scope>provided</scope>
        </dependency>
        <dependency>
            <groupId>org.springframework.boot</groupId>
            <artifactId>spring-boot-starter-aop</artifactId>
            <scope>provided</scope>
        </dependency>
        <!--  Dynamic data sources   The required depend on  ### end-->
        <dependency>
            <groupId>org.springframework.boot</groupId>
            <artifactId>spring-boot-starter-web</artifactId>
        </dependency>
     <dependency>
            <groupId>org.projectlombok</groupId>
            <artifactId>lombok</artifactId>
            <optional>true</optional>
        </dependency>
        <dependency>
            <groupId>com.alibaba</groupId>
            <artifactId>fastjson</artifactId>
            <version>1.2.4</version>
        </dependency>
        <dependency>
            <groupId>org.springframework.boot</groupId>
            <artifactId>spring-boot-starter-test</artifactId>
            <scope>test</scope>
        </dependency>
        <dependency>
            <groupId>org.springframework.boot</groupId>
            <artifactId>spring-boot-starter-data-jpa</artifactId>
        </dependency>
    </dependencies>

Directory structure

After introducing basic dependencies , Sort out the directory structure , The skeleton of the completed project is as follows :

Build table

Create a table user, Execute... In the main library sql Statement is generating the corresponding table data from the database

DROP TABLE IF EXISTS `user`;
CREATE TABLE `user` (
  `user_id` bigint(20) NOT NULL COMMENT ' user id',
  `user_name` varchar(255) DEFAULT '' COMMENT ' User name ',
  `user_phone` varchar(50) DEFAULT '' COMMENT ' User's mobile phone ',
  `address` varchar(255) DEFAULT '' COMMENT ' address ',
  `weight` int(3) NOT NULL DEFAULT '1' COMMENT ' The weight , The bigger is the better ',
  `created_at` datetime NOT NULL DEFAULT CURRENT_TIMESTAMP COMMENT ' Creation time ',
  `updated_at` datetime DEFAULT CURRENT_TIMESTAMP ON UPDATE CURRENT_TIMESTAMP COMMENT ' Update time ',
  PRIMARY KEY (`user_id`)
) ENGINE=InnoDB DEFAULT CHARSET=utf8;
INSERT INTO `user` VALUES ('1196978513958141952', ' test 1', '18826334748', ' Haizhu District, Guangzhou ', '1', '2019-11-20 10:28:51', '2019-11-22 14:28:26');
INSERT INTO `user` VALUES ('1196978513958141953', ' test 2', '18826274230', ' Tianhe District, Guangzhou ', '2', '2019-11-20 10:29:37', '2019-11-22 14:28:14');
INSERT INTO `user` VALUES ('1196978513958141954', ' test 3', '18826273900', ' Tianhe District, Guangzhou ', '1', '2019-11-20 10:30:19', '2019-11-22 14:28:30');

Master and slave data source configuration

application.yml, The main information is the data source configuration of the master-slave database

server:
  port: 8001
spring:
  jackson:
   date-format: yyyy-MM-dd HH:mm:ss
   time-zone: GMT+8
  datasource:
    type: com.alibaba.druid.pool.DruidDataSource
    driver-class-name: com.mysql.cj.jdbc.Driver
    master:
      url: jdbc:mysql://127.0.0.1:3307/user?serverTimezone=Asia/Shanghai&useUnicode=true&characterEncoding=UTF-8&autoReconnect=true&failOverReadOnly=false&useSSL=false&zeroDateTimeBehavior=convertToNull&allowMultiQueries=true
      username: root
      password:
    slave:
      url: jdbc:mysql://127.0.0.1:3308/user?serverTimezone=Asia/Shanghai&useUnicode=true&characterEncoding=UTF-8&autoReconnect=true&failOverReadOnly=false&useSSL=false&zeroDateTimeBehavior=convertToNull&allowMultiQueries=true
      username: root
      password:

Because there are one master and one slave data sources , Let's use enumeration classes instead of , It is convenient for us to use

@Getter
public enum DynamicDataSourceEnum {
    MASTER("master"),
    SLAVE("slave");
    private String dataSourceName;
    DynamicDataSourceEnum(String dataSourceName) {
        this.dataSourceName = dataSourceName;
    }
}

Data source configuration information class DataSourceConfig, There are two data sources configured here ,masterDb and slaveDb

@Configuration
@MapperScan(basePackages = "com.xjt.proxy.mapper", sqlSessionTemplateRef = "sqlTemplate")
public class DataSourceConfig {
    
     //  Main library
      @Bean
      @ConfigurationProperties(prefix = "spring.datasource.master")
      public DataSource masterDb() {
  return DruidDataSourceBuilder.create().build();
      }
    /**
     *  Slave Library
     */
    @Bean
    @ConditionalOnProperty(prefix = "spring.datasource", name = "slave", matchIfMissing = true)
    @ConfigurationProperties(prefix = "spring.datasource.slave")
    public DataSource slaveDb() {
        return DruidDataSourceBuilder.create().build();
    }
    /**
     *  Master slave dynamic configuration
     */
    @Bean
    public DynamicDataSource dynamicDb(@Qualifier("masterDb") DataSource masterDataSource,
        @Autowired(required = false) @Qualifier("slaveDb") DataSource slaveDataSource) {
        DynamicDataSource dynamicDataSource = new DynamicDataSource();
        Map<Object, Object> targetDataSources = new HashMap<>();
        targetDataSources.put(DynamicDataSourceEnum.MASTER.getDataSourceName(), masterDataSource);
        if (slaveDataSource != null) {
            targetDataSources.put(DynamicDataSourceEnum.SLAVE.getDataSourceName(), slaveDataSource);
        }
        dynamicDataSource.setTargetDataSources(targetDataSources);
        dynamicDataSource.setDefaultTargetDataSource(masterDataSource);
        return dynamicDataSource;
    }
    @Bean
    public SqlSessionFactory sessionFactory(@Qualifier("dynamicDb") DataSource dynamicDataSource) throws Exception {
        SqlSessionFactoryBean bean = new SqlSessionFactoryBean();
        bean.setMapperLocations(
            new PathMatchingResourcePatternResolver().getResources("classpath*:mapper/*Mapper.xml"));
        bean.setDataSource(dynamicDataSource);
        return bean.getObject();
    }
    @Bean
    public SqlSessionTemplate sqlTemplate(@Qualifier("sessionFactory") SqlSessionFactory sqlSessionFactory) {
        return new SqlSessionTemplate(sqlSessionFactory);
    }
    @Bean(name = "dataSourceTx")
    public DataSourceTransactionManager dataSourceTx(@Qualifier("dynamicDb") DataSource dynamicDataSource) {
        DataSourceTransactionManager dataSourceTransactionManager = new DataSourceTransactionManager();
        dataSourceTransactionManager.setDataSource(dynamicDataSource);
        return dataSourceTransactionManager;
    }
}

Set the routing

The purpose of setting the route is to find the corresponding data source conveniently , We can use ThreadLocal Save the data source information to each thread , It's convenient for us to get

public class DataSourceContextHolder {
    private static final ThreadLocal<String> DYNAMIC_DATASOURCE_CONTEXT = new ThreadLocal<>();
    public static void set(String datasourceType) {
        DYNAMIC_DATASOURCE_CONTEXT.set(datasourceType);
    }
    public static String get() {
        return DYNAMIC_DATASOURCE_CONTEXT.get();
    }
    public static void clear() {
        DYNAMIC_DATASOURCE_CONTEXT.remove();
    }
}

Get route

public class DynamicDataSource extends AbstractRoutingDataSource {
    @Override
    protected Object determineCurrentLookupKey() {
        return DataSourceContextHolder.get();
    }
}

AbstractRoutingDataSource It's based on finding key Route to the corresponding data source , It internally maintains a set of target data sources , And routing key Mapping to the target data source , Offer based on key How to find the data source .

Data source comments

For the convenience of switching data sources , We can write an annotation , The annotation contains the enumeration values corresponding to the data source , The default is the main database ,

@Retention(RetentionPolicy.RUNTIME)
@Target(ElementType.METHOD)
@Documented
public @interface DataSourceSelector {
    DynamicDataSourceEnum value() default DynamicDataSourceEnum.MASTER;
    boolean clear() default true;
}

aop Switch data source

Come here ,aop Finally I can show up , So let's define one here aop class , Switch data sources for annotated methods , The specific code is as follows :

@Slf4j
@Aspect
@Order(value = 1)
@Component
public class DataSourceContextAop {
 @Around("@annotation(com.xjt.proxy.dynamicdatasource.DataSourceSelector)")
    public Object setDynamicDataSource(ProceedingJoinPoint pjp) throws Throwable {
        boolean clear = true;
        try {
            Method method = this.getMethod(pjp);
            DataSourceSelector dataSourceImport = method.getAnnotation(DataSourceSelector.class);
            clear = dataSourceImport.clear();
            DataSourceContextHolder.set(dataSourceImport.value().getDataSourceName());
            log.info("======== Switch the data source to :{}", dataSourceImport.value().getDataSourceName());
            return pjp.proceed();
        } finally {
            if (clear) {
                DataSourceContextHolder.clear();
            }
        }
    }
    private Method getMethod(JoinPoint pjp) {
        MethodSignature signature = (MethodSignature)pjp.getSignature();
        return signature.getMethod();
    }
}

To this step , Our preparation and configuration work is finished , Let's start to test the effect .

Write well first Service file , There are two ways to read and update ,

@Service
public class UserService {
    @Autowired
    private UserMapper userMapper;
    @DataSourceSelector(value = DynamicDataSourceEnum.SLAVE)
    public List<User> listUser() {
        List<User> users = userMapper.selectAll();
        return users;
    }
    @DataSourceSelector(value = DynamicDataSourceEnum.MASTER)
    public int update() {
        User user = new User();
        user.setUserId(Long.parseLong("1196978513958141952"));
        user.setUserName(" Revised name 2");
        return userMapper.updateByPrimaryKeySelective(user);
    }
    @DataSourceSelector(value = DynamicDataSourceEnum.SLAVE)
    public User find() {
        User user = new User();
        user.setUserId(Long.parseLong("1196978513958141952"));
        return userMapper.selectByPrimaryKey(user);
    }
}

As can be seen from the method notes , The way of reading is from the library , Update the method to go to the main library , The object of the update is userId by 1196978513958141953 The data of ,

Then we write a test class to test whether it can achieve the effect ,

@RunWith(SpringRunner.class)
@SpringBootTest
class UserServiceTest {
    @Autowired
    UserService userService;
    @Test
    void listUser() {
        List<User> users = userService.listUser();
        for (User user : users) {
            System.out.println(user.getUserId());
            System.out.println(user.getUserName());
            System.out.println(user.getUserPhone());
        }
    }
    @Test
    void update() {
        userService.update();
        User user = userService.find();
        System.out.println(user.getUserName());
    }
}

test result :

1、 Reading method

2、 Update method

After performing , Compare the database to find that the master and slave databases have modified the data , It shows that our separation of reading and writing is successful . Of course , The update method can point to from the library , In this way, only the data from the database will be modified , It doesn't involve the main database .

Be careful

The above test example is simple , But it also conforms to the normal read-write separation configuration . It's worth noting that , The purpose of the separation of reading and writing is to ease the writing library , That is, the pressure of the main reservoir , But it must be based on the principle of data consistency , It is to ensure that the data between the master and slave databases must be consistent . If a method involves writing logic , Then all database operations in this method should go to the main database .

Suppose that the data may not be synchronized to the slave database after the write operation , And then the read operation begins to execute , If the read program is still running from the library , Then there will be data inconsistency , This is what we don't allow .

Finally, I will send you the github Address , Interested students can have a look , Remember to give star Oh

Address :https://github.com/Taoxj/mysql-proxy

Refer to :

https://www.cnblogs.com/cjsblog/p/9712457.html