注意:mysql、redis等連接密碼需修改為相應值
Seata-Server
環境
- 版本:1.4.2
- OS: CentOS Linux release 7.5.1804 (Core)
- ip:192.168.1.78
Registry.conf
registry {
# file 、nacos 、eureka、redis、zk、consul、etcd3、sofa
type = "nacos"
nacos {
application = "seata-server"
serverAddr = "192.168.1.78:8848"
group = "SEATA_GROUP"
namespace = ""
cluster = "default"
username = ""
password = ""
}
}
config {
# file、nacos 、apollo、zk、consul、etcd3
type = "nacos"
nacos {
serverAddr = "192.168.1.78:8848"
namespace = ""
group = "SEATA_GROUP"
username = ""
password = ""
dataId = "seataServer.properties"
}
}
seataServer.properties
transport.type=TCP
transport.server=NIO
transport.heartbeat=true
transport.enableClientBatchSendRequest=true
transport.threadFactory.bossThreadPrefix=NettyBoss
transport.threadFactory.workerThreadPrefix=NettyServerNIOWorker
transport.threadFactory.serverExecutorThreadPrefix=NettyServerBizHandler
transport.threadFactory.shareBossWorker=false
transport.threadFactory.clientSelectorThreadPrefix=NettyClientSelector
transport.threadFactory.clientSelectorThreadSize=1
transport.threadFactory.clientWorkerThreadPrefix=NettyClientWorkerThread
transport.threadFactory.bossThreadSize=1
transport.threadFactory.workerThreadSize=default
transport.shutdown.wait=3
service.vgroupMapping.my_test_tx_group=default
#service.my_test_tx_group.grouplist=192.168.1.78:8091
service.default.grouplist=192.168.1.78:8091
service.enableDegrade=false
service.disableGlobalTransaction=false
client.rm.asyncCommitBufferLimit=10000
client.rm.lock.retryInterval=10
client.rm.lock.retryTimes=30
client.rm.lock.retryPolicyBranchRollbackOnConflict=true
client.rm.reportRetryCount=5
client.rm.tableMetaCheckEnable=false
client.rm.tableMetaCheckerInterval=60000
client.rm.sqlParserType=druid
client.rm.reportSuccessEnable=false
client.rm.sagaBranchRegisterEnable=false
client.rm.sagaJsonParser=fastjson
client.rm.tccActionInterceptorOrder=-2147482648
client.tm.commitRetryCount=5
client.tm.rollbackRetryCount=5
client.tm.defaultGlobalTransactionTimeout=60000
client.tm.degradeCheck=false
client.tm.degradeCheckAllowTimes=10
client.tm.degradeCheckPeriod=2000
client.tm.interceptorOrder=-2147482648
store.mode=db
store.lock.mode=file
#store.session.mode=file
store.session.mode=db
store.publicKey=
store.file.dir=file_store/data
store.file.maxBranchSessionSize=16384
store.file.maxGlobalSessionSize=512
store.file.fileWriteBufferCacheSize=16384
store.file.flushDiskMode=async
store.file.sessionReloadReadSize=100
store.db.datasource=druid
store.db.dbType=mysql
store.db.driverClassName=com.mysql.jdbc.Driver
store.db.url=jdbc:mysql://192.168.1.235:3306/seata?useUnicode=true&rewriteBatchedStatements=true
store.db.user=root
store.db.password=123456
store.db.minConn=5
store.db.maxConn=30
store.db.globalTable=global_table
store.db.branchTable=branch_table
store.db.distributedLockTable=distributed_lock
store.db.queryLimit=100
store.db.lockTable=lock_table
store.db.maxWait=5000
store.redis.mode=single
store.redis.single.host=192.168.1.78
store.redis.single.port=6379
store.redis.sentinel.masterName=
store.redis.sentinel.sentinelHosts=
store.redis.maxConn=10
store.redis.minConn=1
store.redis.maxTotal=100
store.redis.database=0
store.redis.password=123456
store.redis.queryLimit=100
server.recovery.committingRetryPeriod=1000
server.recovery.asynCommittingRetryPeriod=1000
server.recovery.rollbackingRetryPeriod=1000
server.recovery.timeoutRetryPeriod=1000
server.maxCommitRetryTimeout=-1
server.maxRollbackRetryTimeout=-1
server.rollbackRetryTimeoutUnlockEnable=false
server.distributedLockExpireTime=10000
client.undo.dataValidation=true
client.undo.logSerialization=jackson
client.undo.onlyCareUpdateColumns=true
server.undo.logSaveDays=7
server.undo.logDeletePeriod=86400000
client.undo.logTable=undo_log
client.undo.compress.enable=true
client.undo.compress.type=zip
client.undo.compress.threshold=64k
log.exceptionRate=100
transport.serialization=seata
transport.compressor=none
metrics.enabled=false
metrics.registryType=compact
metrics.exporterList=prometheus
metrics.exporterPrometheusPort=9898
tcc.fence.logTableName=tcc_fence_log
tcc.fence.cleanPeriod=1h
seata-server.sh
沒有修改過,1.4.2發布版本是什么樣就是什么樣
啟動命令
sh seata-server.sh -m db -h 192.168.1.78 -p 8091
Seata-Client
SeataClient是直接集成到我們的業務微服務中,SeataClient的原理是代理我們jdbc數據源,在應用程序和數據庫之間加了一層,通過添加的這一層來做事務管理。
下面虛擬了兩個業務微服務,一個叫seata-user-service,一個叫seata-car-service。
關於SeataServer和SeataClient的版本對應問題,當前暫未找到相關官方的說明文檔
環境
- SpringBoot:2.5.3
- SpringCloud:2020.0.3
- SpringCloudAlibaba:2021.1
pom
網上有一些文章說引入pom依賴之后,需要手動配置相關容器類,咱們使用的是高版本,引入 spring-cloud-starter-alibaba-seata 后,SpringBoot能夠為我們完成所有的自動配置,這一點需要注意,手動配置和自動配置不可同時存在,這點官網上面有說明。
<!-- SpringCloud 微服務 -->
<dependency>
<groupId>org.springframework.cloud</groupId>
<artifactId>spring-cloud-dependencies</artifactId>
<version>${spring-cloud.version}</version>
<type>pom</type>
<scope>import</scope>
</dependency>
<!-- SpringCloud Alibaba 微服務 -->
<dependency>
<groupId>com.alibaba.cloud</groupId>
<artifactId>spring-cloud-alibaba-dependencies</artifactId>
<version>${spring-cloud-alibaba.version}</version>
<type>pom</type>
<scope>import</scope>
</dependency>
<dependencies>
<!-- seata -->
<dependency>
<groupId>com.alibaba.cloud</groupId>
<artifactId>spring-cloud-starter-alibaba-seata</artifactId>
</dependency>
</dependencies>
seata-user-service
application.yml
重點關注兩個地方
- seata.registry:seataClient通過這個配置來找到SeataServer
- spring.cloud.alibaba.seata.tx-service-group:通過這個值來確定事務分組,當前配置的值my_test_tx_group,對應seataServer.properties中的service.vgroupMapping.my_test_tx_group=default
事務分組是為了邏輯上做高可用而設計的一個東西,從某種意義上來說本身就是一個非必須的東西,關於事務分組具體如何使用后續再整理文檔說明
server:
port: 9302
spring:
application:
name: seata-user-service
cloud:
nacos:
discovery:
# 服務注冊地址
server-addr: 192.168.1.78:8848
alibaba:
seata:
tx-service-group: my_test_tx_group
#logging:
# level:
# io.seata: debug
seata:
registry:
type: nacos
nacos:
server-addr: 192.168.1.78
group: SEATA_GROUP
cluster: default
seata-car-service
application.yml
server:
port: 9301
spring:
application:
name: seata-car-service
cloud:
nacos:
discovery:
# 服務注冊地址
server-addr: 192.168.1.78:8848
alibaba:
seata:
tx-service-group: my_test_tx_group
logging:
level:
io.seata: debug
seata:
registry:
type: nacos
nacos:
server-addr: 192.168.1.78:8848
group: SEATA_GROUP
cluster: default