diff --git a/.git-blame-ignore-revs b/.git-blame-ignore-revs index 81c802cb3..8fec79999 100644 --- a/.git-blame-ignore-revs +++ b/.git-blame-ignore-revs @@ -1,2 +1,5 @@ # reformat with scalafmt 3.8.2 6e52790fdcaa57dd38a88c547739c8b95d8fb04e + +# Scala Steward: Reformat with scalafmt 3.8.4 +bf1dd91ba61a569a4d91843418fa13ff11771a09 diff --git a/.scalafmt.conf b/.scalafmt.conf index 7b38c3ae0..37551d4b2 100644 --- a/.scalafmt.conf +++ b/.scalafmt.conf @@ -1,4 +1,4 @@ -version = 3.8.3 +version = 3.8.4 project.git = true diff --git a/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_4_1_0.scala b/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_4_1_0.scala index 36da54971..afc868537 100644 --- a/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_4_1_0.scala +++ b/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_4_1_0.scala @@ -153,20 +153,20 @@ private[journal] object Batch_4_1_0 { } final case class Appends( - offset: Offset, - records: NonEmptyList[ActionRecord[Action.Append]], + offset: Offset, + records: NonEmptyList[ActionRecord[Action.Append]], ) extends Batch_4_1_0 final case class Delete( - offset: Offset, - to: DeleteTo, - origin: Option[Origin], - version: Option[Version], + offset: Offset, + to: DeleteTo, + origin: Option[Origin], + version: Option[Version], ) extends Batch_4_1_0 final case class Purge( - offset: Offset, - origin: Option[Origin], - version: Option[Version], + offset: Offset, + origin: Option[Origin], + version: Option[Version], ) extends Batch_4_1_0 } diff --git a/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_4_1_0_Alternative_with_Vector.scala b/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_4_1_0_Alternative_with_Vector.scala index c6d827d68..affb590a0 100644 --- a/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_4_1_0_Alternative_with_Vector.scala +++ b/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_4_1_0_Alternative_with_Vector.scala @@ -95,20 +95,20 @@ private[journal] object Batch_4_1_0_Alternative_with_Vector { } final case class Appends( - offset: Offset, - records: NonEmptyList[ActionRecord[Action.Append]], + offset: Offset, + records: NonEmptyList[ActionRecord[Action.Append]], ) extends Batch_4_1_0_Alternative_with_Vector final case class Delete( - offset: Offset, - to: DeleteTo, - origin: Option[Origin], - version: Option[Version], + offset: Offset, + to: DeleteTo, + origin: Option[Origin], + version: Option[Version], ) extends Batch_4_1_0_Alternative_with_Vector final case class Purge( - offset: Offset, - origin: Option[Origin], - version: Option[Version], + offset: Offset, + origin: Option[Origin], + version: Option[Version], ) extends Batch_4_1_0_Alternative_with_Vector } diff --git a/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_4_1_2.scala b/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_4_1_2.scala index 5d1645eb7..18b567d01 100644 --- a/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_4_1_2.scala +++ b/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_4_1_2.scala @@ -153,20 +153,20 @@ private[journal] object Batch_4_1_2 { } final case class Appends( - offset: Offset, - records: NonEmptyList[ActionRecord[Action.Append]], + offset: Offset, + records: NonEmptyList[ActionRecord[Action.Append]], ) extends Batch_4_1_2 final case class Delete( - offset: Offset, - to: DeleteTo, - origin: Option[Origin], - version: Option[Version], + offset: Offset, + to: DeleteTo, + origin: Option[Origin], + version: Option[Version], ) extends Batch_4_1_2 final case class Purge( - offset: Offset, - origin: Option[Origin], // used only for logging - version: Option[Version], // used only for logging + offset: Offset, + origin: Option[Origin], // used only for logging + version: Option[Version], // used only for logging ) extends Batch_4_1_2 } diff --git a/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_Alternative_with_Aggressive_Reshuffling.scala b/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_Alternative_with_Aggressive_Reshuffling.scala index 432e125a4..8c6d76d21 100644 --- a/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_Alternative_with_Aggressive_Reshuffling.scala +++ b/benchmark/src/test/scala/com/evolutiongaming/kafka/journal/replicator/Batch_Alternative_with_Aggressive_Reshuffling.scala @@ -28,9 +28,9 @@ private[journal] object Batch_Alternative_with_Aggressive_Reshuffling { } private final case class State( - private val purge: Option[Purge] = None, - private val appends: Option[Appends] = None, - private val delete: Option[Delete] = None, + private val purge: Option[Purge] = None, + private val appends: Option[Appends] = None, + private val delete: Option[Delete] = None, ) { // Expects records to be provided in reversed order, e.g., youngest first private def handle: ActionRecord[Action] => State = { @@ -117,20 +117,20 @@ private[journal] object Batch_Alternative_with_Aggressive_Reshuffling { } final case class Appends( - offset: Offset, - records: NonEmptyList[ActionRecord[Action.Append]], + offset: Offset, + records: NonEmptyList[ActionRecord[Action.Append]], ) extends Batch_Alternative_with_Aggressive_Reshuffling final case class Delete( - offset: Offset, - to: DeleteTo, - origin: Option[Origin], - version: Option[Version], + offset: Offset, + to: DeleteTo, + origin: Option[Origin], + version: Option[Version], ) extends Batch_Alternative_with_Aggressive_Reshuffling final case class Purge( - offset: Offset, - origin: Option[Origin], - version: Option[Version], + offset: Offset, + origin: Option[Origin], + version: Option[Version], ) extends Batch_Alternative_with_Aggressive_Reshuffling } diff --git a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/CassandraConsistencyConfig.scala b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/CassandraConsistencyConfig.scala index e9d16ba08..31aa18c2b 100644 --- a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/CassandraConsistencyConfig.scala +++ b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/CassandraConsistencyConfig.scala @@ -5,8 +5,8 @@ import pureconfig.ConfigReader import pureconfig.generic.semiauto.deriveReader private[journal] final case class CassandraConsistencyConfig( - read: CassandraConsistencyConfig.Read = CassandraConsistencyConfig.Read.default, - write: CassandraConsistencyConfig.Write = CassandraConsistencyConfig.Write.default, + read: CassandraConsistencyConfig.Read = CassandraConsistencyConfig.Read.default, + write: CassandraConsistencyConfig.Write = CassandraConsistencyConfig.Write.default, ) private[journal] object CassandraConsistencyConfig { diff --git a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/CassandraSync.scala b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/CassandraSync.scala index dfa0222f7..3fb917797 100644 --- a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/CassandraSync.scala +++ b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/CassandraSync.scala @@ -23,9 +23,9 @@ private[journal] object CassandraSync { def apply[F[_]](implicit F: CassandraSync[F]): CassandraSync[F] = F def apply[F[_]: Temporal: CassandraSession]( - keyspace: KeyspaceConfig, - table: String, - origin: Option[Origin], + keyspace: KeyspaceConfig, + table: String, + origin: Option[Origin], ): CassandraSync[F] = { val autoCreate = if (keyspace.autoCreate) AutoCreate.Table else AutoCreate.None @@ -33,10 +33,10 @@ private[journal] object CassandraSync { } def apply[F[_]: Temporal: CassandraSession]( - keyspace: String, - table: String, - autoCreate: AutoCreate, - metadata: Option[String], + keyspace: String, + table: String, + autoCreate: AutoCreate, + metadata: Option[String], ): CassandraSync[F] = { new CassandraSync[F] { @@ -71,9 +71,9 @@ private[journal] object CassandraSync { * @see [[com.evolutiongaming.cassandra.sync.CassandraSync]] for more details. */ def of[F[_]: Temporal: CassandraSession]( - keyspace: KeyspaceConfig, - table: String, - origin: Option[Origin], + keyspace: KeyspaceConfig, + table: String, + origin: Option[Origin], ): F[CassandraSync[F]] = { for { diff --git a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/CreateTables.scala b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/CreateTables.scala index 6449427e1..cd66c0b30 100644 --- a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/CreateTables.scala +++ b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/CreateTables.scala @@ -23,7 +23,7 @@ private[journal] object CreateTables { self => def apply[F[_]](implicit F: CreateTables[F]): CreateTables[F] = F def apply[F[_]: Monad: CassandraCluster: CassandraSession: CassandraSync]( - log: Log[F], + log: Log[F], ): CreateTables[F] = new CreateTables[F] { def apply(keyspace: String, tables: Nel[Table]) = { diff --git a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/KeyspaceConfig.scala b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/KeyspaceConfig.scala index 2c4e163c0..c007a4666 100644 --- a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/KeyspaceConfig.scala +++ b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/KeyspaceConfig.scala @@ -5,9 +5,9 @@ import pureconfig.ConfigReader import pureconfig.generic.semiauto.deriveReader private[journal] final case class KeyspaceConfig( - name: String = "journal", - replicationStrategy: ReplicationStrategyConfig = ReplicationStrategyConfig.Default, - autoCreate: Boolean = true, + name: String = "journal", + replicationStrategy: ReplicationStrategyConfig = ReplicationStrategyConfig.Default, + autoCreate: Boolean = true, ) private[journal] object KeyspaceConfig { diff --git a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/MigrateSchema.scala b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/MigrateSchema.scala index 7690af5ec..446000513 100644 --- a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/MigrateSchema.scala +++ b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/MigrateSchema.scala @@ -45,10 +45,10 @@ private[journal] object MigrateSchema { * The instance of schema migrator. */ def forSettingKey[F[_]: MonadThrow]( - cassandraSync: CassandraSync[F], - settings: Settings[F], - settingKey: String, - migrations: Nel[String], + cassandraSync: CassandraSync[F], + settings: Settings[F], + settingKey: String, + migrations: Nel[String], ): MigrateSchema[F] = new MigrateSchema[F] { def setVersion(version: Int): F[Unit] = diff --git a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/SettingStatements.scala b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/SettingStatements.scala index 70d1bd5cc..d028d88a2 100644 --- a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/SettingStatements.scala +++ b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/SettingStatements.scala @@ -56,8 +56,8 @@ private[journal] object SettingStatements { object Select { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Read, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[Select[F]] = { val query = s"SELECT value, timestamp, origin FROM ${name.toCql} WHERE key = ?" @@ -111,8 +111,8 @@ private[journal] object SettingStatements { object Insert { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[Insert[F]] = { val query = s"INSERT INTO ${name.toCql} (key, value, timestamp, origin) VALUES (?, ?, ?, ?)" @@ -135,8 +135,8 @@ private[journal] object SettingStatements { object Delete { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[Delete[F]] = { val query = s"DELETE FROM ${name.toCql} WHERE key = ?" diff --git a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/SettingsCassandra.scala b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/SettingsCassandra.scala index 703cd95e0..7f0c9168b 100644 --- a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/SettingsCassandra.scala +++ b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/cassandra/SettingsCassandra.scala @@ -12,8 +12,8 @@ import com.evolutiongaming.sstream.Stream private[journal] object SettingsCassandra { def apply[F[_]: Monad: Clock]( - statements: Statements[F], - origin: Option[Origin], + statements: Statements[F], + origin: Option[Origin], ): Settings[F] = new Settings[F] { def get(key: K): F[Option[Setting]] = { @@ -46,9 +46,9 @@ private[journal] object SettingsCassandra { } def of[F[_]: Monad: Parallel: Clock: CassandraSession]( - table: TableName, - origin: Option[Origin], - consistencyConfig: CassandraConsistencyConfig, + table: TableName, + origin: Option[Origin], + consistencyConfig: CassandraConsistencyConfig, ): F[Settings[F]] = { for { statements <- Statements.of[F](table, consistencyConfig) @@ -58,16 +58,16 @@ private[journal] object SettingsCassandra { } final case class Statements[F[_]]( - select: SettingStatements.Select[F], - insert: SettingStatements.Insert[F], - all: SettingStatements.All[F], - delete: SettingStatements.Delete[F], + select: SettingStatements.Select[F], + insert: SettingStatements.Insert[F], + all: SettingStatements.All[F], + delete: SettingStatements.Delete[F], ) object Statements { def of[F[_]: Monad: Parallel: CassandraSession]( - table: TableName, - consistencyConfig: CassandraConsistencyConfig, + table: TableName, + consistencyConfig: CassandraConsistencyConfig, ): F[Statements[F]] = { val statements = ( diff --git a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/CassandraCluster.scala b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/CassandraCluster.scala index 50e8c7647..828ec2cd0 100644 --- a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/CassandraCluster.scala +++ b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/CassandraCluster.scala @@ -20,8 +20,8 @@ object CassandraCluster { def apply[F[_]](implicit F: CassandraCluster[F]): CassandraCluster[F] = F def apply[F[_]: Async: Parallel: FromGFuture]( - cluster: scassandra.CassandraCluster[F], - retries: Int, + cluster: scassandra.CassandraCluster[F], + retries: Int, ): CassandraCluster[F] = new CassandraCluster[F] { def session = { @@ -43,9 +43,9 @@ object CassandraCluster { } def make[F[_]: Async: Parallel: FromGFuture]( - config: CassandraConfig, - cassandraClusterOf: CassandraClusterOf[F], - retries: Int, + config: CassandraConfig, + cassandraClusterOf: CassandraClusterOf[F], + retries: Int, ): Resource[F, CassandraCluster[F]] = { for { diff --git a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/CassandraSession.scala b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/CassandraSession.scala index 1f445a0f7..94e8852ae 100644 --- a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/CassandraSession.scala +++ b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/CassandraSession.scala @@ -32,16 +32,16 @@ object CassandraSession { def apply[F[_]](implicit F: CassandraSession[F]): CassandraSession[F] = F def apply[F[_]]( - session: CassandraSession[F], - retries: Int, - trace: Boolean = false, + session: CassandraSession[F], + retries: Int, + trace: Boolean = false, ): CassandraSession[F] = { val retryPolicy = new LoggingRetryPolicy(NextHostRetryPolicy(retries)) session.configured(retryPolicy, trace) } private def apply[F[_]: Async: FromGFuture]( - session: scassandra.CassandraSession[F], + session: scassandra.CassandraSession[F], ): CassandraSession[F] = { new CassandraSession[F] { @@ -60,7 +60,7 @@ object CassandraSession { } def make[F[_]: Async: Parallel: FromGFuture]( - session: scassandra.CassandraSession[F], + session: scassandra.CassandraSession[F], ): Resource[F, CassandraSession[F]] = { apply[F](session) .enhanceError @@ -70,8 +70,8 @@ object CassandraSession { implicit class CassandraSessionOps[F[_]](val self: CassandraSession[F]) extends AnyVal { def configured( - retryPolicy: RetryPolicy, - trace: Boolean, + retryPolicy: RetryPolicy, + trace: Boolean, ): CassandraSession[F] = new CassandraSession[F] { def prepare(query: String) = { diff --git a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ResultSet.scala b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ResultSet.scala index 5210c14b7..0308add90 100644 --- a/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ResultSet.scala +++ b/cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ResultSet.scala @@ -25,9 +25,9 @@ object ResultSet { } def apply[F[_]: Spawn, A]( - fetch: F[Unit], - fetched: F[Boolean], - next: F[List[A]], + fetch: F[Unit], + fetched: F[Boolean], + next: F[List[A]], ): Stream[F, A] = new Stream[F, A] { def foldWhileM[L, R](l: L)(f: (L, A) => F[Either[L, R]]) = { diff --git a/circe/core/src/main/scala/com/evolutiongaming/kafka/journal/circe/Instances.scala b/circe/core/src/main/scala/com/evolutiongaming/kafka/journal/circe/Instances.scala index 8e43b3ba3..7f5a1a96a 100644 --- a/circe/core/src/main/scala/com/evolutiongaming/kafka/journal/circe/Instances.scala +++ b/circe/core/src/main/scala/com/evolutiongaming/kafka/journal/circe/Instances.scala @@ -17,12 +17,12 @@ import java.nio.charset.StandardCharsets object Instances { implicit def kafkaWrite[F[_]: MonadThrowable]( - implicit payloadJsonToBytes: ToBytes[F, PayloadJson[Json]], + implicit payloadJsonToBytes: ToBytes[F, PayloadJson[Json]], ): KafkaWrite[F, Json] = KafkaWrite.writeJson(EventJsonPayloadAndType(_, PayloadType.Json), payloadJsonToBytes) implicit def kafkaRead[F[_]: MonadThrowable]( - implicit payloadJsonFromBytes: FromBytes[F, PayloadJson[Json]], + implicit payloadJsonFromBytes: FromBytes[F, PayloadJson[Json]], ): KafkaRead[F, Json] = KafkaRead.readJson(payloadJsonFromBytes, (json: EventJsonPayloadAndType[Json]) => json.payload.pure[F]) diff --git a/core/src/main/scala/com/evolutiongaming/kafka/journal/FromBytes.scala b/core/src/main/scala/com/evolutiongaming/kafka/journal/FromBytes.scala index 654d02ab5..99c17722b 100644 --- a/core/src/main/scala/com/evolutiongaming/kafka/journal/FromBytes.scala +++ b/core/src/main/scala/com/evolutiongaming/kafka/journal/FromBytes.scala @@ -44,8 +44,8 @@ object FromBytes { } def fromReads[F[_]: Monad: FromJsResult, A]( - implicit reads: Reads[A], - decode: JsonCodec.Decode[F], + implicit reads: Reads[A], + decode: JsonCodec.Decode[F], ): FromBytes[F, A] = { bytes => for { a <- decode.fromBytes(bytes) diff --git a/core/src/main/scala/com/evolutiongaming/kafka/journal/JournalError.scala b/core/src/main/scala/com/evolutiongaming/kafka/journal/JournalError.scala index ca51c887f..0c8169b72 100644 --- a/core/src/main/scala/com/evolutiongaming/kafka/journal/JournalError.scala +++ b/core/src/main/scala/com/evolutiongaming/kafka/journal/JournalError.scala @@ -3,8 +3,8 @@ package com.evolutiongaming.kafka.journal import cats.syntax.all.* final case class JournalError( - msg: String, - cause: Option[Throwable] = None, + msg: String, + cause: Option[Throwable] = None, ) extends RuntimeException(msg, cause.orNull) object JournalError { diff --git a/core/src/main/scala/com/evolutiongaming/kafka/journal/eventual/EventualPayloadAndType.scala b/core/src/main/scala/com/evolutiongaming/kafka/journal/eventual/EventualPayloadAndType.scala index 0a9d6f94b..67f1efaf0 100644 --- a/core/src/main/scala/com/evolutiongaming/kafka/journal/eventual/EventualPayloadAndType.scala +++ b/core/src/main/scala/com/evolutiongaming/kafka/journal/eventual/EventualPayloadAndType.scala @@ -22,8 +22,8 @@ import scodec.bits.ByteVector * should be treated, i.e. if it should be parsed as JSON. */ final case class EventualPayloadAndType( - payload: Either[String, ByteVector], - payloadType: PayloadType, + payload: Either[String, ByteVector], + payloadType: PayloadType, ) object EventualPayloadAndType { diff --git a/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ForkJoinPoolOf.scala b/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ForkJoinPoolOf.scala index 8b2907464..4659143d2 100644 --- a/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ForkJoinPoolOf.scala +++ b/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ForkJoinPoolOf.scala @@ -9,8 +9,8 @@ import java.util.concurrent.ForkJoinPool.ForkJoinWorkerThreadFactory private[journal] object ForkJoinPoolOf { def apply[F[_]: Sync]( - name: String, - parallelism: Int, + name: String, + parallelism: Int, ): Resource[F, ForkJoinPool] = { val threadFactory = ForkJoinPool.defaultForkJoinWorkerThreadFactory.withPrefix(name) diff --git a/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ScheduledExecutorServiceOf.scala b/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ScheduledExecutorServiceOf.scala index 55eededfe..931c5e7c8 100644 --- a/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ScheduledExecutorServiceOf.scala +++ b/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ScheduledExecutorServiceOf.scala @@ -8,8 +8,8 @@ import java.util.concurrent.{Executors as ExecutorsJ, ScheduledExecutorService, private[journal] object ScheduledExecutorServiceOf { def apply[F[_]: Sync]( - parallelism: Int, - threadFactory: ThreadFactory, + parallelism: Int, + threadFactory: ThreadFactory, ): Resource[F, ScheduledExecutorService] = { val result = for { diff --git a/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ThreadFactoryOf.scala b/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ThreadFactoryOf.scala index 58d48dd0c..6b08586a3 100644 --- a/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ThreadFactoryOf.scala +++ b/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ThreadFactoryOf.scala @@ -8,8 +8,8 @@ import java.util.concurrent.{Executors as ExecutorsJ, ThreadFactory} private[journal] object ThreadFactoryOf { def apply[F[_]: Sync]( - prefix: String, - uncaughtExceptionHandler: Thread.UncaughtExceptionHandler = UncaughtExceptionHandler.default, + prefix: String, + uncaughtExceptionHandler: Thread.UncaughtExceptionHandler = UncaughtExceptionHandler.default, ): F[ThreadFactory] = { for { diff --git a/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ThreadPoolOf.scala b/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ThreadPoolOf.scala index d806f35e5..4267dd94a 100644 --- a/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ThreadPoolOf.scala +++ b/core/src/main/scala/com/evolutiongaming/kafka/journal/execution/ThreadPoolOf.scala @@ -9,10 +9,10 @@ import scala.concurrent.duration.* private[journal] object ThreadPoolOf { def apply[F[_]: Sync]( - minSize: Int, - maxSize: Int, - threadFactory: ThreadFactory, - keepAlive: FiniteDuration = 5.minute, + minSize: Int, + maxSize: Int, + threadFactory: ThreadFactory, + keepAlive: FiniteDuration = 5.minute, ): Resource[F, ThreadPoolExecutor] = { val result = for { diff --git a/core/src/main/scala/com/evolutiongaming/kafka/journal/util/ActorSystemOf.scala b/core/src/main/scala/com/evolutiongaming/kafka/journal/util/ActorSystemOf.scala index 9d324df57..3f4928646 100644 --- a/core/src/main/scala/com/evolutiongaming/kafka/journal/util/ActorSystemOf.scala +++ b/core/src/main/scala/com/evolutiongaming/kafka/journal/util/ActorSystemOf.scala @@ -10,8 +10,8 @@ import com.typesafe.config.Config private[journal] object ActorSystemOf { def apply[F[_]: Sync: FromFuture]( - name: String, - config: Option[Config] = None, + name: String, + config: Option[Config] = None, ): Resource[F, ActorSystem] = { val system = Sync[F].delay { config.fold(ActorSystem(name)) { config => ActorSystem(name, config) } } diff --git a/core/src/main/scala/com/evolutiongaming/kafka/journal/util/Executors.scala b/core/src/main/scala/com/evolutiongaming/kafka/journal/util/Executors.scala index 9620e474e..feb3144f6 100644 --- a/core/src/main/scala/com/evolutiongaming/kafka/journal/util/Executors.scala +++ b/core/src/main/scala/com/evolutiongaming/kafka/journal/util/Executors.scala @@ -11,7 +11,7 @@ import scala.concurrent.{ExecutionContext, ExecutionContextExecutorService} private[journal] object Executors { def blocking[F[_]: Sync]( - name: String, + name: String, ): Resource[F, ExecutionContextExecutorService] = { for { threadFactory <- ThreadFactoryOf[F](name).toResource @@ -22,7 +22,7 @@ private[journal] object Executors { } def nonBlocking[F[_]: Sync]( - name: String, + name: String, ): Resource[F, ExecutionContextExecutorService] = { for { cores <- Runtime[F].availableCores.toResource @@ -34,8 +34,8 @@ private[journal] object Executors { } def scheduled[F[_]: Sync]( - name: String, - parallelism: Int, + name: String, + parallelism: Int, ): Resource[F, ScheduledExecutorService] = { for { threadFactory <- ThreadFactoryOf[F](name).toResource diff --git a/core/src/main/scala/com/evolutiongaming/kafka/journal/util/StartResource.scala b/core/src/main/scala/com/evolutiongaming/kafka/journal/util/StartResource.scala index 0c2ddf7ab..2febb2db2 100644 --- a/core/src/main/scala/com/evolutiongaming/kafka/journal/util/StartResource.scala +++ b/core/src/main/scala/com/evolutiongaming/kafka/journal/util/StartResource.scala @@ -7,7 +7,7 @@ import cats.syntax.all.* private[journal] object StartResource { def apply[F[_]: Concurrent, A, B](res: Resource[F, A])( - use: A => F[B], + use: A => F[B], ): F[Fiber[F, Throwable, B]] = { res diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/CreateSchema.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/CreateSchema.scala index 28bb7d967..ed24079b6 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/CreateSchema.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/CreateSchema.scala @@ -22,7 +22,7 @@ private[journal] object CreateSchema { * present in a keyspace. */ def apply[F[_]: Concurrent: CassandraCluster: CassandraSession: CassandraSync: LogOf]( - config: SchemaConfig, + config: SchemaConfig, ): F[(Schema, MigrateSchema.Fresh)] = { for { createTables <- CreateTables.of[F] @@ -32,9 +32,9 @@ private[journal] object CreateSchema { } private[cassandra] def create[F[_]: Monad]( - config: SchemaConfig, - createKeyspace: CreateKeyspace[F], - createTables: CreateTables[F], + config: SchemaConfig, + createKeyspace: CreateKeyspace[F], + createTables: CreateTables[F], ): F[(Schema, MigrateSchema.Fresh)] = { def createTables1: F[(Schema, Boolean)] = { diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandra.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandra.scala index 9dc363d9f..d0140cccc 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandra.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandra.scala @@ -37,13 +37,13 @@ object EventualCassandra { * Underlying schema is automatically created or migrated if required. */ def make[ - F[_]: Async: Parallel: ToTry: LogOf: FromGFuture: MeasureDuration: JsonCodec.Decode, + F[_]: Async: Parallel: ToTry: LogOf: FromGFuture: MeasureDuration: JsonCodec.Decode, ]( - config: EventualCassandraConfig, - origin: Option[Origin], - metrics: Option[EventualJournal.Metrics[F]], - cassandraClusterOf: CassandraClusterOf[F], - dataIntegrity: DataIntegrityConfig, + config: EventualCassandraConfig, + origin: Option[Origin], + metrics: Option[EventualJournal.Metrics[F]], + cassandraClusterOf: CassandraClusterOf[F], + dataIntegrity: DataIntegrityConfig, ): Resource[F, EventualJournal[F]] = { def journal(implicit cassandraCluster: CassandraCluster[F], cassandraSession: CassandraSession[F]) = { @@ -62,13 +62,13 @@ object EventualCassandra { * Underlying schema is automatically created or migrated if required. */ def of[ - F[_]: Temporal: Parallel: ToTry: LogOf: CassandraCluster: CassandraSession: MeasureDuration: JsonCodec.Decode, + F[_]: Temporal: Parallel: ToTry: LogOf: CassandraCluster: CassandraSession: MeasureDuration: JsonCodec.Decode, ]( - schemaConfig: SchemaConfig, - origin: Option[Origin], - metrics: Option[EventualJournal.Metrics[F]], - consistencyConfig: CassandraConsistencyConfig, - dataIntegrity: DataIntegrityConfig, + schemaConfig: SchemaConfig, + origin: Option[Origin], + metrics: Option[EventualJournal.Metrics[F]], + consistencyConfig: CassandraConsistencyConfig, + dataIntegrity: DataIntegrityConfig, ): F[EventualJournal[F]] = { for { @@ -221,9 +221,9 @@ object EventualCassandra { } private[journal] final case class Statements[F[_]]( - records: JournalStatements.SelectRecords[F], - metaJournal: MetaJournalStatements[F], - selectOffset2: Pointer2Statements.SelectOffset[F], + records: JournalStatements.SelectRecords[F], + metaJournal: MetaJournalStatements[F], + selectOffset2: Pointer2Statements.SelectOffset[F], ) private[journal] object Statements { @@ -231,10 +231,10 @@ object EventualCassandra { def apply[F[_]](implicit F: Statements[F]): Statements[F] = F def of[F[_]: Concurrent: CassandraSession: ToTry: JsonCodec.Decode]( - schema: Schema, - segmentNrsOf: SegmentNrs.Of[F], - segments: Segments, - consistencyConfig: CassandraConsistencyConfig.Read, + schema: Schema, + segmentNrsOf: SegmentNrs.Of[F], + segments: Segments, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[Statements[F]] = { for { selectRecords <- JournalStatements.SelectRecords.of[F](schema.journal, consistencyConfig) @@ -258,19 +258,19 @@ object EventualCassandra { private[journal] object MetaJournalStatements { def of[F[_]: Concurrent: CassandraSession]( - schema: Schema, - segmentNrsOf: SegmentNrs.Of[F], - segments: Segments, - consistencyConfig: CassandraConsistencyConfig.Read, + schema: Schema, + segmentNrsOf: SegmentNrs.Of[F], + segments: Segments, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[MetaJournalStatements[F]] = { of(schema.metaJournal, segmentNrsOf, segments, consistencyConfig) } def of[F[_]: Concurrent: CassandraSession]( - metaJournal: TableName, - segmentNrsOf: SegmentNrs.Of[F], - segments: Segments, - consistencyConfig: CassandraConsistencyConfig.Read, + metaJournal: TableName, + segmentNrsOf: SegmentNrs.Of[F], + segments: Segments, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[MetaJournalStatements[F]] = { for { selectJournalHead <- cassandra.MetaJournalStatements.SelectJournalHead.of[F](metaJournal, consistencyConfig) @@ -282,11 +282,11 @@ object EventualCassandra { } def fromMetaJournal[F[_]: Concurrent]( - segmentNrsOf: SegmentNrs.Of[F], - journalHead: cassandra.MetaJournalStatements.SelectJournalHead[F], - journalPointer: cassandra.MetaJournalStatements.SelectJournalPointer[F], - ids: cassandra.MetaJournalStatements.SelectIds[F], - segments: Segments, + segmentNrsOf: SegmentNrs.Of[F], + journalHead: cassandra.MetaJournalStatements.SelectJournalHead[F], + journalPointer: cassandra.MetaJournalStatements.SelectJournalPointer[F], + ids: cassandra.MetaJournalStatements.SelectIds[F], + segments: Segments, ): MetaJournalStatements[F] = { val journalHead1 = journalHead diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandraConfig.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandraConfig.scala index f6fad5b3c..0f08129e6 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandraConfig.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandraConfig.scala @@ -32,14 +32,14 @@ import pureconfig.generic.semiauto.deriveReader * tests with small number of Cassandra nodes. */ final case class EventualCassandraConfig( - retries: Int = 100, - segmentSize: SegmentSize = SegmentSize.default, - client: CassandraConfig = CassandraConfig( - name = "journal", - query = QueryConfig(consistency = ConsistencyLevel.LOCAL_QUORUM, fetchSize = 1000, defaultIdempotence = true), - ), - schema: SchemaConfig = SchemaConfig.default, - consistencyConfig: CassandraConsistencyConfig = CassandraConsistencyConfig.default, + retries: Int = 100, + segmentSize: SegmentSize = SegmentSize.default, + client: CassandraConfig = CassandraConfig( + name = "journal", + query = QueryConfig(consistency = ConsistencyLevel.LOCAL_QUORUM, fetchSize = 1000, defaultIdempotence = true), + ), + schema: SchemaConfig = SchemaConfig.default, + consistencyConfig: CassandraConsistencyConfig = CassandraConsistencyConfig.default, ) object EventualCassandraConfig { diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ExpiryService.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ExpiryService.scala index 144d310f9..5b445255e 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ExpiryService.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ExpiryService.scala @@ -45,7 +45,7 @@ private[journal] object ExpiryService { } def apply[F[_]: BracketThrowable]( - zoneId: ZoneId, + zoneId: ZoneId, ): ExpiryService[F] = { new ExpiryService[F] { diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/JournalHead.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/JournalHead.scala index 6f43d9ec6..4728d7116 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/JournalHead.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/JournalHead.scala @@ -25,12 +25,12 @@ import com.evolutiongaming.scassandra.{DecodeRow, EncodeRow} * `metajournal` record can be _created_ with `record_id` field, but never updated! */ private[journal] final case class JournalHead( - partitionOffset: PartitionOffset, - segmentSize: SegmentSize, - seqNr: SeqNr, - deleteTo: Option[DeleteTo] = none, - expiry: Option[Expiry] = none, - recordId: Option[RecordId] = none, + partitionOffset: PartitionOffset, + segmentSize: SegmentSize, + seqNr: SeqNr, + deleteTo: Option[DeleteTo] = none, + expiry: Option[Expiry] = none, + recordId: Option[RecordId] = none, ) private[journal] object JournalHead { diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/JournalStatements.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/JournalStatements.scala index a92f98a90..4a5f516ea 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/JournalStatements.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/JournalStatements.scala @@ -64,8 +64,8 @@ private[journal] object JournalStatements { object InsertRecords { def of[F[_]: Monad: CassandraSession: ToTry: JsonCodec.Encode]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[InsertRecords[F]] = { implicit val encodeTry: JsonCodec.Encode[Try] = JsonCodec.Encode.summon[F].mapK(ToTry.functionK) @@ -153,8 +153,8 @@ private[journal] object JournalStatements { object SelectRecords { def of[F[_]: Monad: CassandraSession: ToTry: JsonCodec.Decode]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Read, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[SelectRecords[F]] = { implicit val encodeTry: JsonCodec.Decode[Try] = JsonCodec.Decode.summon[F].mapK(ToTry.functionK) @@ -251,8 +251,8 @@ private[journal] object JournalStatements { object DeleteTo { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[DeleteTo[F]] = { val query = @@ -287,8 +287,8 @@ private[journal] object JournalStatements { object Delete { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[Delete[F]] = { val query = diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/MetaJournalEntry.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/MetaJournalEntry.scala index 5439381b0..359e0631c 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/MetaJournalEntry.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/MetaJournalEntry.scala @@ -8,10 +8,10 @@ import com.evolutiongaming.scassandra.{DecodeRow, EncodeRow} import java.time.Instant private[journal] final case class MetaJournalEntry( - journalHead: JournalHead, - created: Instant, - updated: Instant, - origin: Option[Origin], + journalHead: JournalHead, + created: Instant, + updated: Instant, + origin: Option[Origin], ) private[journal] object MetaJournalEntry { diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/MetaJournalStatements.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/MetaJournalStatements.scala index 315bfb2f3..2a8229a22 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/MetaJournalStatements.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/MetaJournalStatements.scala @@ -58,20 +58,20 @@ private[journal] object MetaJournalStatements { trait Insert[F[_]] { def apply( - key: Key, - segment: SegmentNr, - created: Instant, - updated: Instant, - journalHead: JournalHead, - origin: Option[Origin], + key: Key, + segment: SegmentNr, + created: Instant, + updated: Instant, + journalHead: JournalHead, + origin: Option[Origin], ): F[Unit] } object Insert { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[Insert[F]] = { val query = @@ -100,12 +100,12 @@ private[journal] object MetaJournalStatements { .prepare .map { prepared => ( - key: Key, - segment: SegmentNr, - created: Instant, - updated: Instant, - journalHead: JournalHead, - origin: Option[Origin], + key: Key, + segment: SegmentNr, + created: Instant, + updated: Instant, + journalHead: JournalHead, + origin: Option[Origin], ) => prepared .bind() @@ -131,8 +131,8 @@ private[journal] object MetaJournalStatements { object SelectJournalHead { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Read, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[SelectJournalHead[F]] = { val query = @@ -171,8 +171,8 @@ private[journal] object MetaJournalStatements { object SelectJournalPointer { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Read, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[SelectJournalPointer[F]] = { val query = @@ -210,8 +210,8 @@ private[journal] object MetaJournalStatements { object IdByTopicAndExpireOn { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Read, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[IdByTopicAndExpireOn[F]] = { val query = @@ -245,8 +245,8 @@ private[journal] object MetaJournalStatements { object IdByTopicAndCreated { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Read, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[IdByTopicAndCreated[F]] = { val query = @@ -280,8 +280,8 @@ private[journal] object MetaJournalStatements { object IdByTopicAndSegment { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Read, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[IdByTopicAndSegment[F]] = { val query = @@ -308,20 +308,20 @@ private[journal] object MetaJournalStatements { trait Update[F[_]] { def apply( - key: Key, - segment: SegmentNr, - partitionOffset: PartitionOffset, - timestamp: Instant, - seqNr: SeqNr, - deleteTo: DeleteTo, + key: Key, + segment: SegmentNr, + partitionOffset: PartitionOffset, + timestamp: Instant, + seqNr: SeqNr, + deleteTo: DeleteTo, ): F[Unit] } object Update { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[Update[F]] = { val query = @@ -337,12 +337,12 @@ private[journal] object MetaJournalStatements { .prepare .map { prepared => ( - key: Key, - segment: SegmentNr, - partitionOffset: PartitionOffset, - timestamp: Instant, - seqNr: SeqNr, - deleteTo: DeleteTo, + key: Key, + segment: SegmentNr, + partitionOffset: PartitionOffset, + timestamp: Instant, + seqNr: SeqNr, + deleteTo: DeleteTo, ) => prepared .bind() @@ -361,19 +361,19 @@ private[journal] object MetaJournalStatements { trait UpdateSeqNr[F[_]] { def apply( - key: Key, - segment: SegmentNr, - partitionOffset: PartitionOffset, - timestamp: Instant, - seqNr: SeqNr, + key: Key, + segment: SegmentNr, + partitionOffset: PartitionOffset, + timestamp: Instant, + seqNr: SeqNr, ): F[Unit] } object UpdateSeqNr { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[UpdateSeqNr[F]] = { val query = @@ -405,20 +405,20 @@ private[journal] object MetaJournalStatements { trait UpdateExpiry[F[_]] { def apply( - key: Key, - segment: SegmentNr, - partitionOffset: PartitionOffset, - timestamp: Instant, - seqNr: SeqNr, - expiry: Expiry, + key: Key, + segment: SegmentNr, + partitionOffset: PartitionOffset, + timestamp: Instant, + seqNr: SeqNr, + expiry: Expiry, ): F[Unit] } object UpdateExpiry { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[UpdateExpiry[F]] = { val query = @@ -458,8 +458,8 @@ private[journal] object MetaJournalStatements { object UpdateDeleteTo { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[UpdateDeleteTo[F]] = { val query = @@ -497,8 +497,8 @@ private[journal] object MetaJournalStatements { object UpdatePartitionOffset { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[UpdatePartitionOffset[F]] = { s""" |UPDATE ${name.toCql} @@ -531,8 +531,8 @@ private[journal] object MetaJournalStatements { object Delete { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[Delete[F]] = { val query = @@ -564,8 +564,8 @@ private[journal] object MetaJournalStatements { object DeleteExpiry { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[DeleteExpiry[F]] = { val query = @@ -598,8 +598,8 @@ private[journal] object MetaJournalStatements { object SelectIds { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Read, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[SelectIds[F]] = { for { prepared <- s"SELECT id FROM ${name.toCql} WHERE topic = ? AND segment = ?".prepare diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/Pointer2Statements.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/Pointer2Statements.scala index 56bff30af..81e74c7b6 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/Pointer2Statements.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/Pointer2Statements.scala @@ -35,8 +35,8 @@ private[journal] object Pointer2Statements { object SelectTopics { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Read, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[SelectTopics[F]] = { val query = s"""SELECT DISTINCT topic, partition FROM ${name.toCql}""".stripMargin @@ -76,8 +76,8 @@ private[journal] object Pointer2Statements { } def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Read, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[Select[F]] = { s""" |SELECT created FROM ${name.toCql} @@ -106,8 +106,8 @@ private[journal] object Pointer2Statements { object SelectOffset { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Read, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Read, ): F[SelectOffset[F]] = { val query = @@ -139,8 +139,8 @@ private[journal] object Pointer2Statements { object Insert { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[Insert[F]] = { val query = @@ -174,8 +174,8 @@ private[journal] object Pointer2Statements { object Update { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[Update[F]] = { val query = diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/PointerStatements.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/PointerStatements.scala index 390daa398..c1176857f 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/PointerStatements.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/PointerStatements.scala @@ -33,8 +33,8 @@ private[journal] object PointerStatements { object Insert { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[Insert[F]] = { val query = @@ -68,8 +68,8 @@ private[journal] object PointerStatements { object Update { def of[F[_]: Monad: CassandraSession]( - name: TableName, - consistencyConfig: CassandraConsistencyConfig.Write, + name: TableName, + consistencyConfig: CassandraConsistencyConfig.Write, ): F[Update[F]] = { val query = diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ReplicatedCassandra.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ReplicatedCassandra.scala index fd3e3d365..527ff05c6 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ReplicatedCassandra.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ReplicatedCassandra.scala @@ -29,11 +29,11 @@ private[journal] object ReplicatedCassandra { private sealed abstract class Main def of[ - F[_]: Async: Parallel: ToTry: LogOf: Fail: CassandraCluster: CassandraSession: MeasureDuration: JsonCodec.Encode, + F[_]: Async: Parallel: ToTry: LogOf: Fail: CassandraCluster: CassandraSession: MeasureDuration: JsonCodec.Encode, ]( - config: EventualCassandraConfig, - origin: Option[Origin], - metrics: Option[ReplicatedJournal.Metrics[F]], + config: EventualCassandraConfig, + origin: Option[Origin], + metrics: Option[ReplicatedJournal.Metrics[F]], ): F[ReplicatedJournal[F]] = { for { @@ -52,10 +52,10 @@ private[journal] object ReplicatedCassandra { } def apply[F[_]: Sync: Parallel: Fail: UUIDGen]( - segmentSizeDefault: SegmentSize, - segmentNrsOf: SegmentNrs.Of[F], - statements: Statements[F], - expiryService: ExpiryService[F], + segmentSizeDefault: SegmentSize, + segmentNrsOf: SegmentNrs.Of[F], + statements: Statements[F], + expiryService: ExpiryService[F], ): ReplicatedJournal[F] = { new Main with ReplicatedJournal[F] { @@ -134,10 +134,10 @@ private[journal] object ReplicatedCassandra { new Main with ReplicatedKeyJournal[F] { def append( - offset: Offset, - timestamp: Instant, - expireAfter: Option[ExpireAfter], - events: Nel[EventRecord[EventualPayloadAndType]], + offset: Offset, + timestamp: Instant, + expireAfter: Option[ExpireAfter], + events: Nel[EventRecord[EventualPayloadAndType]], ): F[Changed] = { def partitionOffset = PartitionOffset(partition, offset) @@ -146,9 +146,9 @@ private[journal] object ReplicatedCassandra { @tailrec def loop( - events: List[EventRecord[EventualPayloadAndType]], - s: Option[(Segment, Nel[EventRecord[EventualPayloadAndType]])], - result: F[Unit], + events: List[EventRecord[EventualPayloadAndType]], + s: Option[(Segment, Nel[EventRecord[EventualPayloadAndType]])], + result: F[Unit], ): F[Unit] = { def insert(segment: Segment, events: Nel[EventRecord[EventualPayloadAndType]]) = { @@ -280,10 +280,10 @@ private[journal] object ReplicatedCassandra { } def delete( - offset: Offset, - timestamp: Instant, - deleteTo: DeleteTo, - origin: Option[Origin], + offset: Offset, + timestamp: Instant, + deleteTo: DeleteTo, + origin: Option[Origin], ): F[Changed] = { def partitionOffset = PartitionOffset(partition, offset) @@ -381,8 +381,8 @@ private[journal] object ReplicatedCassandra { } def purge( - offset: Offset, - timestamp: Instant, + offset: Offset, + timestamp: Instant, ): F[Changed] = { for { journalHead <- journalHeadRef.get @@ -478,15 +478,15 @@ private[journal] object ReplicatedCassandra { private[journal] object MetaJournalStatements { def of[F[_]: Monad: CassandraSession]( - schema: Schema, - consistencyConfig: CassandraConsistencyConfig, + schema: Schema, + consistencyConfig: CassandraConsistencyConfig, ): F[MetaJournalStatements[F]] = { of[F](schema.metaJournal, consistencyConfig) } def of[F[_]: Monad: CassandraSession]( - metaJournal: TableName, - consistencyConfig: CassandraConsistencyConfig, + metaJournal: TableName, + consistencyConfig: CassandraConsistencyConfig, ): F[MetaJournalStatements[F]] = { for { @@ -517,15 +517,15 @@ private[journal] object ReplicatedCassandra { private sealed abstract class MetaJournal def apply[F[_]]( - selectJournalHead: cassandra.MetaJournalStatements.SelectJournalHead[F], - insert: cassandra.MetaJournalStatements.Insert[F], - update: cassandra.MetaJournalStatements.Update[F], - updateSeqNr: cassandra.MetaJournalStatements.UpdateSeqNr[F], - updateExpiry: cassandra.MetaJournalStatements.UpdateExpiry[F], - updateDeleteTo: cassandra.MetaJournalStatements.UpdateDeleteTo[F], - updatePartitionOffset: cassandra.MetaJournalStatements.UpdatePartitionOffset[F], - delete: cassandra.MetaJournalStatements.Delete[F], - deleteExpiry: cassandra.MetaJournalStatements.DeleteExpiry[F], + selectJournalHead: cassandra.MetaJournalStatements.SelectJournalHead[F], + insert: cassandra.MetaJournalStatements.Insert[F], + update: cassandra.MetaJournalStatements.Update[F], + updateSeqNr: cassandra.MetaJournalStatements.UpdateSeqNr[F], + updateExpiry: cassandra.MetaJournalStatements.UpdateExpiry[F], + updateDeleteTo: cassandra.MetaJournalStatements.UpdateDeleteTo[F], + updatePartitionOffset: cassandra.MetaJournalStatements.UpdatePartitionOffset[F], + delete: cassandra.MetaJournalStatements.Delete[F], + deleteExpiry: cassandra.MetaJournalStatements.DeleteExpiry[F], ): MetaJournalStatements[F] = { val inset1 = insert @@ -608,17 +608,17 @@ private[journal] object ReplicatedCassandra { } private[journal] final case class Statements[F[_]]( - insertRecords: JournalStatements.InsertRecords[F], - deleteRecordsTo: JournalStatements.DeleteTo[F], - deleteRecords: JournalStatements.Delete[F], - metaJournal: MetaJournalStatements[F], - selectOffset2: Pointer2Statements.SelectOffset[F], - selectPointer2: Pointer2Statements.Select[F], - insertPointer: PointerStatements.Insert[F], - insertPointer2: Pointer2Statements.Insert[F], - updatePointer: PointerStatements.Update[F], - updatePointer2: Pointer2Statements.Update[F], - selectTopics2: Pointer2Statements.SelectTopics[F], + insertRecords: JournalStatements.InsertRecords[F], + deleteRecordsTo: JournalStatements.DeleteTo[F], + deleteRecords: JournalStatements.Delete[F], + metaJournal: MetaJournalStatements[F], + selectOffset2: Pointer2Statements.SelectOffset[F], + selectPointer2: Pointer2Statements.Select[F], + insertPointer: PointerStatements.Insert[F], + insertPointer2: Pointer2Statements.Insert[F], + updatePointer: PointerStatements.Update[F], + updatePointer2: Pointer2Statements.Update[F], + selectTopics2: Pointer2Statements.SelectTopics[F], ) private[journal] object Statements { @@ -626,8 +626,8 @@ private[journal] object ReplicatedCassandra { def apply[F[_]](implicit F: Statements[F]): Statements[F] = F def of[F[_]: Monad: CassandraSession: ToTry: JsonCodec.Encode]( - schema: Schema, - consistencyConfig: CassandraConsistencyConfig, + schema: Schema, + consistencyConfig: CassandraConsistencyConfig, ): F[Statements[F]] = { for { insertRecords <- JournalStatements.InsertRecords.of[F](schema.journal, consistencyConfig.write) diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/Schema.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/Schema.scala index 466f671da..642345356 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/Schema.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/Schema.scala @@ -18,9 +18,9 @@ import com.evolutiongaming.scassandra.TableName * @param setting used to track schema migrations, see [[SetupSchema]] and [[MigrateSchema]] for details */ private[journal] final case class Schema( - journal: TableName, - metaJournal: TableName, - pointer: TableName, // TODO MR remove with next major release - pointer2: TableName, - setting: TableName, + journal: TableName, + metaJournal: TableName, + pointer: TableName, // TODO MR remove with next major release + pointer2: TableName, + setting: TableName, ) diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/SchemaConfig.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/SchemaConfig.scala index b4e3eebfc..15e077a04 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/SchemaConfig.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/SchemaConfig.scala @@ -5,14 +5,14 @@ import pureconfig.ConfigReader import pureconfig.generic.semiauto.deriveReader private[journal] final case class SchemaConfig( - keyspace: KeyspaceConfig = KeyspaceConfig.default, - journalTable: String = "journal", - metaJournalTable: String = "metajournal", - pointerTable: String = "pointer", // should not be used any more - pointer2Table: String = "pointer2", - settingTable: String = "setting", - locksTable: String = "locks", - autoCreate: Boolean = true, + keyspace: KeyspaceConfig = KeyspaceConfig.default, + journalTable: String = "journal", + metaJournalTable: String = "metajournal", + pointerTable: String = "pointer", // should not be used any more + pointer2Table: String = "pointer2", + settingTable: String = "setting", + locksTable: String = "locks", + autoCreate: Boolean = true, ) private[journal] object SchemaConfig { diff --git a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/SetupSchema.scala b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/SetupSchema.scala index 547a5de32..64c7f0e93 100644 --- a/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/SetupSchema.scala +++ b/eventual-cassandra/src/main/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/SetupSchema.scala @@ -49,10 +49,10 @@ private[journal] object SetupSchema { } private[cassandra] def migrate[F[_]: MonadThrow: CassandraSession]( - schema: Schema, - fresh: MigrateSchema.Fresh, - settings: Settings[F], - cassandraSync: CassandraSync[F], + schema: Schema, + fresh: MigrateSchema.Fresh, + settings: Settings[F], + cassandraSync: CassandraSync[F], ): F[Unit] = { val migrateSchema = MigrateSchema.forSettingKey( cassandraSync = cassandraSync, @@ -64,9 +64,9 @@ private[journal] object SetupSchema { } def apply[F[_]: Temporal: Parallel: CassandraCluster: CassandraSession: LogOf]( - config: SchemaConfig, - origin: Option[Origin], - consistencyConfig: CassandraConsistencyConfig, + config: SchemaConfig, + origin: Option[Origin], + consistencyConfig: CassandraConsistencyConfig, ): F[Schema] = { def createSchema(implicit cassandraSync: CassandraSync[F]): F[(Schema, Fresh)] = CreateSchema(config) diff --git a/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandraSpec.scala b/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandraSpec.scala index d0c8a1ac2..1fa26c7c3 100644 --- a/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandraSpec.scala +++ b/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandraSpec.scala @@ -144,9 +144,9 @@ object EventualCassandraSpec { } def replicatedJournalOf( - segmentSize: SegmentSize, - delete: Boolean, - segmentNrsOf: SegmentNrs.Of[StateT], + segmentSize: SegmentSize, + delete: Boolean, + segmentNrsOf: SegmentNrs.Of[StateT], ): ReplicatedJournal[StateT] = { val insertRecords: JournalStatements.InsertRecords[StateT] = { (key, segment, insert) => @@ -381,10 +381,10 @@ object EventualCassandraSpec { } def journalsOf( - segmentSize: SegmentSize, - delete: Boolean, - segmentsFirst: Segments, - segmentsSecond: Segments, + segmentSize: SegmentSize, + delete: Boolean, + segmentsFirst: Segments, + segmentsSecond: Segments, ): EventualAndReplicated[StateT] = { val segmentNrsOf = SegmentNrs.Of[StateT](first = segmentsFirst, second = segmentsSecond) val replicatedJournal = replicatedJournalOf(segmentSize, delete, segmentNrsOf) @@ -393,9 +393,9 @@ object EventualCassandraSpec { } final case class State( - journal: Map[(Key, SegmentNr), List[JournalRecord]], - metaJournal: Map[Key, JournalHead], - pointers: Map[Topic, TopicPointers], + journal: Map[(Key, SegmentNr), List[JournalRecord]], + metaJournal: Map[Key, JournalHead], + pointers: Map[Topic, TopicPointers], ) object State { diff --git a/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandraTest.scala b/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandraTest.scala index d7ea344bf..92d10ce3b 100644 --- a/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandraTest.scala +++ b/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/EventualCassandraTest.scala @@ -545,10 +545,10 @@ object EventualCassandraTest { final case class PointerEntry(offset: Offset, created: Instant, updated: Instant) final case class State( - pointers: Map[Topic, Map[Partition, PointerEntry]] = Map.empty, - metadata: Map[Topic, Map[String, MetaJournalEntry]] = Map.empty, - metaJournal: Map[(Topic, SegmentNr), Map[String, MetaJournalEntry]] = Map.empty, - journal: Map[(Key, SegmentNr), Map[(SeqNr, Instant), JournalRecord]] = Map.empty, + pointers: Map[Topic, Map[Partition, PointerEntry]] = Map.empty, + metadata: Map[Topic, Map[String, MetaJournalEntry]] = Map.empty, + metaJournal: Map[(Topic, SegmentNr), Map[String, MetaJournalEntry]] = Map.empty, + journal: Map[(Key, SegmentNr), Map[(SeqNr, Instant), JournalRecord]] = Map.empty, ) object State { diff --git a/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ReplicatedCassandraMetaJournalStatementsTest.scala b/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ReplicatedCassandraMetaJournalStatementsTest.scala index ec40698f3..3fe0eb32e 100644 --- a/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ReplicatedCassandraMetaJournalStatementsTest.scala +++ b/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ReplicatedCassandraMetaJournalStatementsTest.scala @@ -296,8 +296,8 @@ object ReplicatedCassandraMetaJournalStatementsTest { val expiry: Expiry = Expiry(1.day.toExpireAfter, LocalDate.of(2019, 12, 12).toExpireOn) final case class State( - metadata: Map[Topic, Map[String, MetaJournalEntry]] = Map.empty, - metaJournal: Map[(Topic, SegmentNr), Map[String, MetaJournalEntry]] = Map.empty, + metadata: Map[Topic, Map[String, MetaJournalEntry]] = Map.empty, + metaJournal: Map[(Topic, SegmentNr), Map[String, MetaJournalEntry]] = Map.empty, ) object State { diff --git a/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ReplicatedCassandraTest.scala b/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ReplicatedCassandraTest.scala index 13bbb5d08..70d6052b4 100644 --- a/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ReplicatedCassandraTest.scala +++ b/eventual-cassandra/src/test/scala/com/evolutiongaming/kafka/journal/eventual/cassandra/ReplicatedCassandraTest.scala @@ -2132,53 +2132,53 @@ object ReplicatedCassandraTest { final case class DeleteRecordsTo(key: Key, segment: SegmentNr, seqNr: SeqNr) extends Action final case class InsertMetaJournal( - key: Key, - segment: SegmentNr, - created: Instant, - updated: Instant, - journalHead: JournalHead, - origin: Option[Origin], + key: Key, + segment: SegmentNr, + created: Instant, + updated: Instant, + journalHead: JournalHead, + origin: Option[Origin], ) extends Action final case class Update( - key: Key, - segment: SegmentNr, - partitionOffset: PartitionOffset, - timestamp: Instant, - seqNr: SeqNr, - deleteTo: DeleteTo, + key: Key, + segment: SegmentNr, + partitionOffset: PartitionOffset, + timestamp: Instant, + seqNr: SeqNr, + deleteTo: DeleteTo, ) extends Action final case class UpdateSeqNr( - key: Key, - segment: SegmentNr, - partitionOffset: PartitionOffset, - timestamp: Instant, - seqNr: SeqNr, + key: Key, + segment: SegmentNr, + partitionOffset: PartitionOffset, + timestamp: Instant, + seqNr: SeqNr, ) extends Action final case class UpdateExpiry( - key: Key, - segment: SegmentNr, - partitionOffset: PartitionOffset, - timestamp: Instant, - seqNr: SeqNr, - expiry: Expiry, + key: Key, + segment: SegmentNr, + partitionOffset: PartitionOffset, + timestamp: Instant, + seqNr: SeqNr, + expiry: Expiry, ) extends Action final case class UpdateDeleteTo( - key: Key, - segment: SegmentNr, - partitionOffset: PartitionOffset, - timestamp: Instant, - deleteTo: DeleteTo, + key: Key, + segment: SegmentNr, + partitionOffset: PartitionOffset, + timestamp: Instant, + deleteTo: DeleteTo, ) extends Action final case class UpdatePartitionOffset( - key: Key, - segment: SegmentNr, - partitionOffset: PartitionOffset, - timestamp: Instant, + key: Key, + segment: SegmentNr, + partitionOffset: PartitionOffset, + timestamp: Instant, ) extends Action final case class DeleteMetaJournal(key: Key, segment: SegmentNr) extends Action @@ -2187,10 +2187,10 @@ object ReplicatedCassandraTest { } final case class State( - actions: List[Action] = List.empty, - pointers: Map[Topic, Map[Partition, PointerEntry]] = Map.empty, - metaJournal: Map[(Topic, SegmentNr), Map[String, MetaJournalEntry]] = Map.empty, - journal: Map[(Key, SegmentNr), Map[(SeqNr, Instant), JournalRecord]] = Map.empty, + actions: List[Action] = List.empty, + pointers: Map[Topic, Map[Partition, PointerEntry]] = Map.empty, + metaJournal: Map[(Topic, SegmentNr), Map[String, MetaJournalEntry]] = Map.empty, + journal: Map[(Key, SegmentNr), Map[(SeqNr, Instant), JournalRecord]] = Map.empty, ) object State { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/Action.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/Action.scala index 6234acb5d..451413207 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/Action.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/Action.scala @@ -23,35 +23,35 @@ sealed abstract class Action extends Product { object Action { def append( - key: Key, - timestamp: Instant, - header: ActionHeader.Append, - payload: ByteVector, - headers: Headers, + key: Key, + timestamp: Instant, + header: ActionHeader.Append, + payload: ByteVector, + headers: Headers, ): Action = { Append(key, timestamp, header, payload, headers) } def delete( - key: Key, - timestamp: Instant, - header: ActionHeader.Delete, + key: Key, + timestamp: Instant, + header: ActionHeader.Delete, ): Action = { Delete(key, timestamp, header) } def purge( - key: Key, - timestamp: Instant, - header: ActionHeader.Purge, + key: Key, + timestamp: Instant, + header: ActionHeader.Purge, ): Action = { Purge(key, timestamp, header) } def mark( - key: Key, - timestamp: Instant, - header: ActionHeader.Mark, + key: Key, + timestamp: Instant, + header: ActionHeader.Mark, ): Action = { Mark(key, timestamp, header) } @@ -61,11 +61,11 @@ object Action { sealed abstract class System extends Action final case class Append( - key: Key, - timestamp: Instant, - header: ActionHeader.Append, - payload: ByteVector, - headers: Headers, + key: Key, + timestamp: Instant, + header: ActionHeader.Append, + payload: ByteVector, + headers: Headers, ) extends User { def payloadType: PayloadType.BinaryOrJson = header.payloadType @@ -76,13 +76,13 @@ object Action { object Append { def of[F[_]: Functor, A]( - key: Key, - timestamp: Instant, - origin: Option[Origin], - version: Option[Version], - events: Events[A], - metadata: HeaderMetadata, - headers: Headers, + key: Key, + timestamp: Instant, + origin: Option[Origin], + version: Option[Version], + events: Events[A], + metadata: HeaderMetadata, + headers: Headers, )(implicit kafkaWrite: KafkaWrite[F, A]): F[Append] = { for { payloadAndType <- kafkaWrite(events) @@ -105,9 +105,9 @@ object Action { } final case class Delete( - key: Key, - timestamp: Instant, - header: ActionHeader.Delete, + key: Key, + timestamp: Instant, + header: ActionHeader.Delete, ) extends User { def to: DeleteTo = header.to @@ -116,11 +116,11 @@ object Action { object Delete { def apply( - key: Key, - timestamp: Instant, - to: DeleteTo, - origin: Option[Origin], - version: Option[Version], + key: Key, + timestamp: Instant, + to: DeleteTo, + origin: Option[Origin], + version: Option[Version], ): Delete = { val header = ActionHeader.Delete(to, origin, version) Delete(key, timestamp, header) @@ -128,9 +128,9 @@ object Action { } final case class Purge( - key: Key, - timestamp: Instant, - header: ActionHeader.Purge, + key: Key, + timestamp: Instant, + header: ActionHeader.Purge, ) extends User object Purge { @@ -142,9 +142,9 @@ object Action { } final case class Mark( - key: Key, - timestamp: Instant, - header: ActionHeader.Mark, + key: Key, + timestamp: Instant, + header: ActionHeader.Mark, ) extends System { def id: String = header.id @@ -153,11 +153,11 @@ object Action { object Mark { def apply( - key: Key, - timestamp: Instant, - id: String, - origin: Option[Origin], - version: Option[Version], + key: Key, + timestamp: Instant, + id: String, + origin: Option[Origin], + version: Option[Version], ): Mark = { val header = ActionHeader.Mark(id, origin, version) Mark(key, timestamp, header) diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/ActionHeader.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/ActionHeader.scala index 6da698a8c..f44adc303 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/ActionHeader.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/ActionHeader.scala @@ -94,27 +94,27 @@ object ActionHeader { sealed abstract class AppendOrDelete extends ActionHeader final case class Append( - range: SeqRange, - origin: Option[Origin], - version: Option[Version], - payloadType: PayloadType.BinaryOrJson, - metadata: HeaderMetadata, + range: SeqRange, + origin: Option[Origin], + version: Option[Version], + payloadType: PayloadType.BinaryOrJson, + metadata: HeaderMetadata, ) extends AppendOrDelete final case class Delete( - to: DeleteTo, - origin: Option[Origin], - version: Option[Version], + to: DeleteTo, + origin: Option[Origin], + version: Option[Version], ) extends AppendOrDelete final case class Purge( - origin: Option[Origin], - version: Option[Version], + origin: Option[Origin], + version: Option[Version], ) extends AppendOrDelete final case class Mark( - id: String, - origin: Option[Origin], - version: Option[Version], + id: String, + origin: Option[Origin], + version: Option[Version], ) extends ActionHeader } diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/AppendEvents.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/AppendEvents.scala index 0e211906d..6a644bc47 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/AppendEvents.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/AppendEvents.scala @@ -8,10 +8,10 @@ import com.evolutiongaming.kafka.journal.conversions.KafkaWrite trait AppendEvents[F[_]] { def apply[A]( - key: Key, - events: Nel[Event[A]], - metadata: RecordMetadata, - headers: Headers, + key: Key, + events: Nel[Event[A]], + metadata: RecordMetadata, + headers: Headers, )(implicit kafkaWrite: KafkaWrite[F, A]): F[PartitionOffset] } @@ -19,7 +19,7 @@ object AppendEvents { def apply[F[_]: Monad](produce: Produce[F]): AppendEvents[F] = new AppendEvents[F] { override def apply[A](key: Key, events0: Nel[Event[A]], metadata: RecordMetadata, headers: Headers)( - implicit kafkaWrite: KafkaWrite[F, A], + implicit kafkaWrite: KafkaWrite[F, A], ): F[PartitionOffset] = { val events = Events(events0, metadata.payload) diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/AppendMarker.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/AppendMarker.scala index cc0590784..be6a43a2c 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/AppendMarker.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/AppendMarker.scala @@ -12,7 +12,7 @@ trait AppendMarker[F[_]] { object AppendMarker { def apply[F[_]: FlatMap: RandomIdOf]( - produce: Produce[F], + produce: Produce[F], ): AppendMarker[F] = { (key: Key) => { for { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsRecord.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsRecord.scala index 70002be83..032785170 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsRecord.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsRecord.scala @@ -7,12 +7,12 @@ import scodec.bits.ByteVector object ConsRecord { def apply( - topicPartition: TopicPartition, - offset: Offset, - timestampAndType: Option[TimestampAndType], - key: Option[WithSize[String]] = None, - value: Option[WithSize[ByteVector]] = None, - headers: List[Header] = Nil, + topicPartition: TopicPartition, + offset: Offset, + timestampAndType: Option[TimestampAndType], + key: Option[WithSize[String]] = None, + value: Option[WithSize[ByteVector]] = None, + headers: List[Header] = Nil, ): ConsRecord = { ConsumerRecord( diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsumeActionRecords.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsumeActionRecords.scala index 88aef07c8..0be8fbf68 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsumeActionRecords.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsumeActionRecords.scala @@ -18,7 +18,7 @@ trait ConsumeActionRecords[F[_]] { object ConsumeActionRecords { def apply[F[_]: BracketThrowable]( - consumer: Resource[F, Journals.Consumer[F]], + consumer: Resource[F, Journals.Consumer[F]], )(implicit consRecordToActionRecord: ConsRecordToActionRecord[F]): ConsumeActionRecords[F] = { class Main new Main with ConsumeActionRecords[F] { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsumerPool.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsumerPool.scala index 711a7fa26..3b97b83e9 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsumerPool.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsumerPool.scala @@ -18,10 +18,10 @@ private[journal] object ConsumerPool { * @return The outer Resource is for the pool, the inner is for consumers */ def make[F[_]: Async: Runtime: MeasureDuration]( - poolConfig: ConsumerPoolConfig, - metrics: Option[ConsumerPoolMetrics[F]], - consumer: Resource[F, Consumer[F]], - timeout: FiniteDuration = 1.minute, + poolConfig: ConsumerPoolConfig, + metrics: Option[ConsumerPoolMetrics[F]], + consumer: Resource[F, Consumer[F]], + timeout: FiniteDuration = 1.minute, ): Resource[F, Resource[F, Consumer[F]]] = { for { cores <- Runtime[F].availableCores.toResource diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsumerPoolMetrics.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsumerPoolMetrics.scala index 8dc4ba39f..05597c821 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsumerPoolMetrics.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/ConsumerPoolMetrics.scala @@ -18,8 +18,8 @@ trait ConsumerPoolMetrics[F[_]] { object ConsumerPoolMetrics { def make[F[_]]( - registry: CollectorRegistry[F], - prefix: String = "journal", + registry: CollectorRegistry[F], + prefix: String = "journal", ): Resource[F, ConsumerPoolMetrics[F]] = { registry diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/EventRecord.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/EventRecord.scala index 67a5356a8..a2a090145 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/EventRecord.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/EventRecord.scala @@ -10,13 +10,13 @@ import java.time.Instant * @param origin identifier of event origin, for instance node IP address */ final case class EventRecord[A]( - event: Event[A], - timestamp: Instant, - partitionOffset: PartitionOffset, - origin: Option[Origin], - version: Option[Version], - metadata: RecordMetadata, - headers: Headers, + event: Event[A], + timestamp: Instant, + partitionOffset: PartitionOffset, + origin: Option[Origin], + version: Option[Version], + metadata: RecordMetadata, + headers: Headers, ) { def seqNr: SeqNr = event.seqNr @@ -31,18 +31,18 @@ final case class EventRecord[A]( object EventRecord { def apply[A]( - record: ActionRecord[Action.Append], - event: Event[A], - metadata: PayloadMetadata, + record: ActionRecord[Action.Append], + event: Event[A], + metadata: PayloadMetadata, ): EventRecord[A] = { apply(record.action, event, record.partitionOffset, metadata) } def apply[A]( - action: Action.Append, - event: Event[A], - partitionOffset: PartitionOffset, - metadata: PayloadMetadata, + action: Action.Append, + event: Event[A], + partitionOffset: PartitionOffset, + metadata: PayloadMetadata, ): EventRecord[A] = { EventRecord( event = event, diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/Events.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/Events.scala index c93f43b19..b3ed8d82f 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/Events.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/Events.scala @@ -27,12 +27,12 @@ object Events { } implicit def eventsToBytes[F[_]: FromAttempt, A]( - implicit eventCodec: Codec[Event[A]], - metadataCodec: Codec[PayloadMetadata], + implicit eventCodec: Codec[Event[A]], + metadataCodec: Codec[PayloadMetadata], ): ToBytes[F, Events[A]] = ToBytes.fromEncoder implicit def eventsFromBytes[F[_]: FromAttempt, A]( - implicit eventCodec: Codec[Event[A]], - metadataCodec: Codec[PayloadMetadata], + implicit eventCodec: Codec[Event[A]], + metadataCodec: Codec[PayloadMetadata], ): FromBytes[F, Events[A]] = FromBytes.fromDecoder } diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCache.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCache.scala index d76de4d64..567d230ef 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCache.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCache.scala @@ -99,9 +99,9 @@ object HeadCache { * did not change. */ def make[F[_]: Async: Parallel: Runtime: LogOf: KafkaConsumerOf: MeasureDuration: FromTry: FromJsResult: JsonCodec.Decode]( - consumerConfig: ConsumerConfig, - eventualJournal: EventualJournal[F], - metrics: Option[HeadCacheMetrics[F]], + consumerConfig: ConsumerConfig, + eventualJournal: EventualJournal[F], + metrics: Option[HeadCacheMetrics[F]], ): Resource[F, HeadCache[F]] = { for { log <- LogOf[F].apply(HeadCache.getClass).toResource @@ -141,11 +141,11 @@ object HeadCache { * did not change. */ def make[F[_]: Async: Parallel: Runtime: FromJsResult: MeasureDuration: JsonCodec.Decode]( - eventual: Eventual[F], - log: Log[F], - consumer: Resource[F, TopicCache.Consumer[F]], - metrics: Option[HeadCacheMetrics[F]], - config: HeadCacheConfig = HeadCacheConfig.default, + eventual: Eventual[F], + log: Log[F], + consumer: Resource[F, TopicCache.Consumer[F]], + metrics: Option[HeadCacheMetrics[F]], + config: HeadCacheConfig = HeadCacheConfig.default, ): Resource[F, HeadCache[F]] = { val consRecordToActionHeader = ConsRecordToActionHeader[F] @@ -396,8 +396,8 @@ object HeadCache { * Prefix to use for the registered metrics. */ def make[F[_]: Monad]( - registry: CollectorRegistry[F], - prefix: Prefix = Prefix.default, + registry: CollectorRegistry[F], + prefix: Prefix = Prefix.default, ): Resource[F, Metrics[F]] = { val getLatencySummary = registry.summary( diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheConfig.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheConfig.scala index 785986c90..82f18bee7 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheConfig.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheConfig.scala @@ -23,10 +23,10 @@ import scala.concurrent.duration.* * Partition cache configuration as described in [[HeadCacheConfig.Partition]]. */ final case class HeadCacheConfig( - timeout: FiniteDuration = 1.second, - expiry: FiniteDuration = 10.minutes, - removeInterval: FiniteDuration = 100.millis, - partition: HeadCacheConfig.Partition = HeadCacheConfig.Partition.default, + timeout: FiniteDuration = 1.second, + expiry: FiniteDuration = 10.minutes, + removeInterval: FiniteDuration = 100.millis, + partition: HeadCacheConfig.Partition = HeadCacheConfig.Partition.default, ) object HeadCacheConfig { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheConsumption.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheConsumption.scala index 0fbac6339..8c0a9120f 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheConsumption.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheConsumption.scala @@ -47,10 +47,10 @@ private[journal] object HeadCacheConsumption { * succeeds. */ def apply[F[_]: BracketThrowable: Sleep]( - topic: Topic, - pointers: F[Map[Partition, Offset]], - consumer: Resource[F, TopicCache.Consumer[F]], - log: Log[F], + topic: Topic, + pointers: F[Map[Partition, Offset]], + consumer: Resource[F, TopicCache.Consumer[F]], + log: Log[F], ): Stream[F, ConsumerRecords[String, Unit]] = { def partitions(consumer: Consumer[F], random: Random.State): F[Nes[Partition]] = { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheMetrics.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheMetrics.scala index 429d11544..daefa1c0c 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheMetrics.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheMetrics.scala @@ -12,8 +12,8 @@ object HeadCacheMetrics { def empty[F[_]: Applicative]: HeadCacheMetrics[F] = apply(HeadCache.Metrics.empty, CacheMetrics.empty) def make[F[_]: Monad]( - registry: CollectorRegistry[F], - prefix: HeadCache.Metrics.Prefix = HeadCache.Metrics.Prefix.default, + registry: CollectorRegistry[F], + prefix: HeadCache.Metrics.Prefix = HeadCache.Metrics.Prefix.default, ): Resource[F, HeadCacheMetrics[F]] = { for { headCache <- HeadCache.Metrics.make(registry, prefix) diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheOf.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheOf.scala index ad531ba5b..ab4803df6 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheOf.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadCacheOf.scala @@ -12,8 +12,8 @@ import com.evolutiongaming.skafka.consumer.ConsumerConfig trait HeadCacheOf[F[_]] { def apply( - consumerConfig: ConsumerConfig, - eventualJournal: EventualJournal[F], + consumerConfig: ConsumerConfig, + eventualJournal: EventualJournal[F], ): Resource[F, HeadCache[F]] } @@ -28,7 +28,7 @@ object HeadCacheOf { def apply[F[_]](implicit F: HeadCacheOf[F]): HeadCacheOf[F] = F def apply[F[_]: Async: Parallel: Runtime: LogOf: KafkaConsumerOf: MeasureDuration: FromTry: FromJsResult: JsonCodec.Decode]( - metrics: Option[HeadCacheMetrics[F]], + metrics: Option[HeadCacheMetrics[F]], ): HeadCacheOf[F] = { (consumerConfig: ConsumerConfig, eventualJournal: EventualJournal[F]) => { HeadCache.make[F](consumerConfig, eventualJournal, metrics) diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadInfo.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadInfo.scala index 422cbef59..da5d2f4a8 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadInfo.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/HeadInfo.scala @@ -159,9 +159,9 @@ object HeadInfo { * [[SeqNr]] of the _last_ deleted event, if any. */ final case class Append( - offset: Offset, - seqNr: SeqNr, - deleteTo: Option[DeleteTo], + offset: Offset, + seqNr: SeqNr, + deleteTo: Option[DeleteTo], ) extends NonEmpty /** The only non-replicated records are delete actions. @@ -180,7 +180,7 @@ object HeadInfo { * [[SeqNr]] of the _last_ deleted event. */ final case class Delete( - deleteTo: DeleteTo, + deleteTo: DeleteTo, ) extends NonEmpty /** The last non-replicated record was a journal purge action. diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/Journal.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/Journal.scala index f714d0716..53e4f2886 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/Journal.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/Journal.scala @@ -18,13 +18,13 @@ import scala.concurrent.duration.* trait Journal[F[_]] { def append[A]( - events: Nel[Event[A]], - metadata: RecordMetadata = RecordMetadata.empty, - headers: Headers = Headers.empty, + events: Nel[Event[A]], + metadata: RecordMetadata = RecordMetadata.empty, + headers: Headers = Headers.empty, )(implicit kafkaWrite: KafkaWrite[F, A]): F[PartitionOffset] def read[A]( - from: SeqNr = SeqNr.min, + from: SeqNr = SeqNr.min, )(implicit kafkaRead: KafkaRead[F, A], eventualRead: EventualRead[F, A]): Stream[F, EventRecord[A]] def pointer: F[Option[SeqNr]] @@ -75,8 +75,8 @@ object Journal { implicit class JournalOps[F[_]](val self: Journal[F]) extends AnyVal { def withLog(key: Key, log: Log[F], config: CallTimeThresholds = CallTimeThresholds.default)( - implicit F: FlatMap[F], - measureDuration: MeasureDuration[F], + implicit F: FlatMap[F], + measureDuration: MeasureDuration[F], ): Journal[F] = { val functionKId = FunctionK.id[F] @@ -88,7 +88,7 @@ object Journal { new WithLog with Journal[F] { def append[A](events: Nel[Event[A]], metadata: RecordMetadata, headers: Headers)( - implicit kafkaWrite: KafkaWrite[F, A], + implicit kafkaWrite: KafkaWrite[F, A], ) = { for { d <- MeasureDuration[F].start @@ -167,7 +167,7 @@ object Journal { new WithLogError with Journal[F] { def append[A](events: Nel[Event[A]], metadata: RecordMetadata, headers: Headers)( - implicit kafkaWrite: KafkaWrite[F, A], + implicit kafkaWrite: KafkaWrite[F, A], ) = { logError { self.append(events, metadata, headers) @@ -214,8 +214,8 @@ object Journal { } def withMetrics( - topic: Topic, - metrics: JournalMetrics[F], + topic: Topic, + metrics: JournalMetrics[F], )(implicit F: MonadThrowable[F], measureDuration: MeasureDuration[F]): Journal[F] = { val functionKId = FunctionK.id[F] @@ -231,7 +231,7 @@ object Journal { new WithMetrics with Journal[F] { def append[A](events: Nel[Event[A]], metadata: RecordMetadata, headers: Headers)( - implicit kafkaWrite: KafkaWrite[F, A], + implicit kafkaWrite: KafkaWrite[F, A], ) = { def append = self.append(events, metadata, headers) for { @@ -293,7 +293,7 @@ object Journal { new MapK with Journal[G] { def append[A](events: Nel[Event[A]], metadata: RecordMetadata, headers: Headers)( - implicit kafkaWrite: KafkaWrite[G, A], + implicit kafkaWrite: KafkaWrite[G, A], ) = { fg(self.append(events, metadata, headers)(kafkaWrite.mapK(gf))) } @@ -311,11 +311,11 @@ object Journal { } final case class CallTimeThresholds( - append: FiniteDuration = 500.millis, - read: FiniteDuration = 5.seconds, - pointer: FiniteDuration = 1.second, - delete: FiniteDuration = 1.second, - purge: FiniteDuration = 1.second, + append: FiniteDuration = 500.millis, + read: FiniteDuration = 5.seconds, + pointer: FiniteDuration = 1.second, + delete: FiniteDuration = 1.second, + purge: FiniteDuration = 1.second, ) object CallTimeThresholds { @@ -335,8 +335,8 @@ object Journal { * @param idleTimeout if idle for this time, Kafka consumers are closed */ final case class ConsumerPoolConfig( - multiplier: Double, - idleTimeout: FiniteDuration, + multiplier: Double, + idleTimeout: FiniteDuration, ) object ConsumerPoolConfig { @@ -347,16 +347,16 @@ object Journal { } final case class DataIntegrityConfig( - /** + /** * On recovery, if true, duplicated [[SeqNr]] in events will cause [[JournalError]] `Data integrity violated` */ - seqNrUniqueness: Boolean, + seqNrUniqueness: Boolean, - /** + /** * On recovery, if true, events with [[RecordId]] different from the one in the current metadata record * will be filtered out and logged as an error. */ - correlateEventsWithMeta: Boolean, + correlateEventsWithMeta: Boolean, ) object DataIntegrityConfig { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/JournalConfig.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/JournalConfig.scala index e37e29314..4ce77a68f 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/JournalConfig.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/JournalConfig.scala @@ -23,9 +23,9 @@ import scala.concurrent.duration.* * and how it works. */ final case class JournalConfig( - pollTimeout: FiniteDuration = 10.millis, - kafka: KafkaConfig = KafkaConfig("journal"), - headCache: JournalConfig.HeadCache = JournalConfig.HeadCache.default, + pollTimeout: FiniteDuration = 10.millis, + kafka: KafkaConfig = KafkaConfig("journal"), + headCache: JournalConfig.HeadCache = JournalConfig.HeadCache.default, ) object JournalConfig { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/JournalMetrics.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/JournalMetrics.scala index fe0c1a407..0a0dfefcb 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/JournalMetrics.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/JournalMetrics.scala @@ -51,8 +51,8 @@ object JournalMetrics { } def make[F[_]: Monad]( - registry: CollectorRegistry[F], - prefix: String = "journal", + registry: CollectorRegistry[F], + prefix: String = "journal", ): Resource[F, JournalMetrics[F]] = { val latencySummary = registry.summary( diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/JournalReadWrite.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/JournalReadWrite.scala index 419575b03..4854839fb 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/JournalReadWrite.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/JournalReadWrite.scala @@ -4,17 +4,17 @@ import com.evolutiongaming.kafka.journal.conversions.{KafkaRead, KafkaWrite} import com.evolutiongaming.kafka.journal.eventual.EventualRead final case class JournalReadWrite[F[_], A]( - kafkaRead: KafkaRead[F, A], - kafkaWrite: KafkaWrite[F, A], - eventualRead: EventualRead[F, A], + kafkaRead: KafkaRead[F, A], + kafkaWrite: KafkaWrite[F, A], + eventualRead: EventualRead[F, A], ) object JournalReadWrite { def of[F[_], A]( - implicit kafkaRead: KafkaRead[F, A], - kafkaWrite: KafkaWrite[F, A], - eventualRead: EventualRead[F, A], + implicit kafkaRead: KafkaRead[F, A], + kafkaWrite: KafkaWrite[F, A], + eventualRead: EventualRead[F, A], ): JournalReadWrite[F, A] = { JournalReadWrite(kafkaRead, kafkaWrite, eventualRead) } diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/Journals.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/Journals.scala index 187c9f783..0d7701c2d 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/Journals.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/Journals.scala @@ -40,16 +40,16 @@ object Journals { } def make[ - F[_]: Async: FromTry: Fail: LogOf: KafkaConsumerOf: KafkaProducerOf: HeadCacheOf: RandomIdOf: MeasureDuration: JsonCodec, + F[_]: Async: FromTry: Fail: LogOf: KafkaConsumerOf: KafkaProducerOf: HeadCacheOf: RandomIdOf: MeasureDuration: JsonCodec, ]( - config: JournalConfig, - origin: Option[Origin], - eventualJournal: EventualJournal[F], - journalMetrics: Option[JournalMetrics[F]], - conversionMetrics: Option[ConversionMetrics[F]], - consumerPoolConfig: ConsumerPoolConfig, - consumerPoolMetrics: Option[ConsumerPoolMetrics[F]], - callTimeThresholds: Journal.CallTimeThresholds, + config: JournalConfig, + origin: Option[Origin], + eventualJournal: EventualJournal[F], + journalMetrics: Option[JournalMetrics[F]], + conversionMetrics: Option[ConversionMetrics[F]], + consumerPoolConfig: ConsumerPoolConfig, + consumerPoolMetrics: Option[ConsumerPoolMetrics[F]], + callTimeThresholds: Journal.CallTimeThresholds, ): Resource[F, Journals[F]] = { val consumer = Consumer.make[F](config.kafka.consumer, config.pollTimeout) @@ -83,13 +83,13 @@ object Journals { } def apply[F[_]: Clock: RandomIdOf: Fail: JsonCodec: MeasureDuration]( - origin: Option[Origin], - producer: Producer[F], - consumer: Resource[F, Consumer[F]], - eventualJournal: EventualJournal[F], - headCache: HeadCache[F], - log: Log[F], - conversionMetrics: Option[ConversionMetrics[F]], + origin: Option[Origin], + producer: Producer[F], + consumer: Resource[F, Consumer[F]], + eventualJournal: EventualJournal[F], + headCache: HeadCache[F], + log: Log[F], + conversionMetrics: Option[ConversionMetrics[F]], )(implicit F: MonadCancel[F, Throwable]): Journals[F] = { implicit val fromAttempt: FromAttempt[F] = FromAttempt.lift[F] implicit val fromJsResult: FromJsResult[F] = FromJsResult.lift[F] @@ -105,12 +105,12 @@ object Journals { } def apply[F[_]: RandomIdOf: MeasureDuration]( - eventual: EventualJournal[F], - consumeActionRecords: ConsumeActionRecords[F], - produce: Produce[F], - headCache: HeadCache[F], - log: Log[F], - conversionMetrics: Option[ConversionMetrics[F]], + eventual: EventualJournal[F], + consumeActionRecords: ConsumeActionRecords[F], + produce: Produce[F], + headCache: HeadCache[F], + log: Log[F], + conversionMetrics: Option[ConversionMetrics[F]], )(implicit F: MonadCancel[F, Throwable]): Journals[F] = { val appendMarker = AppendMarker(produce) @@ -170,7 +170,7 @@ object Journals { new Main with Journal[F] { def append[A](events: Nel[Event[A]], metadata: RecordMetadata, headers: Headers)( - implicit kafkaWrite: KafkaWrite[F, A], + implicit kafkaWrite: KafkaWrite[F, A], ) = { appendEvents(key, events, metadata, headers)(kafkaWriteWithMetrics) } @@ -355,7 +355,7 @@ object Journals { } def apply[F[_]: Monad: Fail]( - producer: KafkaProducer[F], + producer: KafkaProducer[F], )(implicit toBytesKey: skafka.ToBytes[F, String], toBytesValue: skafka.ToBytes[F, ByteVector]): Producer[F] = { (record: ProducerRecord[String, ByteVector]) => { @@ -388,8 +388,8 @@ object Journals { object Consumer { def make[F[_]: MonadThrowable: KafkaConsumerOf: FromTry]( - config: ConsumerConfig, - pollTimeout: FiniteDuration, + config: ConsumerConfig, + pollTimeout: FiniteDuration, ): Resource[F, Consumer[F]] = { import com.evolutiongaming.kafka.journal.util.SkafkaHelper.* @@ -403,8 +403,8 @@ object Journals { } def apply[F[_]]( - consumer: KafkaConsumer[F, String, ByteVector], - pollTimeout: FiniteDuration, + consumer: KafkaConsumer[F, String, ByteVector], + pollTimeout: FiniteDuration, ): Consumer[F] = { class Main new Main with Consumer[F] { @@ -427,8 +427,8 @@ object Journals { implicit class JournalsOps[F[_]](val self: Journals[F]) extends AnyVal { def withLog( - log: Log[F], - config: Journal.CallTimeThresholds = Journal.CallTimeThresholds.default, + log: Log[F], + config: Journal.CallTimeThresholds = Journal.CallTimeThresholds.default, )(implicit F: FlatMap[F], measureDuration: MeasureDuration[F]): Journals[F] = { (key: Key) => self(key).withLog(key, log, config) } @@ -438,7 +438,7 @@ object Journals { } def withMetrics( - metrics: JournalMetrics[F], + metrics: JournalMetrics[F], )(implicit F: MonadThrowable[F], measureDuration: MeasureDuration[F]): Journals[F] = { (key: Key) => self(key).withMetrics(key.topic, metrics) } diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaConsumer.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaConsumer.scala index 56722ef4c..b520fb619 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaConsumer.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaConsumer.scala @@ -35,7 +35,7 @@ object KafkaConsumer { def apply[F[_], K, V](implicit F: KafkaConsumer[F, K, V]): KafkaConsumer[F, K, V] = F def make[F[_]: Sync, K, V]( - consumer: Resource[F, Consumer[F, K, V]], + consumer: Resource[F, Consumer[F, K, V]], ): Resource[F, KafkaConsumer[F, K, V]] = { val result = for { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaConsumerOf.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaConsumerOf.scala index d27568bba..f35e64b76 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaConsumerOf.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaConsumerOf.scala @@ -8,7 +8,7 @@ import com.evolutiongaming.skafka.consumer.{ConsumerConfig, ConsumerMetrics, Con trait KafkaConsumerOf[F[_]] { def apply[K, V]( - config: ConsumerConfig, + config: ConsumerConfig, )(implicit fromBytesK: skafka.FromBytes[F, K], fromBytesV: skafka.FromBytes[F, V]): Resource[F, KafkaConsumer[F, K, V]] } @@ -17,7 +17,7 @@ object KafkaConsumerOf { def apply[F[_]](implicit F: KafkaConsumerOf[F]): KafkaConsumerOf[F] = F def apply[F[_]: Async: ToTry: ToFuture: MeasureDuration]( - metrics: Option[ConsumerMetrics[F]] = None, + metrics: Option[ConsumerMetrics[F]] = None, ): KafkaConsumerOf[F] = { val consumerOf = ConsumerOf.apply1(metrics) diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaProducer.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaProducer.scala index b67918e3d..4c4df0dde 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaProducer.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaProducer.scala @@ -9,7 +9,7 @@ import com.evolutiongaming.skafka.producer.* trait KafkaProducer[F[_]] { def send[K, V]( - record: ProducerRecord[K, V], + record: ProducerRecord[K, V], )(implicit toBytesK: skafka.ToBytes[F, K], toBytesV: skafka.ToBytes[F, V]): F[producer.RecordMetadata] } diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaProducerOf.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaProducerOf.scala index eabe45ede..a0d2ee794 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaProducerOf.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/KafkaProducerOf.scala @@ -15,7 +15,7 @@ object KafkaProducerOf { def apply[F[_]](implicit F: KafkaProducerOf[F]): KafkaProducerOf[F] = F def apply[F[_]: Async: MeasureDuration: ToTry]( - metrics: Option[ProducerMetrics[F]] = None, + metrics: Option[ProducerMetrics[F]] = None, ): KafkaProducerOf[F] = { val producerOf = ProducerOf.apply1(metrics) apply(producerOf) diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/PartitionCache.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/PartitionCache.scala index f8b48f898..c9524ace9 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/PartitionCache.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/PartitionCache.scala @@ -116,18 +116,18 @@ private[journal] object PartitionCache { * parameters did not change. */ def make[F[_]: Async]( - maxSize: Int = 10000, - dropUponLimit: Double = 0.1, - timeout: FiniteDuration, + maxSize: Int = 10000, + dropUponLimit: Double = 0.1, + timeout: FiniteDuration, ): Resource[F, PartitionCache[F]] = { main(maxSize = maxSize.max(1), dropUponLimit = dropUponLimit.max(0.01).min(1.0), timeout = timeout) } /** Same as [[#of]], but without default parameters */ private def main[F[_]: Async]( - maxSize: Int, - dropUponLimit: Double, - timeout: FiniteDuration, + maxSize: Int, + dropUponLimit: Double, + timeout: FiniteDuration, ): Resource[F, PartitionCache[F]] = { /** Listener waiting for latest [[HeadInfo]] to appear in the cache. @@ -155,10 +155,10 @@ private[journal] object PartitionCache { * Listener waiting for latest [[HeadInfo]] to appear in the cache. */ final case class State( - listenerId: ListenerId, - offset: Option[Offset], - entries: Option[Entries], - listeners: Map[ListenerId, Listener], + listenerId: ListenerId, + offset: Option[Offset], + entries: Option[Entries], + listeners: Map[ListenerId, Listener], ) { self => /** Checks if current state is ahead of Cassandra. diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/PartitionOffset.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/PartitionOffset.scala index 0ba63b658..79e3a653e 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/PartitionOffset.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/PartitionOffset.scala @@ -9,8 +9,8 @@ import com.evolutiongaming.skafka.consumer.ConsumerRecord import com.evolutiongaming.skafka.{Offset, Partition} final case class PartitionOffset( - partition: Partition = Partition.min, - offset: Offset = Offset.min, + partition: Partition = Partition.min, + offset: Offset = Offset.min, ) { override def toString = s"$partition:$offset" } diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/PayloadAndType.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/PayloadAndType.scala index 8ffec0e0e..ec72ae0ae 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/PayloadAndType.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/PayloadAndType.scala @@ -49,10 +49,10 @@ object PayloadAndType { * `Json` for Circe. */ final case class EventJson[A]( - seqNr: SeqNr, - tags: Tags, - payloadType: Option[PayloadType.TextOrJson] = None, - payload: Option[A] = None, + seqNr: SeqNr, + tags: Tags, + payloadType: Option[PayloadType.TextOrJson] = None, + payload: Option[A] = None, ) object EventJson { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/Produce.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/Produce.scala index fa95e8122..1aa6c6180 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/Produce.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/Produce.scala @@ -13,11 +13,11 @@ import java.time.Instant private[journal] trait Produce[F[_]] { def append( - key: Key, - range: SeqRange, - payloadAndType: PayloadAndType, - metadata: HeaderMetadata, - headers: Headers, + key: Key, + range: SeqRange, + payloadAndType: PayloadAndType, + metadata: HeaderMetadata, + headers: Headers, ): F[PartitionOffset] def delete(key: Key, to: DeleteTo): F[PartitionOffset] @@ -35,11 +35,11 @@ private[journal] object Produce { class Const new Const with Produce[F] { def append( - key: Key, - range: SeqRange, - payloadAndType: PayloadAndType, - metadata: HeaderMetadata, - headers: Headers, + key: Key, + range: SeqRange, + payloadAndType: PayloadAndType, + metadata: HeaderMetadata, + headers: Headers, ): F[PartitionOffset] = { partitionOffset } @@ -53,16 +53,16 @@ private[journal] object Produce { } def apply[F[_]: MonadThrowable: Clock](producer: Journals.Producer[F], origin: Option[Origin])( - implicit actionToProducerRecord: ActionToProducerRecord[F], + implicit actionToProducerRecord: ActionToProducerRecord[F], ): Produce[F] = { val produceAction = ProduceAction(producer) apply(produceAction, origin) } def apply[F[_]: MonadThrowable: Clock]( - produceAction: ProduceAction[F], - origin: Option[Origin], - version: Version = Version.current, + produceAction: ProduceAction[F], + origin: Option[Origin], + version: Version = Version.current, ): Produce[F] = { def send(action: Action) = { @@ -76,11 +76,11 @@ private[journal] object Produce { new Main with Produce[F] { def append( - key: Key, - range: SeqRange, - payloadAndType: PayloadAndType, - metadata: HeaderMetadata, - headers: Headers, + key: Key, + range: SeqRange, + payloadAndType: PayloadAndType, + metadata: HeaderMetadata, + headers: Headers, ): F[PartitionOffset] = { def actionOf(timestamp: Instant) = { @@ -140,11 +140,11 @@ private[journal] object Produce { def mapK[G[_]](f: F ~> G): Produce[G] = new MapK with Produce[G] { def append( - key: Key, - range: SeqRange, - payloadAndType: PayloadAndType, - metadata: HeaderMetadata, - headers: Headers, + key: Key, + range: SeqRange, + payloadAndType: PayloadAndType, + metadata: HeaderMetadata, + headers: Headers, ): G[PartitionOffset] = { f(self.append(key, range, payloadAndType, metadata, headers)) } diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/ProduceAction.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/ProduceAction.scala index 2b8f0e125..27307b485 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/ProduceAction.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/ProduceAction.scala @@ -12,7 +12,7 @@ trait ProduceAction[F[_]] { object ProduceAction { def apply[F[_]: Monad]( - producer: Journals.Producer[F], + producer: Journals.Producer[F], )(implicit actionToProducerRecord: ActionToProducerRecord[F]): ProduceAction[F] = { (action: Action) => { for { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/StreamActionRecords.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/StreamActionRecords.scala index f6f67791b..a0874949f 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/StreamActionRecords.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/StreamActionRecords.scala @@ -60,12 +60,12 @@ object StreamActionRecords { * Underlying reader of Kafka records. */ def apply[F[_]: BracketThrowable]( - // TODO add range argument - key: Key, - from: SeqNr, - marker: Marker, - offsetReplicated: Option[Offset], - consumeActionRecords: ConsumeActionRecords[F], + // TODO add range argument + key: Key, + from: SeqNr, + marker: Marker, + offsetReplicated: Option[Offset], + consumeActionRecords: ConsumeActionRecords[F], ): StreamActionRecords[F] = { // TODO compare partitions ! diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/TopicCache.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/TopicCache.scala index 08e295b1f..1eb6b7ff3 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/TopicCache.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/TopicCache.scala @@ -77,13 +77,13 @@ private[journal] object TopicCache { * did not change. */ def make[F[_]: Async: Parallel]( - eventual: Eventual[F], - topic: Topic, - log: Log[F], - consumer: Resource[F, Consumer[F]], - config: HeadCacheConfig, - consRecordToActionHeader: ConsRecordToActionHeader[F], - metrics: Option[HeadCache.Metrics[F]], + eventual: Eventual[F], + topic: Topic, + log: Log[F], + consumer: Resource[F, Consumer[F]], + config: HeadCacheConfig, + consRecordToActionHeader: ConsRecordToActionHeader[F], + metrics: Option[HeadCache.Metrics[F]], ): Resource[F, TopicCache[F]] = { for { @@ -319,8 +319,8 @@ private[journal] object TopicCache { * @param pollTimeout The timeout to use for [[KafkaConsumer#poll]]. */ def apply[F[_]: Monad]( - consumer: KafkaConsumer[F, String, Unit], - pollTimeout: FiniteDuration, + consumer: KafkaConsumer[F, String, Unit], + pollTimeout: FiniteDuration, ): Consumer[F] = { class Main @@ -358,8 +358,8 @@ private[journal] object TopicCache { * The timeout to use for [[KafkaConsumer#poll]]. */ def make[F[_]: Monad: KafkaConsumerOf: FromTry]( - config: ConsumerConfig, - pollTimeout: FiniteDuration = 10.millis, + config: ConsumerConfig, + pollTimeout: FiniteDuration = 10.millis, ): Resource[F, Consumer[F]] = { val config1 = config.copy(autoOffsetReset = AutoOffsetReset.Earliest, groupId = None, autoCommit = false) for { @@ -482,8 +482,8 @@ private[journal] object TopicCache { /** Wrap instance in a class, which logs metrics to [[HeadCache.Metrics]] */ def withMetrics( - topic: Topic, - metrics: HeadCache.Metrics[F], + topic: Topic, + metrics: HeadCache.Metrics[F], )(implicit F: MonadThrowable[F], measureDuration: MeasureDuration[F]): TopicCache[F] = { new WithMetrics with TopicCache[F] { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ActionHeaderToHeader.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ActionHeaderToHeader.scala index 1396e3c0e..a2e7e8959 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ActionHeaderToHeader.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ActionHeaderToHeader.scala @@ -13,7 +13,7 @@ trait ActionHeaderToHeader[F[_]] { object ActionHeaderToHeader { implicit def apply[F[_]: ApplicativeThrowable]( - implicit actionHeaderToBytes: ToBytes[F, ActionHeader], + implicit actionHeaderToBytes: ToBytes[F, ActionHeader], ): ActionHeaderToHeader[F] = { (actionHeader: ActionHeader) => { val result = for { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ActionToProducerRecord.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ActionToProducerRecord.scala index 2b08d04c5..fb8956440 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ActionToProducerRecord.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ActionToProducerRecord.scala @@ -15,8 +15,8 @@ trait ActionToProducerRecord[F[_]] { object ActionToProducerRecord { implicit def apply[F[_]: MonadThrowable]( - implicit actionHeaderToHeader: ActionHeaderToHeader[F], - tupleToHeader: TupleToHeader[F], + implicit actionHeaderToHeader: ActionHeaderToHeader[F], + tupleToHeader: TupleToHeader[F], ): ActionToProducerRecord[F] = { (action: Action) => { val key = action.key diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ConsRecordToActionRecord.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ConsRecordToActionRecord.scala index c4904e6f3..53a19b024 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ConsRecordToActionRecord.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ConsRecordToActionRecord.scala @@ -16,8 +16,8 @@ trait ConsRecordToActionRecord[F[_]] { object ConsRecordToActionRecord { implicit def apply[F[_]: MonadThrowable]( - implicit consRecordToActionHeader: ConsRecordToActionHeader[F], - headerToTuple: HeaderToTuple[F], + implicit consRecordToActionHeader: ConsRecordToActionHeader[F], + headerToTuple: HeaderToTuple[F], ): ConsRecordToActionRecord[F] = { (consRecord: ConsRecord) => { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ConversionMetrics.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ConversionMetrics.scala index 833f2ed52..be2b48db5 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ConversionMetrics.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/ConversionMetrics.scala @@ -1,6 +1,6 @@ package com.evolutiongaming.kafka.journal.conversions final case class ConversionMetrics[F[_]]( - kafkaRead: KafkaReadMetrics[F], - kafkaWrite: KafkaWriteMetrics[F], + kafkaRead: KafkaReadMetrics[F], + kafkaWrite: KafkaWriteMetrics[F], ) diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaRead.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaRead.scala index 24379841c..81614ccf2 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaRead.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaRead.scala @@ -23,8 +23,8 @@ object KafkaRead { def summon[F[_], A](implicit kafkaRead: KafkaRead[F, A]): KafkaRead[F, A] = kafkaRead implicit def payloadKafkaRead[F[_]: MonadThrowable: FromJsResult]( - implicit eventsFromBytes: FromBytes[F, Events[Payload]], - payloadJsonFromBytes: FromBytes[F, PayloadJson[JsValue]], + implicit eventsFromBytes: FromBytes[F, Events[Payload]], + payloadJsonFromBytes: FromBytes[F, PayloadJson[JsValue]], ): KafkaRead[F, Payload] = { (payloadAndType: PayloadAndType) => { @@ -55,8 +55,8 @@ object KafkaRead { } def readJson[F[_]: MonadThrowable, A, B]( - payloadJsonFromBytes: FromBytes[F, PayloadJson[A]], - fromEventJsonPayload: EventJsonPayloadAndType[A] => F[B], + payloadJsonFromBytes: FromBytes[F, PayloadJson[A]], + fromEventJsonPayload: EventJsonPayloadAndType[A] => F[B], ): KafkaRead[F, B] = { (payloadAndType: PayloadAndType) => { @@ -105,10 +105,10 @@ object KafkaRead { implicit class KafkaReadOps[F[_], A](val self: KafkaRead[F, A]) extends AnyVal { def withMetrics( - metrics: KafkaReadMetrics[F], + metrics: KafkaReadMetrics[F], )( - implicit F: Monad[F], - measureDuration: MeasureDuration[F], + implicit F: Monad[F], + measureDuration: MeasureDuration[F], ): KafkaRead[F, A] = { payloadAndType => for { d <- MeasureDuration[F].start diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaReadMetrics.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaReadMetrics.scala index 5ae5fb3c6..378968b2e 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaReadMetrics.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaReadMetrics.scala @@ -18,8 +18,8 @@ object KafkaReadMetrics { def empty[F[_]: Applicative]: KafkaReadMetrics[F] = (_, _) => Applicative[F].unit def make[F[_]]( - registry: CollectorRegistry[F], - prefix: String = "journal", + registry: CollectorRegistry[F], + prefix: String = "journal", ): Resource[F, KafkaReadMetrics[F]] = { val durationSummary = registry.summary( diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaWrite.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaWrite.scala index f1f4e628d..d0b731301 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaWrite.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaWrite.scala @@ -25,8 +25,8 @@ object KafkaWrite { def summon[F[_], A](implicit kafkaWrite: KafkaWrite[F, A]): KafkaWrite[F, A] = kafkaWrite implicit def payloadKafkaWrite[F[_]: MonadThrowable]( - implicit eventsToBytes: ToBytes[F, Events[Payload]], - payloadJsonToBytes: ToBytes[F, PayloadJson[JsValue]], + implicit eventsToBytes: ToBytes[F, Events[Payload]], + payloadJsonToBytes: ToBytes[F, PayloadJson[JsValue]], ): KafkaWrite[F, Payload] = { (events: Events[Payload]) => { @@ -95,8 +95,8 @@ object KafkaWrite { } def writeJson[F[_]: MonadThrowable, A, B]( - toEventJsonPayload: A => EventJsonPayloadAndType[B], - payloadJsonToBytes: ToBytes[F, PayloadJson[B]], + toEventJsonPayload: A => EventJsonPayloadAndType[B], + payloadJsonToBytes: ToBytes[F, PayloadJson[B]], ): KafkaWrite[F, A] = { (events: Events[A]) => { @@ -130,10 +130,10 @@ object KafkaWrite { implicit class KafkaWriteOps[F[_], A](val self: KafkaWrite[F, A]) extends AnyVal { def withMetrics( - metrics: KafkaWriteMetrics[F], + metrics: KafkaWriteMetrics[F], )( - implicit F: Monad[F], - measureDuration: MeasureDuration[F], + implicit F: Monad[F], + measureDuration: MeasureDuration[F], ): KafkaWrite[F, A] = { events => for { d <- MeasureDuration[F].start diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaWriteMetrics.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaWriteMetrics.scala index 753aeccc8..fe90cf72f 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaWriteMetrics.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/conversions/KafkaWriteMetrics.scala @@ -21,8 +21,8 @@ object KafkaWriteMetrics { } def make[F[_]]( - registry: CollectorRegistry[F], - prefix: String = "journal", + registry: CollectorRegistry[F], + prefix: String = "journal", ): Resource[F, KafkaWriteMetrics[F]] = { val durationSummary = registry.summary( diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/EventualJournal.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/EventualJournal.scala index 43d09c507..00453ce5d 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/EventualJournal.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/EventualJournal.scala @@ -133,8 +133,8 @@ object EventualJournal { private sealed abstract class Main def make[F[_]]( - registry: CollectorRegistry[F], - prefix: String = "eventual_journal", + registry: CollectorRegistry[F], + prefix: String = "eventual_journal", ): Resource[F, Metrics[F]] = { val versionGauge = registry.gauge( diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedJournal.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedJournal.scala index 36270408d..13c41c181 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedJournal.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedJournal.scala @@ -177,8 +177,8 @@ object ReplicatedJournal { } def make[F[_]: Monad]( - registry: CollectorRegistry[F], - prefix: String = "replicated_journal", + registry: CollectorRegistry[F], + prefix: String = "replicated_journal", ): Resource[F, Metrics[F]] = { val versionGauge = registry.gauge( diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedJournalFlat.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedJournalFlat.scala index 3feca39eb..bca6e8bb2 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedJournalFlat.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedJournalFlat.scala @@ -31,28 +31,28 @@ trait ReplicatedJournalFlat[F[_]] { def offsetUpdate(topic: Topic, partition: Partition, offset: Offset, timestamp: Instant): F[Unit] def append( - key: Key, - partition: Partition, - offset: Offset, - timestamp: Instant, - expireAfter: Option[ExpireAfter], - events: Nel[EventRecord[EventualPayloadAndType]], + key: Key, + partition: Partition, + offset: Offset, + timestamp: Instant, + expireAfter: Option[ExpireAfter], + events: Nel[EventRecord[EventualPayloadAndType]], ): F[Changed] def delete( - key: Key, - partition: Partition, - offset: Offset, - timestamp: Instant, - deleteTo: DeleteTo, - origin: Option[Origin], + key: Key, + partition: Partition, + offset: Offset, + timestamp: Instant, + deleteTo: DeleteTo, + origin: Option[Origin], ): F[Changed] def purge( - key: Key, - partition: Partition, - offset: Offset, - timestamp: Instant, + key: Key, + partition: Partition, + offset: Offset, + timestamp: Instant, ): F[Changed] } @@ -107,12 +107,12 @@ object ReplicatedJournalFlat { } def append( - key: Key, - partition: Partition, - offset: Offset, - timestamp: Instant, - expireAfter: Option[ExpireAfter], - events: Nel[EventRecord[EventualPayloadAndType]], + key: Key, + partition: Partition, + offset: Offset, + timestamp: Instant, + expireAfter: Option[ExpireAfter], + events: Nel[EventRecord[EventualPayloadAndType]], ) = { replicatedJournal .journal(key.topic) @@ -128,12 +128,12 @@ object ReplicatedJournalFlat { } def delete( - key: Key, - partition: Partition, - offset: Offset, - timestamp: Instant, - deleteTo: DeleteTo, - origin: Option[Origin], + key: Key, + partition: Partition, + offset: Offset, + timestamp: Instant, + deleteTo: DeleteTo, + origin: Option[Origin], ) = { replicatedJournal .journal(key.topic) @@ -149,10 +149,10 @@ object ReplicatedJournalFlat { } def purge( - key: Key, - partition: Partition, - offset: Offset, - timestamp: Instant, + key: Key, + partition: Partition, + offset: Offset, + timestamp: Instant, ) = { replicatedJournal .journal(key.topic) diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedKeyJournal.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedKeyJournal.scala index 4b98172c4..59e01933d 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedKeyJournal.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedKeyJournal.scala @@ -13,22 +13,22 @@ import java.time.Instant trait ReplicatedKeyJournal[F[_]] { def append( - offset: Offset, - timestamp: Instant, - expireAfter: Option[ExpireAfter], - events: Nel[EventRecord[EventualPayloadAndType]], + offset: Offset, + timestamp: Instant, + expireAfter: Option[ExpireAfter], + events: Nel[EventRecord[EventualPayloadAndType]], ): F[Changed] def delete( - offset: Offset, - timestamp: Instant, - deleteTo: DeleteTo, - origin: Option[Origin], + offset: Offset, + timestamp: Instant, + deleteTo: DeleteTo, + origin: Option[Origin], ): F[Changed] def purge( - offset: Offset, - timestamp: Instant, + offset: Offset, + timestamp: Instant, ): F[Changed] } @@ -43,22 +43,22 @@ object ReplicatedKeyJournal { new Const with ReplicatedKeyJournal[F] { def append( - offset: Offset, - timestamp: Instant, - expireAfter: Option[ExpireAfter], - events: Nel[EventRecord[EventualPayloadAndType]], + offset: Offset, + timestamp: Instant, + expireAfter: Option[ExpireAfter], + events: Nel[EventRecord[EventualPayloadAndType]], ) = value def delete( - offset: Offset, - timestamp: Instant, - deleteTo: DeleteTo, - origin: Option[Origin], + offset: Offset, + timestamp: Instant, + deleteTo: DeleteTo, + origin: Option[Origin], ) = value def purge( - offset: Offset, - timestamp: Instant, + offset: Offset, + timestamp: Instant, ) = value } } @@ -75,26 +75,26 @@ object ReplicatedKeyJournal { new MapK with ReplicatedKeyJournal[G] { def append( - offset: Offset, - timestamp: Instant, - expireAfter: Option[ExpireAfter], - events: Nel[EventRecord[EventualPayloadAndType]], + offset: Offset, + timestamp: Instant, + expireAfter: Option[ExpireAfter], + events: Nel[EventRecord[EventualPayloadAndType]], ) = { f(self.append(offset, timestamp, expireAfter, events)) } def delete( - offset: Offset, - timestamp: Instant, - deleteTo: DeleteTo, - origin: Option[Origin], + offset: Offset, + timestamp: Instant, + deleteTo: DeleteTo, + origin: Option[Origin], ) = { f(self.delete(offset, timestamp, deleteTo, origin)) } def purge( - offset: Offset, - timestamp: Instant, + offset: Offset, + timestamp: Instant, ) = { f(self.purge(offset, timestamp)) } @@ -102,17 +102,17 @@ object ReplicatedKeyJournal { } def withLog(key: Key, partition: Partition, log: Log[F])( - implicit F: FlatMap[F], - measureDuration: MeasureDuration[F], + implicit F: FlatMap[F], + measureDuration: MeasureDuration[F], ): ReplicatedKeyJournal[F] = { new WithLog with ReplicatedKeyJournal[F] { def append( - offset: Offset, - timestamp: Instant, - expireAfter: Option[ExpireAfter], - events: Nel[EventRecord[EventualPayloadAndType]], + offset: Offset, + timestamp: Instant, + expireAfter: Option[ExpireAfter], + events: Nel[EventRecord[EventualPayloadAndType]], ) = { for { d <- MeasureDuration[F].start @@ -130,10 +130,10 @@ object ReplicatedKeyJournal { } def delete( - offset: Offset, - timestamp: Instant, - deleteTo: DeleteTo, - origin: Option[Origin], + offset: Offset, + timestamp: Instant, + deleteTo: DeleteTo, + origin: Option[Origin], ) = { for { d <- MeasureDuration[F].start @@ -147,8 +147,8 @@ object ReplicatedKeyJournal { } def purge( - offset: Offset, - timestamp: Instant, + offset: Offset, + timestamp: Instant, ) = { for { d <- MeasureDuration[F].start @@ -161,16 +161,16 @@ object ReplicatedKeyJournal { } def withMetrics( - topic: Topic, - metrics: ReplicatedJournal.Metrics[F], + topic: Topic, + metrics: ReplicatedJournal.Metrics[F], )(implicit F: FlatMap[F], measureDuration: MeasureDuration[F]): ReplicatedKeyJournal[F] = { new WithMetrics with ReplicatedKeyJournal[F] { def append( - offset: Offset, - timestamp: Instant, - expireAfter: Option[ExpireAfter], - events: Nel[EventRecord[EventualPayloadAndType]], + offset: Offset, + timestamp: Instant, + expireAfter: Option[ExpireAfter], + events: Nel[EventRecord[EventualPayloadAndType]], ) = { for { d <- MeasureDuration[F].start @@ -181,10 +181,10 @@ object ReplicatedKeyJournal { } def delete( - offset: Offset, - timestamp: Instant, - deleteTo: DeleteTo, - origin: Option[Origin], + offset: Offset, + timestamp: Instant, + deleteTo: DeleteTo, + origin: Option[Origin], ) = { for { d <- MeasureDuration[F].start @@ -195,8 +195,8 @@ object ReplicatedKeyJournal { } def purge( - offset: Offset, - timestamp: Instant, + offset: Offset, + timestamp: Instant, ) = { for { d <- MeasureDuration[F].start @@ -217,10 +217,10 @@ object ReplicatedKeyJournal { new ReplicatedKeyJournal[F] { def append( - offset: Offset, - timestamp: Instant, - expireAfter: Option[ExpireAfter], - events: Nel[EventRecord[EventualPayloadAndType]], + offset: Offset, + timestamp: Instant, + expireAfter: Option[ExpireAfter], + events: Nel[EventRecord[EventualPayloadAndType]], ) = { self .append(offset, timestamp, expireAfter, events) @@ -238,10 +238,10 @@ object ReplicatedKeyJournal { } def delete( - offset: Offset, - timestamp: Instant, - deleteTo: DeleteTo, - origin: Option[Origin], + offset: Offset, + timestamp: Instant, + deleteTo: DeleteTo, + origin: Option[Origin], ) = { self .delete(offset, timestamp, deleteTo, origin) @@ -259,8 +259,8 @@ object ReplicatedKeyJournal { } def purge( - offset: Offset, - timestamp: Instant, + offset: Offset, + timestamp: Instant, ) = { self .purge(offset, timestamp) diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedPartitionJournal.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedPartitionJournal.scala index c3a6c6e6c..570ef1fcc 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedPartitionJournal.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedPartitionJournal.scala @@ -85,8 +85,8 @@ object ReplicatedPartitionJournal { } def withLog(topic: Topic, partition: Partition, log: Log[F])( - implicit F: Monad[F], - measureDuration: MeasureDuration[F], + implicit F: Monad[F], + measureDuration: MeasureDuration[F], ): ReplicatedPartitionJournal[F] = { new WithLog with ReplicatedPartitionJournal[F] { @@ -135,8 +135,8 @@ object ReplicatedPartitionJournal { } def withMetrics( - topic: Topic, - metrics: ReplicatedJournal.Metrics[F], + topic: Topic, + metrics: ReplicatedJournal.Metrics[F], )(implicit F: Monad[F], measureDuration: MeasureDuration[F]): ReplicatedPartitionJournal[F] = { new WithMetrics with ReplicatedPartitionJournal[F] { diff --git a/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedTopicJournal.scala b/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedTopicJournal.scala index 490a66034..2a7f17242 100644 --- a/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedTopicJournal.scala +++ b/journal/src/main/scala/com/evolutiongaming/kafka/journal/eventual/ReplicatedTopicJournal.scala @@ -69,8 +69,8 @@ object ReplicatedTopicJournal { } def withLog( - topic: Topic, - log: Log[F], + topic: Topic, + log: Log[F], )(implicit F: Monad[F], measureDuration: MeasureDuration[F]): ReplicatedTopicJournal[F] = { new WithLog with ReplicatedTopicJournal[F] { @@ -83,8 +83,8 @@ object ReplicatedTopicJournal { } def withMetrics( - topic: Topic, - metrics: ReplicatedJournal.Metrics[F], + topic: Topic, + metrics: ReplicatedJournal.Metrics[F], )(implicit F: Monad[F], measureDuration: MeasureDuration[F]): ReplicatedTopicJournal[F] = { new WithMetrics with ReplicatedTopicJournal[F] { diff --git a/journal/src/test/scala/com/evolutiongaming/kafka/journal/ConsumerRecordOf.scala b/journal/src/test/scala/com/evolutiongaming/kafka/journal/ConsumerRecordOf.scala index ad0c538ae..7fdd12fbc 100644 --- a/journal/src/test/scala/com/evolutiongaming/kafka/journal/ConsumerRecordOf.scala +++ b/journal/src/test/scala/com/evolutiongaming/kafka/journal/ConsumerRecordOf.scala @@ -9,7 +9,7 @@ import com.evolutiongaming.skafka.{Offset, TimestampAndType, TimestampType, Topi object ConsumerRecordOf { def apply[F[_]: Functor](action: Action, topicPartition: TopicPartition, offset: Offset)( - implicit actionToProducerRecord: ActionToProducerRecord[F], + implicit actionToProducerRecord: ActionToProducerRecord[F], ): F[ConsRecord] = { for { diff --git a/journal/src/test/scala/com/evolutiongaming/kafka/journal/HeadCacheSpec.scala b/journal/src/test/scala/com/evolutiongaming/kafka/journal/HeadCacheSpec.scala index fed2e8f65..350ba6f41 100644 --- a/journal/src/test/scala/com/evolutiongaming/kafka/journal/HeadCacheSpec.scala +++ b/journal/src/test/scala/com/evolutiongaming/kafka/journal/HeadCacheSpec.scala @@ -245,7 +245,7 @@ object HeadCacheSpec { val headers: Headers = Headers.empty def consumerRecordOf(action: Action, topicPartition: TopicPartition, offset: Offset)( - implicit actionToProducerRecord: ActionToProducerRecord[Try], + implicit actionToProducerRecord: ActionToProducerRecord[Try], ): ConsumerRecord[String, Unit] = { ConsumerRecordOf[Try](action, topicPartition, offset) .get @@ -278,9 +278,9 @@ object HeadCacheSpec { } def headCacheOf( - eventual: HeadCache.Eventual[IO], - consumer: Resource[IO, TopicCache.Consumer[IO]], - config: HeadCacheConfig = config, + eventual: HeadCache.Eventual[IO], + consumer: Resource[IO, TopicCache.Consumer[IO]], + config: HeadCacheConfig = config, ): Resource[IO, HeadCache[IO]] = { for { @@ -347,9 +347,9 @@ object HeadCacheSpec { } final case class State( - actions: List[Action] = List.empty, - topics: Map[Topic, List[Partition]] = Map.empty, - records: Queue[Try[ConsumerRecords[String, Unit]]] = Queue.empty, + actions: List[Action] = List.empty, + topics: Map[Topic, List[Partition]] = Map.empty, + records: Queue[Try[ConsumerRecords[String, Unit]]] = Queue.empty, ) object State { diff --git a/journal/src/test/scala/com/evolutiongaming/kafka/journal/JournalSpec.scala b/journal/src/test/scala/com/evolutiongaming/kafka/journal/JournalSpec.scala index 28875c3da..1e8ee9c01 100644 --- a/journal/src/test/scala/com/evolutiongaming/kafka/journal/JournalSpec.scala +++ b/journal/src/test/scala/com/evolutiongaming/kafka/journal/JournalSpec.scala @@ -258,10 +258,10 @@ class JournalSpec extends AnyWordSpec with Matchers { } { def test( - eventual: EventualJournal[StateT], - consumeActionRecords: ConsumeActionRecords[StateT], - produceAction: ProduceAction[StateT], - headCache: HeadCache[StateT], + eventual: EventualJournal[StateT], + consumeActionRecords: ConsumeActionRecords[StateT], + produceAction: ProduceAction[StateT], + headCache: HeadCache[StateT], ): Unit = { val journal = SeqNrJournal(eventualJournalOf(eventual), consumeActionRecordsOf(consumeActionRecords), produceAction, headCache) @@ -411,13 +411,13 @@ object JournalSpec { object SeqNrJournal { def apply[F[_]: Monad, A]( - journals: Journals[F], + journals: Journals[F], )(implicit kafkaRead: KafkaRead[F, A], eventualRead: EventualRead[F, A], kafkaWrite: KafkaWrite[F, A]): SeqNrJournal[F] = { apply(journals(key)) } def apply[F[_]: Monad, A]( - journal: Journal[F], + journal: Journal[F], )(implicit kafkaRead: KafkaRead[F, A], eventualRead: EventualRead[F, A], kafkaWrite: KafkaWrite[F, A]): SeqNrJournal[F] = { new SeqNrJournal[F] { @@ -469,10 +469,10 @@ object JournalSpec { } def apply[F[_]: Sync]( - eventual: EventualJournal[F], - consumeActionRecords: ConsumeActionRecords[F], - produceAction: ProduceAction[F], - headCache: HeadCache[F], + eventual: EventualJournal[F], + consumeActionRecords: ConsumeActionRecords[F], + produceAction: ProduceAction[F], + headCache: HeadCache[F], ): SeqNrJournal[F] = { implicit val clock = Clock.const[F](nanos = 0, millis = timestamp.toEpochMilli) implicit val randomIdOf = RandomIdOf.uuid[F] @@ -498,9 +498,9 @@ object JournalSpec { } final case class State( - records: Queue[ActionRecord[Action]] = Queue.empty, - replicatedState: EventualJournalOf.State = EventualJournalOf.State.empty, - recordsToRead: Queue[ActionRecord[Action]] = Queue.empty, + records: Queue[ActionRecord[Action]] = Queue.empty, + replicatedState: EventualJournalOf.State = EventualJournalOf.State.empty, + recordsToRead: Queue[ActionRecord[Action]] = Queue.empty, ) object State { @@ -615,9 +615,9 @@ object JournalSpec { object EventualJournalOf { final case class State( - events: Queue[EventRecord[EventualPayloadAndType]] = Queue.empty, - deleteTo: Option[DeleteTo] = None, - offset: Option[Offset] = None, + events: Queue[EventRecord[EventualPayloadAndType]] = Queue.empty, + deleteTo: Option[DeleteTo] = None, + offset: Option[Offset] = None, ) { def apply(record: ActionRecord[Action]): State = { diff --git a/journal/src/test/scala/com/evolutiongaming/kafka/journal/PartitionCacheSpec.scala b/journal/src/test/scala/com/evolutiongaming/kafka/journal/PartitionCacheSpec.scala index 35af2d1b8..0ac0a962b 100644 --- a/journal/src/test/scala/com/evolutiongaming/kafka/journal/PartitionCacheSpec.scala +++ b/journal/src/test/scala/com/evolutiongaming/kafka/journal/PartitionCacheSpec.scala @@ -19,9 +19,9 @@ class PartitionCacheSpec extends AsyncFunSuite with Matchers { import PartitionCacheSpec.* private def partitionCacheOf( - maxSize: Int = 10, - dropUponLimit: Double = 0.1, - timeout: FiniteDuration = 1.minute, + maxSize: Int = 10, + dropUponLimit: Double = 0.1, + timeout: FiniteDuration = 1.minute, ) = { PartitionCache.make[IO](maxSize, dropUponLimit, timeout) } diff --git a/journal/src/test/scala/com/evolutiongaming/kafka/journal/StreamActionRecordsSpec.scala b/journal/src/test/scala/com/evolutiongaming/kafka/journal/StreamActionRecordsSpec.scala index 015f7b051..325af2d50 100644 --- a/journal/src/test/scala/com/evolutiongaming/kafka/journal/StreamActionRecordsSpec.scala +++ b/journal/src/test/scala/com/evolutiongaming/kafka/journal/StreamActionRecordsSpec.scala @@ -63,9 +63,9 @@ object StreamActionRecordsSpec { } private def seqNrs( - replicated: Option[Long], - offset: Option[Long], - pointers: List[Pointer], + replicated: Option[Long], + offset: Option[Long], + pointers: List[Pointer], ) = { val timestamp = Instant.now() diff --git a/journal/src/test/scala/com/evolutiongaming/kafka/journal/eventual/EventualJournalSpec.scala b/journal/src/test/scala/com/evolutiongaming/kafka/journal/eventual/EventualJournalSpec.scala index fd5f53bed..e435266e0 100644 --- a/journal/src/test/scala/com/evolutiongaming/kafka/journal/eventual/EventualJournalSpec.scala +++ b/journal/src/test/scala/com/evolutiongaming/kafka/journal/eventual/EventualJournalSpec.scala @@ -57,7 +57,7 @@ trait EventualJournalSpec extends AnyWordSpec with Matchers { } private def test1[F[_]: MonadThrowable: Fail]( - withJournals: (Key, Instant) => ((Eventual[F], Replicated[F]) => F[Assertion]) => F[Assertion], + withJournals: (Key, Instant) => ((Eventual[F], Replicated[F]) => F[Assertion]) => F[Assertion], ): Unit = { implicit val monoidUnit = Applicative.monoid[F, Unit] @@ -539,9 +539,9 @@ object EventualJournalSpec { object Replicated { def apply[F[_]: FlatMap]( - journal: ReplicatedJournalFlat[F], - key: Key, - timestamp: Instant, + journal: ReplicatedJournalFlat[F], + key: Key, + timestamp: Instant, ): Replicated[F] = { new Replicated[F] { diff --git a/persistence/src/main/scala/akka/persistence/kafka/journal/EventSerializer.scala b/persistence/src/main/scala/akka/persistence/kafka/journal/EventSerializer.scala index c1f30f8f3..54bc20fad 100644 --- a/persistence/src/main/scala/akka/persistence/kafka/journal/EventSerializer.scala +++ b/persistence/src/main/scala/akka/persistence/kafka/journal/EventSerializer.scala @@ -131,7 +131,7 @@ object EventSerializer { * `Any` type and will accept anything assigned). */ def apply[F[_]: MonadThrowable: FromAttempt: FromJsResult]( - serializer: SerializedMsgSerializer[F], + serializer: SerializedMsgSerializer[F], ): EventSerializer[F, Payload] = { def toEventPayload(repr: PersistentRepresentation): F[Payload] = { @@ -208,8 +208,8 @@ object EventSerializer { * be used. */ def apply[F[_]: MonadThrowable, A]( - toEventPayload: PersistentRepresentation => F[A], - fromEventPayload: A => F[PersistentRepresentation], + toEventPayload: PersistentRepresentation => F[A], + fromEventPayload: A => F[PersistentRepresentation], ): EventSerializer[F, A] = new EventSerializer[F, A] { implicit val fail: Fail[F] = Fail.lift[F] diff --git a/persistence/src/main/scala/akka/persistence/kafka/journal/JournalAdapter.scala b/persistence/src/main/scala/akka/persistence/kafka/journal/JournalAdapter.scala index ed70284e8..bb94dff0f 100644 --- a/persistence/src/main/scala/akka/persistence/kafka/journal/JournalAdapter.scala +++ b/persistence/src/main/scala/akka/persistence/kafka/journal/JournalAdapter.scala @@ -32,21 +32,21 @@ trait JournalAdapter[F[_]] { object JournalAdapter { def make[ - F[ - _, - ]: Async: ToFuture: Parallel: LogOf: RandomIdOf: FromGFuture: MeasureDuration: ToTry: FromTry: FromJsResult: Fail: JsonCodec, - A, + F[ + _, + ]: Async: ToFuture: Parallel: LogOf: RandomIdOf: FromGFuture: MeasureDuration: ToTry: FromTry: FromJsResult: Fail: JsonCodec, + A, ]( - toKey: ToKey[F], - origin: Option[Origin], - serializer: EventSerializer[F, A], - journalReadWrite: JournalReadWrite[F, A], - config: KafkaJournalConfig, - metrics: Metrics[F], - log: Log[F], - batching: Batching[F], - appendMetadataOf: AppendMetadataOf[F], - cassandraClusterOf: CassandraClusterOf[F], + toKey: ToKey[F], + origin: Option[Origin], + serializer: EventSerializer[F, A], + journalReadWrite: JournalReadWrite[F, A], + config: KafkaJournalConfig, + metrics: Metrics[F], + log: Log[F], + batching: Batching[F], + appendMetadataOf: AppendMetadataOf[F], + cassandraClusterOf: CassandraClusterOf[F], ): Resource[F, JournalAdapter[F]] = { def clientIdOf(config: CommonConfig) = config.clientId getOrElse "journal" @@ -76,9 +76,9 @@ object JournalAdapter { } def journal(eventualJournal: EventualJournal[F])( - implicit kafkaConsumerOf: KafkaConsumerOf[F], - kafkaProducerOf: KafkaProducerOf[F], - headCacheOf: HeadCacheOf[F], + implicit kafkaConsumerOf: KafkaConsumerOf[F], + kafkaProducerOf: KafkaProducerOf[F], + headCacheOf: HeadCacheOf[F], ): Resource[F, Journals[F]] = { Journals .make[F]( @@ -111,11 +111,11 @@ object JournalAdapter { } def apply[F[_]: Monad, A]( - journals: Journals[F], - toKey: ToKey[F], - serializer: EventSerializer[F, A], - journalReadWrite: JournalReadWrite[F, A], - appendMetadataOf: AppendMetadataOf[F], + journals: Journals[F], + toKey: ToKey[F], + serializer: EventSerializer[F, A], + journalReadWrite: JournalReadWrite[F, A], + appendMetadataOf: AppendMetadataOf[F], ): JournalAdapter[F] = { implicit val kafkaRead = journalReadWrite.kafkaRead @@ -233,13 +233,13 @@ object JournalAdapter { } final case class Metrics[F[_]]( - journal: Option[JournalMetrics[F]] = none, - eventual: Option[EventualJournal.Metrics[F]] = none, - headCache: Option[HeadCacheMetrics[F]] = none, - producer: Option[ClientId => ProducerMetrics[F]] = none, - consumer: Option[ClientId => ConsumerMetrics[F]] = none, - conversion: Option[ConversionMetrics[F]] = none, - consumerPool: Option[ConsumerPoolMetrics[F]] = none, + journal: Option[JournalMetrics[F]] = none, + eventual: Option[EventualJournal.Metrics[F]] = none, + headCache: Option[HeadCacheMetrics[F]] = none, + producer: Option[ClientId => ProducerMetrics[F]] = none, + consumer: Option[ClientId => ConsumerMetrics[F]] = none, + conversion: Option[ConversionMetrics[F]] = none, + consumerPool: Option[ConsumerPoolMetrics[F]] = none, ) object Metrics { diff --git a/persistence/src/main/scala/akka/persistence/kafka/journal/KafkaJournal.scala b/persistence/src/main/scala/akka/persistence/kafka/journal/KafkaJournal.scala index 1f3ce33bc..e7a4fcc19 100644 --- a/persistence/src/main/scala/akka/persistence/kafka/journal/KafkaJournal.scala +++ b/persistence/src/main/scala/akka/persistence/kafka/journal/KafkaJournal.scala @@ -185,8 +185,8 @@ class KafkaJournal(config: Config) extends AsyncWriteJournal { actor => } def adapterIO[A]( - serializer: EventSerializer[IO, A], - journalReadWrite: JournalReadWrite[IO, A], + serializer: EventSerializer[IO, A], + journalReadWrite: JournalReadWrite[IO, A], ): Resource[IO, JournalAdapter[IO]] = { for { config <- kafkaJournalConfig.toResource @@ -195,9 +195,9 @@ class KafkaJournal(config: Config) extends AsyncWriteJournal { actor => } def adapterIO[A]( - config: KafkaJournalConfig, - serializer: EventSerializer[IO, A], - journalReadWrite: JournalReadWrite[IO, A], + config: KafkaJournalConfig, + serializer: EventSerializer[IO, A], + journalReadWrite: JournalReadWrite[IO, A], ): Resource[IO, JournalAdapter[IO]] = { for { logOf <- logOf @@ -259,21 +259,21 @@ class KafkaJournal(config: Config) extends AsyncWriteJournal { actor => } def adapterOf[A]( - toKey: ToKey[IO], - origin: Option[Origin], - serializer: EventSerializer[IO, A], - journalReadWrite: JournalReadWrite[IO, A], - config: KafkaJournalConfig, - metrics: JournalAdapter.Metrics[IO], - appendMetadataOf: AppendMetadataOf[IO], - batching: Batching[IO], - log: Log[IO], - cassandraClusterOf: CassandraClusterOf[IO], + toKey: ToKey[IO], + origin: Option[Origin], + serializer: EventSerializer[IO, A], + journalReadWrite: JournalReadWrite[IO, A], + config: KafkaJournalConfig, + metrics: JournalAdapter.Metrics[IO], + appendMetadataOf: AppendMetadataOf[IO], + batching: Batching[IO], + log: Log[IO], + cassandraClusterOf: CassandraClusterOf[IO], )( - implicit logOf: LogOf[IO], - randomIdOf: RandomIdOf[IO], - measureDuration: MeasureDuration[IO], - jsonCodec: JsonCodec[IO], + implicit logOf: LogOf[IO], + randomIdOf: RandomIdOf[IO], + measureDuration: MeasureDuration[IO], + jsonCodec: JsonCodec[IO], ): Resource[IO, JournalAdapter[IO]] = { JournalAdapter.make[IO, A]( @@ -308,7 +308,7 @@ class KafkaJournal(config: Config) extends AsyncWriteJournal { actor => } def asyncReplayMessages(persistenceId: PersistenceId, from: Long, to: Long, max: Long)( - f: PersistentRepr => Unit, + f: PersistentRepr => Unit, ): Future[Unit] = { val seqNrFrom = SeqNr .of[Option](from) diff --git a/persistence/src/main/scala/akka/persistence/kafka/journal/KafkaJournalConfig.scala b/persistence/src/main/scala/akka/persistence/kafka/journal/KafkaJournalConfig.scala index d78ae3199..b106cb611 100644 --- a/persistence/src/main/scala/akka/persistence/kafka/journal/KafkaJournalConfig.scala +++ b/persistence/src/main/scala/akka/persistence/kafka/journal/KafkaJournalConfig.scala @@ -52,15 +52,15 @@ import scala.concurrent.duration.* * @see [[KafkaJournal]] for more details. */ final case class KafkaJournalConfig( - journal: JournalConfig = JournalConfig.default, - cassandra: EventualCassandraConfig = EventualCassandraConfig.default, - startTimeout: FiniteDuration = 1.minute, - stopTimeout: FiniteDuration = 1.minute, - maxEventsInBatch: Int = 100, - callTimeThresholds: CallTimeThresholds = CallTimeThresholds.default, - jsonCodec: KafkaJournalConfig.JsonCodec = KafkaJournalConfig.JsonCodec.Default, - consumerPool: ConsumerPoolConfig = ConsumerPoolConfig.Default, - dataIntegrity: DataIntegrityConfig = DataIntegrityConfig.Default, + journal: JournalConfig = JournalConfig.default, + cassandra: EventualCassandraConfig = EventualCassandraConfig.default, + startTimeout: FiniteDuration = 1.minute, + stopTimeout: FiniteDuration = 1.minute, + maxEventsInBatch: Int = 100, + callTimeThresholds: CallTimeThresholds = CallTimeThresholds.default, + jsonCodec: KafkaJournalConfig.JsonCodec = KafkaJournalConfig.JsonCodec.Default, + consumerPool: ConsumerPoolConfig = ConsumerPoolConfig.Default, + dataIntegrity: DataIntegrityConfig = DataIntegrityConfig.Default, ) object KafkaJournalConfig { diff --git a/persistence/src/main/scala/akka/persistence/kafka/journal/PersistentJson.scala b/persistence/src/main/scala/akka/persistence/kafka/journal/PersistentJson.scala index 3c9cd9b6c..761849106 100644 --- a/persistence/src/main/scala/akka/persistence/kafka/journal/PersistentJson.scala +++ b/persistence/src/main/scala/akka/persistence/kafka/journal/PersistentJson.scala @@ -4,10 +4,10 @@ import com.evolutiongaming.kafka.journal.PayloadType import play.api.libs.json.{Format, Json, OFormat} final case class PersistentJson[A]( - manifest: Option[String], - writerUuid: String, - payloadType: Option[PayloadType.TextOrJson], - payload: A, + manifest: Option[String], + writerUuid: String, + payloadType: Option[PayloadType.TextOrJson], + payload: A, ) object PersistentJson { diff --git a/persistence/src/test/scala/akka/persistence/kafka/journal/JournalAdapterSpec.scala b/persistence/src/test/scala/akka/persistence/kafka/journal/JournalAdapterSpec.scala index fcbf71159..5fb631e2d 100644 --- a/persistence/src/test/scala/akka/persistence/kafka/journal/JournalAdapterSpec.scala +++ b/persistence/src/test/scala/akka/persistence/kafka/journal/JournalAdapterSpec.scala @@ -120,13 +120,13 @@ object JournalAdapterSpec { final case class Pointer(key: Key) final case class State( - events: List[EventRecord[EventualPayloadAndType]] = List.empty, - appends: List[Append] = List.empty, - pointers: List[Pointer] = List.empty, - deletes: List[Delete] = List.empty, - reads: List[Read] = List.empty, - replayed: List[PersistentRepr] = List.empty, - actions: List[Action] = List.empty, + events: List[EventRecord[EventualPayloadAndType]] = List.empty, + appends: List[Append] = List.empty, + pointers: List[Pointer] = List.empty, + deletes: List[Delete] = List.empty, + reads: List[Read] = List.empty, + replayed: List[PersistentRepr] = List.empty, + actions: List[Action] = List.empty, ) object State { @@ -161,7 +161,7 @@ object JournalAdapterSpec { def apply(key: Key) = new Journal[StateT] { def append[A](events: Nel[Event[A]], metadata: RecordMetadata, headers: Headers)( - implicit kafkaWrite: KafkaWrite[StateT, A], + implicit kafkaWrite: KafkaWrite[StateT, A], ) = for { payloadAndType <- kafkaWrite(Events(events, metadata.payload)) diff --git a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/Batch.scala b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/Batch.scala index b50393b1e..26b676863 100644 --- a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/Batch.scala +++ b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/Batch.scala @@ -153,20 +153,20 @@ private[journal] object Batch { } final case class Appends( - offset: Offset, - records: NonEmptyList[ActionRecord[Action.Append]], + offset: Offset, + records: NonEmptyList[ActionRecord[Action.Append]], ) extends Batch final case class Delete( - offset: Offset, - to: DeleteTo, - origin: Option[Origin], - version: Option[Version], + offset: Offset, + to: DeleteTo, + origin: Option[Origin], + version: Option[Version], ) extends Batch final case class Purge( - offset: Offset, - origin: Option[Origin], // used only for logging - version: Option[Version], // used only for logging + offset: Offset, + origin: Option[Origin], // used only for logging + version: Option[Version], // used only for logging ) extends Batch } diff --git a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/CacheOf.scala b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/CacheOf.scala index 170e3fa29..fbdb7440a 100644 --- a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/CacheOf.scala +++ b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/CacheOf.scala @@ -40,8 +40,8 @@ private[journal] object CacheOf { } def apply[F[_]: Temporal: Runtime: Parallel: MeasureDuration]( - expireAfter: FiniteDuration, - cacheMetrics: Option[CacheMetrics.Name => CacheMetrics[F]], + expireAfter: FiniteDuration, + cacheMetrics: Option[CacheMetrics.Name => CacheMetrics[F]], ): CacheOf[F] = { class Main new Main with CacheOf[F] { diff --git a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/ConsumeTopic.scala b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/ConsumeTopic.scala index d9fe6fab9..cb75b95e0 100644 --- a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/ConsumeTopic.scala +++ b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/ConsumeTopic.scala @@ -16,10 +16,10 @@ import scala.concurrent.duration.* private[journal] object ConsumeTopic { def apply[F[_]: BracketThrowable: Sleep]( - topic: Topic, - consumer: Resource[F, TopicConsumer[F]], - topicFlowOf: TopicFlowOf[F], - log: Log[F], + topic: Topic, + consumer: Resource[F, TopicConsumer[F]], + topicFlowOf: TopicFlowOf[F], + log: Log[F], ): F[Unit] = { for { random <- Random.State.fromClock[F]() @@ -35,11 +35,11 @@ private[journal] object ConsumeTopic { } def apply[F[_]: BracketThrowable]( - topic: Topic, - consumer: Resource[F, TopicConsumer[F]], - topicFlowOf: TopicFlowOf[F], - log: Log[F], - retry: Retry[F], + topic: Topic, + consumer: Resource[F, TopicConsumer[F]], + topicFlowOf: TopicFlowOf[F], + log: Log[F], + retry: Retry[F], ): F[Unit] = { def rebalanceListenerOf(topicFlow: TopicFlow[F]): RebalanceListener1[F] = { diff --git a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/DistributeJob.scala b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/DistributeJob.scala index 54af644a8..2168dd7db 100644 --- a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/DistributeJob.scala +++ b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/DistributeJob.scala @@ -70,10 +70,10 @@ object DistributeJob { type Assigned = Boolean def apply[F[_]: Concurrent: Defer: Sleep: FromTry: LogOf: Parallel]( - groupId: String, - topic: Topic, - consumerConfig: ConsumerConfig, - kafkaConsumerOf: KafkaConsumerOf[F], + groupId: String, + topic: Topic, + consumerConfig: ConsumerConfig, + kafkaConsumerOf: KafkaConsumerOf[F], ): Resource[F, DistributeJob[F]] = { val consumerConfig1 = consumerConfig.copy(autoCommit = true, groupId = groupId.some, autoOffsetReset = AutoOffsetReset.Latest) @@ -143,9 +143,9 @@ object DistributeJob { consumer <- kafkaConsumerOf[String, Unit](consumerConfig1) partitionsAll <- consumer.partitions(topic).toResource active = ( - deferred: Deferred[F, String => F[Unit]], - jobs: Map[String, (Job, Release)], - partitions: Map[Partition, Assigned], + deferred: Deferred[F, String => F[Unit]], + jobs: Map[String, (Job, Release)], + partitions: Map[Partition, Assigned], ) => { val jobs1 = jobs.map { case (name, (job, _)) => @@ -343,8 +343,8 @@ object DistributeJob { } final case class DistributeJobError( - msg: String, - cause: Option[Throwable] = None, + msg: String, + cause: Option[Throwable] = None, ) extends RuntimeException(msg, cause.orNull) object DistributeJobError { diff --git a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/PurgeExpired.scala b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/PurgeExpired.scala index dd796b38d..6690776e3 100644 --- a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/PurgeExpired.scala +++ b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/PurgeExpired.scala @@ -28,11 +28,11 @@ trait PurgeExpired[F[_]] { object PurgeExpired { def make[F[_]: MonadThrowable: KafkaProducerOf: CassandraSession: FromTry: Fail: Clock: MeasureDuration: JsonCodec.Encode]( - origin: Option[Origin], - producerConfig: ProducerConfig, - tableName: TableName, - metrics: Option[Metrics[F]], - consistencyConfig: CassandraConsistencyConfig.Read, + origin: Option[Origin], + producerConfig: ProducerConfig, + tableName: TableName, + metrics: Option[Metrics[F]], + consistencyConfig: CassandraConsistencyConfig.Read, ): Resource[F, PurgeExpired[F]] = { implicit val fromAttempt = FromAttempt.lift[F] @@ -49,8 +49,8 @@ object PurgeExpired { } def apply[F[_]: Monad]( - selectExpired: MetaJournalStatements.IdByTopicAndExpireOn[F], - produce: Produce[F], + selectExpired: MetaJournalStatements.IdByTopicAndExpireOn[F], + produce: Produce[F], ): PurgeExpired[F] = { new PurgeExpired[F] { @@ -129,8 +129,8 @@ object PurgeExpired { def const[F[_]](unit: F[Unit]): Metrics[F] = (_: Topic, _: FiniteDuration, _: Long) => unit def make[F[_]: Monad]( - registry: CollectorRegistry[F], - prefix: String = "purge_expired", + registry: CollectorRegistry[F], + prefix: String = "purge_expired", ): Resource[F, Metrics[F]] = { val latencySummary = registry.summary( diff --git a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/ReplicateRecords.scala b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/ReplicateRecords.scala index bbb90262c..4595577d5 100644 --- a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/ReplicateRecords.scala +++ b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/ReplicateRecords.scala @@ -26,12 +26,12 @@ private[journal] trait ReplicateRecords[F[_]] { private[journal] object ReplicateRecords { def apply[F[_]: BracketThrowable: Clock, A]( - consRecordToActionRecord: ConsRecordToActionRecord[F], - journal: ReplicatedKeyJournal[F], - metrics: TopicReplicatorMetrics[F], - kafkaRead: KafkaRead[F, A], - eventualWrite: EventualWrite[F, A], - log: Log[F], + consRecordToActionRecord: ConsRecordToActionRecord[F], + journal: ReplicatedKeyJournal[F], + metrics: TopicReplicatorMetrics[F], + kafkaRead: KafkaRead[F, A], + eventualWrite: EventualWrite[F, A], + log: Log[F], ): ReplicateRecords[F] = { (records: Nel[ConsRecord], timestamp: Instant) => { @@ -58,9 +58,9 @@ private[journal] object ReplicateRecords { val bytes = records.foldLeft(0L) { case (bytes, record) => bytes + record.action.payload.size } def msg( - events: Nel[EventRecord[EventualPayloadAndType]], - latency: FiniteDuration, - expireAfter: Option[ExpireAfter], + events: Nel[EventRecord[EventualPayloadAndType]], + latency: FiniteDuration, + expireAfter: Option[ExpireAfter], ): String = { val seqNrs = if (events.tail.isEmpty) s"seqNr: ${events.head.seqNr}" diff --git a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/Replicator.scala b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/Replicator.scala index daab82d55..5589e9b88 100644 --- a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/Replicator.scala +++ b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/Replicator.scala @@ -41,10 +41,10 @@ trait Replicator[F[_]] { object Replicator { def make[F[_]: Async: Parallel: FromTry: ToTry: Fail: LogOf: KafkaConsumerOf: FromGFuture: MeasureDuration: JsonCodec]( - config: ReplicatorConfig, - cassandraClusterOf: CassandraClusterOf[F], - hostName: Option[HostName], - metrics: Option[Metrics[F]] = none, + config: ReplicatorConfig, + cassandraClusterOf: CassandraClusterOf[F], + hostName: Option[HostName], + metrics: Option[Metrics[F]] = none, ): Resource[F, F[Unit]] = { def replicatedJournal(implicit cassandraCluster: CassandraCluster[F], cassandraSession: CassandraSession[F]) = { @@ -61,12 +61,12 @@ object Replicator { } def make[ - F[_]: Temporal: Parallel: Runtime: FromTry: ToTry: Fail: LogOf: KafkaConsumerOf: MeasureDuration: JsonCodec, + F[_]: Temporal: Parallel: Runtime: FromTry: ToTry: Fail: LogOf: KafkaConsumerOf: MeasureDuration: JsonCodec, ]( - config: ReplicatorConfig, - metrics: Option[Metrics[F]], - journal: ReplicatedJournal[F], - hostName: Option[HostName], + config: ReplicatorConfig, + metrics: Option[Metrics[F]], + journal: ReplicatedJournal[F], + hostName: Option[HostName], ): Resource[F, F[Unit]] = { val topicReplicator: Topic => Resource[F, F[Outcome[F, Throwable, Unit]]] = @@ -87,9 +87,9 @@ object Replicator { } def make[F[_]: Concurrent: Sleep: Parallel: LogOf: MeasureDuration]( - config: Config, - consumer: Resource[F, Consumer[F]], - topicReplicatorOf: Topic => Resource[F, F[Outcome[F, Throwable, Unit]]], + config: Config, + consumer: Resource[F, Consumer[F]], + topicReplicatorOf: Topic => Resource[F, F[Outcome[F, Throwable, Unit]]], ): Resource[F, F[Unit]] = { def retry(log: Log[F]) = for { @@ -145,11 +145,11 @@ object Replicator { } def start[F[_]: Concurrent: Sleep: Parallel: MeasureDuration]( - config: Config, - consumer: Consumer[F], - start: Topic => F[Unit], - continue: F[Unit], - log: Log[F], + config: Config, + consumer: Consumer[F], + start: Topic => F[Unit], + continue: F[Unit], + log: Log[F], ): F[Unit] = { type State = Set[Topic] diff --git a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/ReplicatorConfig.scala b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/ReplicatorConfig.scala index b9d2c6012..472126fdf 100644 --- a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/ReplicatorConfig.scala +++ b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/ReplicatorConfig.scala @@ -13,17 +13,17 @@ import pureconfig.{ConfigCursor, ConfigReader, ConfigSource} import scala.concurrent.duration.* final case class ReplicatorConfig( - topicPrefixes: Nel[String] = Nel.of("journal"), - topicDiscoveryInterval: FiniteDuration = 3.seconds, - cacheExpireAfter: FiniteDuration = 1.minute, - kafka: KafkaConfig = KafkaConfig("replicator"), - cassandra: EventualCassandraConfig = EventualCassandraConfig( - client = CassandraConfig( - name = "replicator", - query = QueryConfig(consistency = ConsistencyLevel.LOCAL_QUORUM, defaultIdempotence = true), + topicPrefixes: Nel[String] = Nel.of("journal"), + topicDiscoveryInterval: FiniteDuration = 3.seconds, + cacheExpireAfter: FiniteDuration = 1.minute, + kafka: KafkaConfig = KafkaConfig("replicator"), + cassandra: EventualCassandraConfig = EventualCassandraConfig( + client = CassandraConfig( + name = "replicator", + query = QueryConfig(consistency = ConsistencyLevel.LOCAL_QUORUM, defaultIdempotence = true), + ), ), - ), - pollTimeout: FiniteDuration = 10.millis, + pollTimeout: FiniteDuration = 10.millis, ) object ReplicatorConfig { diff --git a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicCommit.scala b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicCommit.scala index 3bcaf5171..a14710b5a 100644 --- a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicCommit.scala +++ b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicCommit.scala @@ -25,9 +25,9 @@ object TopicCommit { def empty[F[_]: Applicative]: TopicCommit[F] = (_: Nem[Partition, Offset]) => ().pure[F] def apply[F[_]]( - topic: Topic, - metadata: String, - consumer: KafkaConsumer[F, _, _], + topic: Topic, + metadata: String, + consumer: KafkaConsumer[F, _, _], ): TopicCommit[F] = { (offsets: Nem[Partition, Offset]) => { val offsets1 = offsets.mapKV { (partition, offset) => @@ -40,8 +40,8 @@ object TopicCommit { } def delayed[F[_]: Concurrent: Clock]( - delay: FiniteDuration, - commit: TopicCommit[F], + delay: FiniteDuration, + commit: TopicCommit[F], ): F[TopicCommit[F]] = { case class State(until: Instant, offsets: SortedMap[Partition, Offset] = SortedMap.empty) diff --git a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicConsumer.scala b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicConsumer.scala index 463b5c393..57625b7a0 100644 --- a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicConsumer.scala +++ b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicConsumer.scala @@ -23,10 +23,10 @@ private[journal] trait TopicConsumer[F[_]] { private[journal] object TopicConsumer { def apply[F[_]: Monad]( - topic: Topic, - pollTimeout: FiniteDuration, - commit: TopicCommit[F], - consumer: KafkaConsumer[F, String, ByteVector], + topic: Topic, + pollTimeout: FiniteDuration, + commit: TopicCommit[F], + consumer: KafkaConsumer[F, String, ByteVector], ): TopicConsumer[F] = { val commit1 = commit diff --git a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicReplicator.scala b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicReplicator.scala index 07e9a0871..3ef1930b5 100644 --- a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicReplicator.scala +++ b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicReplicator.scala @@ -31,11 +31,11 @@ import scala.util.Try private[journal] object TopicReplicator { def make[F[_]: Concurrent: Sleep: ToTry: LogOf: Fail: MeasureDuration: JsonCodec]( - topic: Topic, - journal: ReplicatedJournal[F], - consumer: Resource[F, TopicConsumer[F]], - metrics: TopicReplicatorMetrics[F], - cacheOf: CacheOf[F], + topic: Topic, + journal: ReplicatedJournal[F], + consumer: Resource[F, TopicConsumer[F]], + metrics: TopicReplicatorMetrics[F], + cacheOf: CacheOf[F], ): Resource[F, F[Outcome[F, Throwable, Unit]]] = { implicit val fromAttempt: FromAttempt[F] = FromAttempt.lift[F] @@ -46,8 +46,8 @@ private[journal] object TopicReplicator { val eventualWrite = EventualWrite.summon[F, Payload] def consume( - consumer: Resource[F, TopicConsumer[F]], - log: Log[F], + consumer: Resource[F, TopicConsumer[F]], + log: Log[F], ) = { val consRecordToActionRecord = ConsRecordToActionRecord[F] @@ -77,15 +77,15 @@ private[journal] object TopicReplicator { } def of[F[_]: Concurrent: Sleep: MeasureDuration, A]( - topic: Topic, - consumer: Resource[F, TopicConsumer[F]], - consRecordToActionRecord: ConsRecordToActionRecord[F], - kafkaRead: KafkaRead[F, A], - eventualWrite: EventualWrite[F, A], - journal: ReplicatedJournal[F], - metrics: TopicReplicatorMetrics[F], - log: Log[F], - cacheOf: CacheOf[F], + topic: Topic, + consumer: Resource[F, TopicConsumer[F]], + consRecordToActionRecord: ConsRecordToActionRecord[F], + kafkaRead: KafkaRead[F, A], + eventualWrite: EventualWrite[F, A], + journal: ReplicatedJournal[F], + metrics: TopicReplicatorMetrics[F], + log: Log[F], + cacheOf: CacheOf[F], ): F[Unit] = { trait PartitionFlow { @@ -236,10 +236,10 @@ private[journal] object TopicReplicator { object ConsumerOf { def make[F[_]: Concurrent: KafkaConsumerOf: FromTry: Clock]( - topic: Topic, - config: ConsumerConfig, - pollTimeout: FiniteDuration, - hostName: Option[HostName], + topic: Topic, + config: ConsumerConfig, + pollTimeout: FiniteDuration, + hostName: Option[HostName], ): Resource[F, TopicConsumer[F]] = { val groupId = { diff --git a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicReplicatorMetrics.scala b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicReplicatorMetrics.scala index 53734e09b..a6a61852d 100644 --- a/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicReplicatorMetrics.scala +++ b/replicator/src/main/scala/com/evolutiongaming/kafka/journal/replicator/TopicReplicatorMetrics.scala @@ -47,8 +47,8 @@ object TopicReplicatorMetrics { } def make[F[_]: Monad]( - registry: CollectorRegistry[F], - prefix: String = "replicator", + registry: CollectorRegistry[F], + prefix: String = "replicator", ): Resource[F, Topic => TopicReplicatorMetrics[F]] = { val replicationSummary = registry.summary( diff --git a/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/ConsumeTopicTest.scala b/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/ConsumeTopicTest.scala index fc053d7cd..a4a524fa8 100644 --- a/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/ConsumeTopicTest.scala +++ b/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/ConsumeTopicTest.scala @@ -126,8 +126,8 @@ object ConsumeTopicTest { val partitions: Nes[Partition] = Nes.of(Partition.min) final case class State( - commands: List[Command] = List.empty, - actions: List[Action] = List.empty, + commands: List[Command] = List.empty, + actions: List[Action] = List.empty, ) { def +(action: Action): State = copy(actions = action :: actions) @@ -317,8 +317,8 @@ object ConsumeTopicTest { } def recordOf( - partition: Int, - offset: Long, + partition: Int, + offset: Long, ): ConsRecord = { ConsRecord( topicPartition = TopicPartition(topic = topic, partition = Partition.unsafe(partition)), diff --git a/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/DistributeJobTest.scala b/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/DistributeJobTest.scala index 1e928c6ee..d21295ca7 100644 --- a/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/DistributeJobTest.scala +++ b/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/DistributeJobTest.scala @@ -34,7 +34,7 @@ class DistributeJobTest extends AsyncFunSuite with Matchers { deferred <- Deferred[IO, RebalanceListener1[IO]] kafkaConsumerOf = new KafkaConsumerOf[IO] { def apply[K, V]( - config: ConsumerConfig, + config: ConsumerConfig, )(implicit fromBytesK: skafka.FromBytes[IO, K], fromBytesV: skafka.FromBytes[IO, V]) = { val consumer: KafkaConsumer[IO, K, V] = new KafkaConsumer[IO, K, V] { def assign(partitions: Nes[TopicPartition]) = ().pure[IO] diff --git a/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/EmptyRebalanceConsumer.scala b/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/EmptyRebalanceConsumer.scala index db130bca5..0895b12fe 100644 --- a/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/EmptyRebalanceConsumer.scala +++ b/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/EmptyRebalanceConsumer.scala @@ -16,8 +16,8 @@ object EmptyRebalanceConsumer extends RebalanceConsumer { ) override def beginningOffsets( - partitions: NonEmptySet[TopicPartition], - timeout: FiniteDuration, + partitions: NonEmptySet[TopicPartition], + timeout: FiniteDuration, ): Try[Map[TopicPartition, Offset]] = Failure(new NotImplementedError) override def commit(): Try[Unit] = Failure(new NotImplementedError) @@ -35,8 +35,8 @@ object EmptyRebalanceConsumer extends RebalanceConsumer { ) override def committed( - partitions: NonEmptySet[TopicPartition], - timeout: FiniteDuration, + partitions: NonEmptySet[TopicPartition], + timeout: FiniteDuration, ): Try[Map[TopicPartition, OffsetAndMetadata]] = Failure(new NotImplementedError) override def endOffsets(partitions: NonEmptySet[TopicPartition]): Try[Map[TopicPartition, Offset]] = Failure( @@ -53,12 +53,12 @@ object EmptyRebalanceConsumer extends RebalanceConsumer { override def topics(timeout: FiniteDuration): Try[Map[Topic, List[PartitionInfo]]] = Failure(new NotImplementedError) override def offsetsForTimes( - timestampsToSearch: NonEmptyMap[TopicPartition, Instant], + timestampsToSearch: NonEmptyMap[TopicPartition, Instant], ): Try[Map[TopicPartition, Option[OffsetAndTimestamp]]] = Failure(new NotImplementedError) override def offsetsForTimes( - timestampsToSearch: NonEmptyMap[TopicPartition, Instant], - timeout: FiniteDuration, + timestampsToSearch: NonEmptyMap[TopicPartition, Instant], + timeout: FiniteDuration, ): Try[Map[TopicPartition, Option[OffsetAndTimestamp]]] = Failure(new NotImplementedError) override def partitionsFor(topic: Topic): Try[List[PartitionInfo]] = Failure(new NotImplementedError) diff --git a/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/TopicReplicatorSpec.scala b/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/TopicReplicatorSpec.scala index 6ad4a4e1c..e70c898f6 100644 --- a/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/TopicReplicatorSpec.scala +++ b/replicator/src/test/scala/com/evolutiongaming/kafka/journal/replicator/TopicReplicatorSpec.scala @@ -696,9 +696,9 @@ class TopicReplicatorSpec extends AsyncWordSpec with Matchers { } private def consumerRecordOf( - action: Action, - topicPartition: TopicPartition, - offset: Long, + action: Action, + topicPartition: TopicPartition, + offset: Long, ) = { val producerRecord = ActionToProducerRecord[Try].apply(action).get @@ -713,10 +713,10 @@ class TopicReplicatorSpec extends AsyncWordSpec with Matchers { } private def record( - seqNr: Int, - partition: Int, - offset: Long, - expireAfter: Option[ExpireAfter] = none, + seqNr: Int, + partition: Int, + offset: Long, + expireAfter: Option[ExpireAfter] = none, ) = { val partitionOffset = PartitionOffset(Partition.unsafe(partition), Offset.unsafe(offset)) val event = Event[EventualPayloadAndType](SeqNr.unsafe(seqNr), Set(seqNr.toString)) @@ -786,11 +786,11 @@ object TopicReplicatorSpec { def keyOf(id: String) = Key(id = id, topic = topic) def metaJournalOf( - id: String, - partition: Int, - offset: Long, - deleteTo: Option[Int] = none, - expireAfter: Option[ExpireAfter] = none, + id: String, + partition: Int, + offset: Long, + deleteTo: Option[Int] = none, + expireAfter: Option[ExpireAfter] = none, ): (String, MetaJournal) = { val deleteToSeqNr = deleteTo .flatMap { deleteTo => SeqNr.opt(deleteTo.toLong) } @@ -860,10 +860,10 @@ object TopicReplicatorSpec { val result = new ReplicatedKeyJournal[StateT] { def append( - offset: Offset, - timestamp: Instant, - expireAfter: Option[ExpireAfter], - events: Nel[EventRecord[EventualPayloadAndType]], + offset: Offset, + timestamp: Instant, + expireAfter: Option[ExpireAfter], + events: Nel[EventRecord[EventualPayloadAndType]], ) = { StateT { state => val records = events.toList ++ state.journal.getOrElse(id, Nil) @@ -1016,13 +1016,13 @@ object TopicReplicatorSpec { } final case class State( - topics: List[Topic] = Nil, - commits: List[Nem[Int, Long]] = Nil, - records: List[ConsRecords] = Nil, - pointers: Map[Topic, Map[Int, Long]] = Map.empty, - journal: Map[String, List[EventRecord[EventualPayloadAndType]]] = Map.empty, - metaJournal: Map[String, MetaJournal] = Map.empty, - metrics: List[Metrics] = Nil, + topics: List[Topic] = Nil, + commits: List[Nem[Int, Long]] = Nil, + records: List[ConsRecords] = Nil, + pointers: Map[Topic, Map[Int, Long]] = Map.empty, + journal: Map[String, List[EventRecord[EventualPayloadAndType]]] = Map.empty, + metaJournal: Map[String, MetaJournal] = Map.empty, + metrics: List[Metrics] = Nil, ) { self => def +(metrics: Metrics): (State, Unit) = { @@ -1072,10 +1072,10 @@ object TopicReplicatorSpec { } final case class MetaJournal( - offset: PartitionOffset, - deleteTo: Option[DeleteTo], - expireAfter: Option[ExpireAfter], - origin: Option[Origin], + offset: PartitionOffset, + deleteTo: Option[DeleteTo], + expireAfter: Option[ExpireAfter], + origin: Option[Origin], ) case object NotImplemented extends RuntimeException with NoStackTrace diff --git a/snapshot/src/main/scala/com/evolutiongaming/kafka/journal/Snapshot.scala b/snapshot/src/main/scala/com/evolutiongaming/kafka/journal/Snapshot.scala index 8b25b6906..3dc762403 100644 --- a/snapshot/src/main/scala/com/evolutiongaming/kafka/journal/Snapshot.scala +++ b/snapshot/src/main/scala/com/evolutiongaming/kafka/journal/Snapshot.scala @@ -8,6 +8,6 @@ package com.evolutiongaming.kafka.journal * Actual contents of a snapshot. */ final case class Snapshot[A]( - seqNr: SeqNr, - payload: A, + seqNr: SeqNr, + payload: A, ) diff --git a/snapshot/src/main/scala/com/evolutiongaming/kafka/journal/SnapshotRecord.scala b/snapshot/src/main/scala/com/evolutiongaming/kafka/journal/SnapshotRecord.scala index 19dd17237..431ec8aa5 100644 --- a/snapshot/src/main/scala/com/evolutiongaming/kafka/journal/SnapshotRecord.scala +++ b/snapshot/src/main/scala/com/evolutiongaming/kafka/journal/SnapshotRecord.scala @@ -20,8 +20,8 @@ import java.time.Instant * The version of a library, which produced a snapshot. See [[Version]] for more details. */ final case class SnapshotRecord[A]( - snapshot: Snapshot[A], - timestamp: Instant, - origin: Option[Origin], - version: Option[Version], + snapshot: Snapshot[A], + timestamp: Instant, + origin: Option[Origin], + version: Option[Version], ) diff --git a/snapshot/src/main/scala/com/evolutiongaming/kafka/journal/SnapshotSelectionCriteria.scala b/snapshot/src/main/scala/com/evolutiongaming/kafka/journal/SnapshotSelectionCriteria.scala index bd71e2587..52e71e917 100644 --- a/snapshot/src/main/scala/com/evolutiongaming/kafka/journal/SnapshotSelectionCriteria.scala +++ b/snapshot/src/main/scala/com/evolutiongaming/kafka/journal/SnapshotSelectionCriteria.scala @@ -10,10 +10,10 @@ import java.time.Instant * they are given in a reverse order in a constructor. */ final case class SnapshotSelectionCriteria( - maxSeqNr: SeqNr = SeqNr.max, - maxTimestamp: Instant = Instant.MAX, - minSeqNr: SeqNr = SeqNr.min, - minTimestamp: Instant = Instant.MIN, + maxSeqNr: SeqNr = SeqNr.max, + maxTimestamp: Instant = Instant.MAX, + minSeqNr: SeqNr = SeqNr.min, + minTimestamp: Instant = Instant.MIN, ) object SnapshotSelectionCriteria { diff --git a/tests/src/test/scala/com/evolutiongaming/kafka/journal/AppendReplicateApp.scala b/tests/src/test/scala/com/evolutiongaming/kafka/journal/AppendReplicateApp.scala index 75d36d534..e2184e537 100644 --- a/tests/src/test/scala/com/evolutiongaming/kafka/journal/AppendReplicateApp.scala +++ b/tests/src/test/scala/com/evolutiongaming/kafka/journal/AppendReplicateApp.scala @@ -40,7 +40,7 @@ object AppendReplicateApp extends IOApp { } private def runF[ - F[_]: Async: Parallel: ToFuture: FromGFuture: MeasureDuration: FromAttempt: FromTry: ToTry: Fail, + F[_]: Async: Parallel: ToFuture: FromGFuture: MeasureDuration: FromAttempt: FromTry: ToTry: Fail, ](topic: Topic)(implicit system: ActorSystem): F[Unit] = { implicit val logOf = LogOfFromAkka[F](system) @@ -53,8 +53,8 @@ object AppendReplicateApp extends IOApp { .liftTo[F] def journal(config: JournalConfig, hostName: Option[HostName], log: Log[F])( - implicit kafkaConsumerOf: KafkaConsumerOf[F], - kafkaProducerOf: KafkaProducerOf[F], + implicit kafkaConsumerOf: KafkaConsumerOf[F], + kafkaProducerOf: KafkaProducerOf[F], ) = { for { @@ -99,7 +99,7 @@ object AppendReplicateApp extends IOApp { } private def append[F[_]: Concurrent: Sleep: Parallel](topic: Topic, journals: Journals[F])( - implicit kafkaWrite: KafkaWrite[F, Payload], + implicit kafkaWrite: KafkaWrite[F, Payload], ) = { def append(id: String) = { diff --git a/tests/src/test/scala/com/evolutiongaming/kafka/journal/IntegrationSuite.scala b/tests/src/test/scala/com/evolutiongaming/kafka/journal/IntegrationSuite.scala index 7fe4d8be1..36b8f5100 100644 --- a/tests/src/test/scala/com/evolutiongaming/kafka/journal/IntegrationSuite.scala +++ b/tests/src/test/scala/com/evolutiongaming/kafka/journal/IntegrationSuite.scala @@ -18,7 +18,7 @@ import com.typesafe.config.ConfigFactory object IntegrationSuite { def startF[F[_]: Async: ToFuture: LogOf: MeasureDuration: FromTry: ToTry: Fail]( - cassandraClusterOf: CassandraClusterOf[F], + cassandraClusterOf: CassandraClusterOf[F], ): Resource[F, Unit] = { def cassandra(log: Log[F]) = { diff --git a/tests/src/test/scala/com/evolutiongaming/kafka/journal/JournalSuite.scala b/tests/src/test/scala/com/evolutiongaming/kafka/journal/JournalSuite.scala index 47b3fa8b5..2f9ba83e9 100644 --- a/tests/src/test/scala/com/evolutiongaming/kafka/journal/JournalSuite.scala +++ b/tests/src/test/scala/com/evolutiongaming/kafka/journal/JournalSuite.scala @@ -84,9 +84,9 @@ object JournalSuite { trait JournalTest[F[_]] { def append[A]( - events: Nel[Event[A]], - metadata: RecordMetadata = RecordMetadata.empty, - headers: Headers = Headers.empty, + events: Nel[Event[A]], + metadata: RecordMetadata = RecordMetadata.empty, + headers: Headers = Headers.empty, )(implicit kafkaWrite: KafkaWrite[F, A]): F[PartitionOffset] def read[A](implicit kafkaRead: KafkaRead[F, A], eventualRead: EventualRead[F, A]): F[List[EventRecord[A]]] @@ -103,8 +103,8 @@ object JournalSuite { object JournalTest { def apply[F[_]: Monad]( - journal: Journal[F], - timestamp: Instant, + journal: Journal[F], + timestamp: Instant, ): JournalTest[F] = new JournalTest[F] { def append[A](events: Nel[Event[A]], metadata: RecordMetadata, headers: Headers)(implicit kafkaWrite: KafkaWrite[F, A]) = { diff --git a/tests/src/test/scala/com/evolutiongaming/kafka/journal/ReadEventsApp.scala b/tests/src/test/scala/com/evolutiongaming/kafka/journal/ReadEventsApp.scala index 5a4d53baf..ab51f49ec 100644 --- a/tests/src/test/scala/com/evolutiongaming/kafka/journal/ReadEventsApp.scala +++ b/tests/src/test/scala/com/evolutiongaming/kafka/journal/ReadEventsApp.scala @@ -45,9 +45,9 @@ object ReadEventsApp extends IOApp { } private def runF[F[ - _, + _, ]: Async: ToFuture: Parallel: LogOf: FromGFuture: MeasureDuration: FromTry: ToTry: FromAttempt: FromJsResult: Fail]( - log: Log[F], + log: Log[F], ): F[Unit] = { implicit val kafkaConsumerOf = KafkaConsumerOf[F]() diff --git a/tests/src/test/scala/com/evolutiongaming/kafka/journal/SettingsIntSpec.scala b/tests/src/test/scala/com/evolutiongaming/kafka/journal/SettingsIntSpec.scala index b0ea1f7bd..fc77894f7 100644 --- a/tests/src/test/scala/com/evolutiongaming/kafka/journal/SettingsIntSpec.scala +++ b/tests/src/test/scala/com/evolutiongaming/kafka/journal/SettingsIntSpec.scala @@ -31,8 +31,8 @@ class SettingsIntSpec extends AsyncWordSpec with BeforeAndAfterAll with Matchers } private def resources[F[_]: Async: LogOf: Parallel: FromFuture]( - origin: Option[Origin], - cassandraClusterOf: CassandraClusterOf[F], + origin: Option[Origin], + cassandraClusterOf: CassandraClusterOf[F], ) = { def settings(config: SchemaConfig)(implicit cassandraCluster: CassandraCluster[F], cassandraSession: CassandraSession[F]) = { diff --git a/tests/src/test/scala/com/evolutiongaming/kafka/journal/replicator/ReplicatorIntSpec.scala b/tests/src/test/scala/com/evolutiongaming/kafka/journal/replicator/ReplicatorIntSpec.scala index cd40cf782..8fcd630a7 100644 --- a/tests/src/test/scala/com/evolutiongaming/kafka/journal/replicator/ReplicatorIntSpec.scala +++ b/tests/src/test/scala/com/evolutiongaming/kafka/journal/replicator/ReplicatorIntSpec.scala @@ -45,7 +45,7 @@ class ReplicatorIntSpec extends AsyncWordSpec with BeforeAndAfterAll with Matche private implicit val randomIdOf: RandomIdOf[IO] = RandomIdOf.uuid[IO] private def resources[F[_]: Async: LogOf: Parallel: FromFuture: ToFuture: RandomIdOf: MeasureDuration: FromTry: ToTry: Fail]( - cassandraClusterOf: CassandraClusterOf[F], + cassandraClusterOf: CassandraClusterOf[F], ) = { def eventualJournal(conf: Config) = { @@ -61,8 +61,8 @@ class ReplicatorIntSpec extends AsyncWordSpec with BeforeAndAfterAll with Matche } def journal( - conf: Config, - eventualJournal: EventualJournal[F], + conf: Config, + eventualJournal: EventualJournal[F], ) = { val config = ConfigSource