Constant Field Values
Contents
com.rabbitmq.*
-
com.rabbitmq.client.AMQPModifier and TypeConstant FieldValue
public static final int403public static final int504public static final int503public static final int320public static final int311public static final int3public static final int206public static final int501public static final int2public static final int8public static final int1public static final int4096public static final int541public static final int402public static final int313public static final int312public static final int530public static final int404public static final int540public static final int406public static final int200public static final int506public static final int405public static final int502public static final int505 -
com.rabbitmq.client.AMQP.PROTOCOL
-
com.rabbitmq.client.ConnectionFactoryModifier and TypeConstant FieldValue
public static final int5671public static final int5672public static final int2047public static final int60000public static final int0public static final int10000public static final int60public static final String"localhost"public static final long5000Lpublic static final String"guest"public static final int10000public static final String"guest"public static final String"/"public static final int-1public static final int-1 -
com.rabbitmq.client.ConnectionFactoryConfiguratorModifier and TypeConstant FieldValue
public static final String"channel.rpc.timeout"public static final String"channel.should.check.rpc.response.type"public static final String"client.properties."public static final String"connection.channel.max"public static final String"connection.frame.max"public static final String"connection.heartbeat"public static final String"connection.recovery.enabled"public static final String"connection.recovery.interval"public static final String"connection.timeout"public static final String"rabbitmq."public static final String"handshake.timeout"public static final String"host"public static final String"nio.nb.io.threads"public static final String"nio.read.byte.buffer.size"public static final String"nio.write.byte.buffer.size"public static final String"nio.write.enqueuing.timeout.in.ms"public static final String"nio.write.queue.capacity"public static final String"password"public static final String"port"public static final String"shutdown.timeout"public static final String"ssl.algorithm"public static final String"ssl.enabled"public static final String"ssl.key.store"public static final String"ssl.key.store.algorithm"public static final String"ssl.key.store.password"public static final String"ssl.key.store.type"public static final String"ssl.trust.store"public static final String"ssl.trust.store.algorithm"public static final String"ssl.trust.store.password"public static final String"ssl.trust.store.type"public static final String"ssl.validate.server.certificate"public static final String"ssl.verify.hostname"public static final String"topology.recovery.enabled"public static final String"use.nio"public static final String"username"public static final String"virtual.host" -
com.rabbitmq.client.LongString
-
com.rabbitmq.client.RpcClient
-
com.rabbitmq.client.StringRpcServer
-
com.rabbitmq.client.impl.AMQChannel
-
com.rabbitmq.client.impl.AMQCommand
-
com.rabbitmq.client.impl.AMQConnectionModifier and TypeConstant FieldValue
public static final double1.05 -
com.rabbitmq.client.impl.AMQImpl.Access
-
com.rabbitmq.client.impl.AMQImpl.Access.Request
-
com.rabbitmq.client.impl.AMQImpl.Access.RequestOk
-
com.rabbitmq.client.impl.AMQImpl.Basic
-
com.rabbitmq.client.impl.AMQImpl.Basic.Ack
-
com.rabbitmq.client.impl.AMQImpl.Basic.Cancel
-
com.rabbitmq.client.impl.AMQImpl.Basic.CancelOk
-
com.rabbitmq.client.impl.AMQImpl.Basic.Consume
-
com.rabbitmq.client.impl.AMQImpl.Basic.ConsumeOk
-
com.rabbitmq.client.impl.AMQImpl.Basic.Deliver
-
com.rabbitmq.client.impl.AMQImpl.Basic.Get
-
com.rabbitmq.client.impl.AMQImpl.Basic.GetEmpty
-
com.rabbitmq.client.impl.AMQImpl.Basic.GetOk
-
com.rabbitmq.client.impl.AMQImpl.Basic.Nack
-
com.rabbitmq.client.impl.AMQImpl.Basic.Publish
-
com.rabbitmq.client.impl.AMQImpl.Basic.Qos
-
com.rabbitmq.client.impl.AMQImpl.Basic.QosOk
-
com.rabbitmq.client.impl.AMQImpl.Basic.Recover
-
com.rabbitmq.client.impl.AMQImpl.Basic.RecoverAsync
-
com.rabbitmq.client.impl.AMQImpl.Basic.RecoverOk
-
com.rabbitmq.client.impl.AMQImpl.Basic.Reject
-
com.rabbitmq.client.impl.AMQImpl.Basic.Return
-
com.rabbitmq.client.impl.AMQImpl.Channel
-
com.rabbitmq.client.impl.AMQImpl.Channel.Close
-
com.rabbitmq.client.impl.AMQImpl.Channel.CloseOk
-
com.rabbitmq.client.impl.AMQImpl.Channel.Flow
-
com.rabbitmq.client.impl.AMQImpl.Channel.FlowOk
-
com.rabbitmq.client.impl.AMQImpl.Channel.Open
-
com.rabbitmq.client.impl.AMQImpl.Channel.OpenOk
-
com.rabbitmq.client.impl.AMQImpl.Confirm
-
com.rabbitmq.client.impl.AMQImpl.Confirm.Select
-
com.rabbitmq.client.impl.AMQImpl.Confirm.SelectOk
-
com.rabbitmq.client.impl.AMQImpl.Connection
-
com.rabbitmq.client.impl.AMQImpl.Connection.Blocked
-
com.rabbitmq.client.impl.AMQImpl.Connection.Close
-
com.rabbitmq.client.impl.AMQImpl.Connection.CloseOk
-
com.rabbitmq.client.impl.AMQImpl.Connection.Open
-
com.rabbitmq.client.impl.AMQImpl.Connection.OpenOk
-
com.rabbitmq.client.impl.AMQImpl.Connection.Secure
-
com.rabbitmq.client.impl.AMQImpl.Connection.SecureOk
-
com.rabbitmq.client.impl.AMQImpl.Connection.Start
-
com.rabbitmq.client.impl.AMQImpl.Connection.StartOk
-
com.rabbitmq.client.impl.AMQImpl.Connection.Tune
-
com.rabbitmq.client.impl.AMQImpl.Connection.TuneOk
-
com.rabbitmq.client.impl.AMQImpl.Connection.Unblocked
-
com.rabbitmq.client.impl.AMQImpl.Connection.UpdateSecret
-
com.rabbitmq.client.impl.AMQImpl.Connection.UpdateSecretOk
-
com.rabbitmq.client.impl.AMQImpl.Exchange
-
com.rabbitmq.client.impl.AMQImpl.Exchange.Bind
-
com.rabbitmq.client.impl.AMQImpl.Exchange.BindOk
-
com.rabbitmq.client.impl.AMQImpl.Exchange.Declare
-
com.rabbitmq.client.impl.AMQImpl.Exchange.DeclareOk
-
com.rabbitmq.client.impl.AMQImpl.Exchange.Delete
-
com.rabbitmq.client.impl.AMQImpl.Exchange.DeleteOk
-
com.rabbitmq.client.impl.AMQImpl.Exchange.Unbind
-
com.rabbitmq.client.impl.AMQImpl.Exchange.UnbindOk
-
com.rabbitmq.client.impl.AMQImpl.Queue
-
com.rabbitmq.client.impl.AMQImpl.Queue.Bind
-
com.rabbitmq.client.impl.AMQImpl.Queue.BindOk
-
com.rabbitmq.client.impl.AMQImpl.Queue.Declare
-
com.rabbitmq.client.impl.AMQImpl.Queue.DeclareOk
-
com.rabbitmq.client.impl.AMQImpl.Queue.Delete
-
com.rabbitmq.client.impl.AMQImpl.Queue.DeleteOk
-
com.rabbitmq.client.impl.AMQImpl.Queue.Purge
-
com.rabbitmq.client.impl.AMQImpl.Queue.PurgeOk
-
com.rabbitmq.client.impl.AMQImpl.Queue.Unbind
-
com.rabbitmq.client.impl.AMQImpl.Queue.UnbindOk
-
com.rabbitmq.client.impl.AMQImpl.Tx
-
com.rabbitmq.client.impl.AMQImpl.Tx.Commit
-
com.rabbitmq.client.impl.AMQImpl.Tx.CommitOk
-
com.rabbitmq.client.impl.AMQImpl.Tx.Rollback
-
com.rabbitmq.client.impl.AMQImpl.Tx.RollbackOk
-
com.rabbitmq.client.impl.AMQImpl.Tx.Select
-
com.rabbitmq.client.impl.AMQImpl.Tx.SelectOk
-
com.rabbitmq.client.impl.SocketFrameHandler
-
com.rabbitmq.client.impl.recovery.RecordedQueue
fr.codinbox.*
-
fr.codinbox.connector.commons.utils.EnvUtilsModifier and TypeConstant FieldValue
"CONNECTOR_KAFKA_""CONNECTOR_DB_""CONNECTOR_RABBITMQ_""CONNECTOR_REDIS_"
org.apache.*
-
org.apache.kafka.clients.admin.AdminClientConfigModifier and TypeConstant FieldValue
public static final String"auto.include.jmx.reporter"public static final String"Deprecated. Whether to automatically include JmxReporter even if it\'s not listed in <code>metric.reporters</code>. This configuration will be removed in Kafka 4.0, users should instead include <code>org.apache.kafka.common.metrics.JmxReporter</code> in <code>metric.reporters</code> in order to enable the JmxReporter."public static final String"bootstrap.controllers"public static final String"A list of host/port pairs to use for establishing the initial connection to the KRaft controller quorum. This list should be in the form <code>host1:port1,host2:port2,...</code>."public static final String"bootstrap.servers"public static final String"client.dns.lookup"public static final String"client.id"public static final String"connections.max.idle.ms"public static final String"default.api.timeout.ms"public static final String"PLAINTEXT"public static final String"enable.metrics.push"public static final String"Whether to enable pushing of client metrics to the cluster, if the cluster has a client metrics subscription which matches this client."public static final String"metadata.max.age.ms"public static final String"metric.reporters"public static final String"metrics.num.samples"public static final String"metrics.recording.level"public static final String"metrics.sample.window.ms"public static final String"receive.buffer.bytes"public static final String"reconnect.backoff.max.ms"public static final String"reconnect.backoff.ms"public static final String"request.timeout.ms"public static final String"retries"public static final String"retry.backoff.max.ms"public static final String"retry.backoff.ms"public static final String"security.protocol"public static final String"security.providers"public static final String"send.buffer.bytes"public static final String"socket.connection.setup.timeout.max.ms"public static final String"socket.connection.setup.timeout.ms"
-
org.apache.kafka.clients.consumer.ConsumerConfigModifier and TypeConstant FieldValue
public static final String"allow.auto.create.topics"public static final String"auto.commit.interval.ms"public static final String"auto.include.jmx.reporter"public static final String"auto.offset.reset"public static final String"What to do when there is no initial offset in Kafka or if the current offset does not exist any more on the server (e.g. because that data has been deleted): <ul><li>earliest: automatically reset the offset to the earliest offset<li>latest: automatically reset the offset to the latest offset</li><li>none: throw exception to the consumer if no previous offset is found for the consumer\'s group</li><li>anything else: throw exception to the consumer.</li></ul><p>Note that altering partition numbers while setting this config to latest may cause message delivery loss since producers could start to send messages to newly added partitions (i.e. no initial offsets exist yet) before consumers reset their offsets."public static final String"bootstrap.servers"public static final String"check.crcs"public static final String"client.dns.lookup"public static final String"client.id"public static final String"client.rack"public static final String"connections.max.idle.ms"public static final booleantruepublic static final String"default.api.timeout.ms"public static final String""public static final booleantruepublic static final int52428800public static final int500public static final int1public static final int1048576public static final int500public static final String"enable.auto.commit"public static final String"enable.metrics.push"public static final String"Whether to enable pushing of client metrics to the cluster, if the cluster has a client metrics subscription which matches this client."public static final String"exclude.internal.topics"public static final String"fetch.max.bytes"public static final String"fetch.max.wait.ms"public static final String"fetch.min.bytes"public static final String"group.id"public static final String"group.instance.id"public static final String"group.protocol"public static final String"The group protocol consumer should use. We currently support \"classic\" or \"consumer\". If \"consumer\" is specified, then the consumer group protocol will be used. Otherwise, the classic group protocol will be used."public static final String"group.remote.assignor"public static final String"The server-side assignor to use. If no assignor is specified, the group coordinator will pick one. This configuration is applied only if <code>group.protocol</code> is set to \"consumer\"."public static final String"heartbeat.interval.ms"public static final String"interceptor.classes"public static final String"A list of classes to use as interceptors. Implementing the <code>org.apache.kafka.clients.consumer.ConsumerInterceptor</code> interface allows you to intercept (and possibly mutate) records received by the consumer. By default, there are no interceptors."public static final String"isolation.level"public static final String"Controls how to read messages written transactionally. If set to <code>read_committed</code>, consumer.poll() will only return transactional messages which have been committed. If set to <code>read_uncommitted</code> (the default), consumer.poll() will return all messages, even transactional messages which have been aborted. Non-transactional messages will be returned unconditionally in either mode. <p>Messages will always be returned in offset order. Hence, in <code>read_committed</code> mode, consumer.poll() will only return messages up to the last stable offset (LSO), which is the one less than the offset of the first open transaction. In particular any messages appearing after messages belonging to ongoing transactions will be withheld until the relevant transaction has been completed. As a result, <code>read_committed</code> consumers will not be able to read up to the high watermark when there are in flight transactions.</p><p> Further, when in <code>read_committed</code> the seekToEnd method will return the LSO</p>"public static final String"key.deserializer"public static final String"Deserializer class for key that implements the <code>org.apache.kafka.common.serialization.Deserializer</code> interface."public static final String"max.partition.fetch.bytes"public static final String"max.poll.interval.ms"public static final String"max.poll.records"public static final String"metadata.max.age.ms"public static final String"metric.reporters"public static final String"metrics.num.samples"public static final String"metrics.recording.level"public static final String"metrics.sample.window.ms"public static final String"partition.assignment.strategy"public static final String"receive.buffer.bytes"public static final String"reconnect.backoff.max.ms"public static final String"reconnect.backoff.ms"public static final String"request.timeout.ms"public static final String"retry.backoff.max.ms"public static final String"retry.backoff.ms"public static final String"security.providers"public static final String"send.buffer.bytes"public static final String"session.timeout.ms"public static final String"socket.connection.setup.timeout.max.ms"public static final String"socket.connection.setup.timeout.ms"public static final String"value.deserializer"public static final String"Deserializer class for value that implements the <code>org.apache.kafka.common.serialization.Deserializer</code> interface." -
org.apache.kafka.clients.consumer.ConsumerRecord<K,
V> Modifier and TypeConstant FieldValuepublic static final long-1Lpublic static final int-1public static final int-1 -
org.apache.kafka.clients.consumer.CooperativeStickyAssignor
-
org.apache.kafka.clients.consumer.RangeAssignor
-
org.apache.kafka.clients.consumer.RoundRobinAssignor
-
org.apache.kafka.clients.consumer.StickyAssignor
-
org.apache.kafka.clients.consumer.internals.AbstractCoordinatorModifier and TypeConstant FieldValue
public static final String"kafka-coordinator-heartbeat-thread"public static final int5000 -
org.apache.kafka.clients.consumer.internals.AbstractStickyAssignor
-
org.apache.kafka.clients.consumer.internals.ConsumerProtocol
-
org.apache.kafka.clients.consumer.internals.ConsumerUtils
-
org.apache.kafka.clients.consumer.internals.LegacyKafkaConsumer<K,
V> -
org.apache.kafka.clients.consumer.internals.NetworkClientDelegate.PollResult
-
org.apache.kafka.clients.producer.KafkaProducer<K,
V> -
org.apache.kafka.clients.producer.ProducerConfigModifier and TypeConstant FieldValue
public static final String"acks"public static final String"auto.include.jmx.reporter"public static final String"batch.size"public static final String"bootstrap.servers"public static final String"buffer.memory"public static final String"client.dns.lookup"public static final String"client.id"public static final String"compression.type"public static final String"connections.max.idle.ms"public static final String"delivery.timeout.ms"public static final String"enable.idempotence"public static final String"When set to \'true\', the producer will ensure that exactly one copy of each message is written in the stream. If \'false\', producer retries due to broker failures, etc., may write duplicates of the retried message in the stream. Note that enabling idempotence requires <code>max.in.flight.requests.per.connection</code> to be less than or equal to 5 (with message ordering preserved for any allowable value), <code>retries</code> to be greater than 0, and <code>acks</code> must be \'all\'. <p>Idempotence is enabled by default if no conflicting configurations are set. If conflicting configurations are set and idempotence is not explicitly enabled, idempotence is disabled. If idempotence is explicitly enabled and conflicting configurations are set, a <code>ConfigException</code> is thrown."public static final String"enable.metrics.push"public static final String"Whether to enable pushing of client metrics to the cluster, if the cluster has a client metrics subscription which matches this client."public static final String"interceptor.classes"public static final String"A list of classes to use as interceptors. Implementing the <code>org.apache.kafka.clients.producer.ProducerInterceptor</code> interface allows you to intercept (and possibly mutate) the records received by the producer before they are published to the Kafka cluster. By default, there are no interceptors."public static final String"key.serializer"public static final String"Serializer class for key that implements the <code>org.apache.kafka.common.serialization.Serializer</code> interface."public static final String"linger.ms"public static final String"max.block.ms"public static final String"max.in.flight.requests.per.connection"public static final String"max.request.size"public static final String"metadata.max.age.ms"public static final String"metadata.max.idle.ms"public static final String"metric.reporters"public static final String"metrics.num.samples"public static final String"metrics.recording.level"public static final String"metrics.sample.window.ms"public static final String"partitioner.adaptive.partitioning.enable"public static final String"partitioner.availability.timeout.ms"public static final String"partitioner.class"public static final String"partitioner.ignore.keys"public static final String"receive.buffer.bytes"public static final String"reconnect.backoff.max.ms"public static final String"reconnect.backoff.ms"public static final String"request.timeout.ms"public static final String"retries"public static final String"retry.backoff.max.ms"public static final String"retry.backoff.ms"public static final String"security.providers"public static final String"send.buffer.bytes"public static final String"socket.connection.setup.timeout.max.ms"public static final String"socket.connection.setup.timeout.ms"public static final String"transaction.timeout.ms"public static final String"The maximum amount of time in milliseconds that a transaction will remain open before the coordinator proactively aborts it. The start of the transaction is set at the time that the first partition is added to it. If this value is larger than the <code>transaction.max.timeout.ms</code> setting in the broker, the request will fail with a <code>InvalidTxnTimeoutException</code> error."public static final String"transactional.id"public static final String"The TransactionalId to use for transactional delivery. This enables reliability semantics which span multiple producer sessions since it allows the client to guarantee that transactions using the same TransactionalId have been completed prior to starting any new transactions. If no TransactionalId is provided, then the producer is limited to idempotent delivery. If a TransactionalId is configured, <code>enable.idempotence</code> is implied. By default the TransactionId is not configured, which means transactions cannot be used. Note that, by default, transactions require a cluster of at least three brokers which is the recommended setting for production; for development you can change this, by adjusting broker setting <code>transaction.state.log.replication.factor</code>."public static final String"value.serializer"public static final String"Serializer class for value that implements the <code>org.apache.kafka.common.serialization.Serializer</code> interface." -
org.apache.kafka.clients.producer.RecordMetadata
-
org.apache.kafka.clients.producer.internals.KafkaProducerMetrics
org.redisson.*
-
org.redisson.api.RExecutorService