Akka.Management
1.5.31
Prefix Reserved
dotnet add package Akka.Management --version 1.5.31
NuGet\Install-Package Akka.Management -Version 1.5.31
<PackageReference Include="Akka.Management" Version="1.5.31" />
paket add Akka.Management --version 1.5.31
#r "nuget: Akka.Management, 1.5.31"
// Install Akka.Management as a Cake Addin #addin nuget:?package=Akka.Management&version=1.5.31 // Install Akka.Management as a Cake Tool #tool nuget:?package=Akka.Management&version=1.5.31
Table Of Contents
- Akka Management
- Akka.Management.Cluster.Bootstrap
Akka Management
Akka Management is the core module of the management utilities which provides a central HTTP endpoint for Akka management extensions.
Basic Usage
With a few exceptions, Akka Management does not start automatically and the routes will only be exposed once you trigger:
AkkaManagement.Get(system).Start();
This allows users to prepare anything further before exposing routes for the bootstrap joining process and other purposes.
NOTE
Once Akka.Management started, you can not add or expose more routes on the HTTP endpoint.
Basic Configuration
Configure Using Akka.Hosting
Setting up Akka.Management through Akka.Hosting is quite simple.
var hostBuilder = new HostBuilder()
.ConfigureServices((context, services) =>
{
services.AddAkka("managementDemo", (builder, provider) =>
{
builder.WithAkkaManagement();
});
});
using(var host = hostBuilder.Build())
{
await host.RunAsync();
}
You can configure hostname and port to use for the HTTP Cluster management by overriding the following:
builder.WithAkkaManagement(
hostName: "127.0.0.1",
port: 8558);
or
builder.WithAkkaManagement(setup =>
{
setup.Http.Hostname = "127.0.0.1";
setup.Http.Port = 8558;
});
or
builder.WithAkkaManagement(new AkkaManagementSetup {
Http = new HttpSetup {
Hostname = "127.0.0.1",
Port = 8558
}
});
Note that the default value for hostname is localhost
When running Akka nodes behind NATs or inside docker containers in bridge mode, it is necessary to set different hostname and port number to bind for the HTTP Server for Http Cluster Management:
builder.WithAkkaManagement(
hostName: "my-public-host-name",
port: 8558,
// Bind to 0.0.0.0:8558 'internally':
bindHostname: "0.0.0.0",
bindPort: 8558);
or
builder.WithAkkaManagement(setup =>
{
setup.Http.Hostname = "my-public-host-name";
setup.Http.Port = 8558;
// Bind to 0.0.0.0:8558 'internally':
setup.Http.BindHostname = "0.0.0.0";
setup.Http.BindPort = 8558;
});
or
builder.WithAkkaManagement(new AkkaManagementSetup {
Http = new HttpSetup {
Hostname = "my-public-host-name",
Port = 8558,
// Bind to 0.0.0.0:8558 'internally':
BindHostname = "0.0.0.0",
BindPort = 8558
}
});
Configure Using HOCON Configuration
You can configure hostname and port to use for the HTTP Cluster management by overriding the following:
akka.management.http.hostname = "127.0.0.1"
akka.management.http.port = 8558
Note that the default value for hostname is localhost
When running Akka nodes behind NATs or inside docker containers in bridge mode, it is necessary to set different hostname and port number to bind for the HTTP Server for Http Cluster Management:
akka.management.http.hostname = "my-public-host-name"
akka.management.http.port = 8558
# Bind to 0.0.0.0:8558 'internally':
akka.management.http.bind-hostname = 0.0.0.0
akka.management.http.bind-port = 8558
Exposed REST API Endpoints
Akka.Management.Cluster.Bootstrap
Akka.Management REST API endpoint are exposed by default, it is being mapped to http://{host}:{port}/bootstrap/seed-nodes
.
The end point will return 200-OK with JSON serialized data of current seed nodes inside the HTTP body if the ActorSystem
actor provider is set to "cluster" and returns 503-ServiceUnavailable if clustering is not available.
Security
Note that http protocol is used by default and, as of now, there is no way to set up security on any HTTP endpoints. Management endpoints are not designed to be and should never be opened to the public.
Stopping Akka Management
In a dynamic environment you might stop instances of Akka Management, for example if you want to free up resources taken by the HTTP server serving the Management routes.
You can do so by calling Stop() on AkkaManagement. This method return a Task to inform when the server has been stopped.
var management = AkkaManagement.Get(system);
await management.Start();
//...
await management.Stop();
Reference HOCON Configuration
######################################################
# Akka Http Cluster Management Reference Config File #
######################################################
# This is the reference config file that contains all the default settings.
# Make your edits/overrides in your application.conf.
akka.management {
http {
# The hostname where the HTTP Server for Http Cluster Management will be started.
# This defines the interface to use.
# akka.remote.dot-netty.tcp.public-hostname is used if not overriden or empty.
# if akka.remote.dot-netty.tcp.public-hostname is empty, Dns.GetHostName is used.
hostname = "<hostname>"
# The port where the HTTP Server for Http Cluster Management will be bound.
# The value will need to be from 0 to 65535.
port = 8558
# Use this setting to bind a network interface to a different hostname or ip
# than the HTTP Server for Http Cluster Management.
# Use "0.0.0.0" to bind to all interfaces.
# akka.management.http.hostname if empty
bind-hostname = ""
# Use this setting to bind a network interface to a different port
# than the HTTP Server for Http Cluster Management. This may be used
# when running akka nodes in a separated networks (under NATs or docker containers).
# Use 0 if you want a random available port.
#
# akka.management.http.port if empty
bind-port = ""
# path prefix for all management routes, usually best to keep the default value here. If
# specified, you'll want to use the same value for all nodes that use akka management so
# that they can know which path to access each other on.
base-path = ""
# Definition of management route providers which shall contribute routes to the management HTTP endpoint.
# Management route providers should be regular extensions that additionally extend the
# `Akka.Management.Dsl.IManagementRoutesProvider` interface
#
# Libraries may register routes into the management routes by defining entries to this setting
# the library `reference.conf`:
#
# akka.management.http.routes {
# name = "FQCN"
# }
#
# Where the `name` of the entry should be unique to allow different route providers to be registered
# by different libraries and applications.
#
# The FQCN is the fully qualified class name of the `ManagementRoutesProvider`.
#
# Unlike the scala version, Akka.NET Akka.Management does not provide any health check capability but
# provides cluster bootstrap functionality instead. If you need any health check capability,
# please install the Akka.HealthCheck NuGet package.
#
routes {
# registers bootstrap routes to be included in akka-management's http endpoint
cluster-bootstrap = "Akka.Management.Cluster.Bootstrap.ClusterBootstrapProvider, Akka.Management"
}
# Should Management route providers only expose read only endpoints? It is up to each route provider
# to adhere to this property
route-providers-read-only = true
}
cluster.bootstrap {
# Cluster Bootstrap will always attempt to join an existing cluster if possible. However
# if no contact point advertises any seed-nodes a new cluster will be formed by the
# node with the lowest address as decided by [[LowestAddressJoinDecider]].
# Setting `new-cluster-enabled=off` after an initial cluster has formed is recommended to prevent new clusters
# forming during a network partition when nodes are redeployed or restarted.
new-cluster-enabled = on
# Configuration for the first phase of bootstrapping, during which contact points are discovered
# using the configured service discovery mechanism (e.g. DNS records).
contact-point-discovery {
# Define this name to be looked up in service discovery for "neighboring" nodes
# If undefined, the name will be taken from the AKKA__CLUSTER__BOOTSTRAP__SERVICE_NAME
# environment variable or extracted from the ActorSystem name
service-name = "<service-name>"
# The portName passed to discovery. This should be set to the name of the port for Akka Management
# If set to "", `null` is passed to the discovery mechanism.
port-name = ""
# The protocol passed to discovery.
# If set to "" None is passed.
protocol = "tcp"
# Added as suffix to the service-name to build the effective-service name used in the contact-point service lookups
# If undefined, nothing will be appended to the service-name.
#
# Examples, set this to:
# "default.svc.cluster.local" or "my-namespace.svc.cluster.local" for kubernetes clusters.
service-namespace = "<service-namespace>"
# The effective service name is the exact string that will be used to perform service discovery.
#
# Set this value to a specific string to override the default behaviour of building the effective name by
# concatenating the `service-name` with the optional `service-namespace` (e.g. "name.default").
effective-name = "<effective-name>"
# Config path of discovery method to be used to locate the initial contact points.
# It must be a fully qualified config path to the discovery's config section.
#
# By setting this to `akka.discovery` we ride on the configuration mechanisms that akka-discovery has,
# and reuse what is configured for it. You can set it explicitly to something else here, if you want to
# use a different discovery mechanism for the bootstrap than for the rest of the application.
discovery-method = akka.discovery
# Amount of time for which a discovery observation must remain "stable"
# (i.e. discovered contact-points list did not change) before a join decision can be made.
# This is done to decrease the likelihood of performing decisions on fluctuating observations.
#
# This timeout represents a tradeoff between safety and quickness of forming a new cluster.
stable-margin = 5s
# Interval at which service discovery will be polled in search for new contact-points
#
# Note that actual timing of lookups will be the following:
# - perform initial lookup; interval is this base interval
# - await response within resolve-timeout
# (this can be larger than interval, which means interval effectively is resolveTimeout + interval,
# this has been specifically made so, to not hit discovery services with requests while the lookup is being serviced)
# - if failure happens apply backoff to interval (the backoff growth is exponential)
# - if no failure happened, and we receive a resolved list of services, schedule another lookup in interval time
# - if previously failures happened during discovery, a successful lookup resets the interval to `interval` again
# - repeat until stable-margin is reached
interval = 1s
# Adds "noise" to vary the intervals between retries slightly (0.2 means 20% of base value).
# This is important in order to avoid the various nodes performing lookups in the same interval,
# potentially causing a thundering heard effect. Usually there is no need to tweak this parameter.
exponential-backoff-random-factor = 0.2
# Maximum interval to which the exponential backoff is allowed to grow
exponential-backoff-max = 15s
# The smallest number of contact points that need to be discovered before the bootstrap process can start.
# For optimal safety during cluster formation, you may want to set these value to the number of initial
# nodes that you know will participate in the cluster (e.g. the value of `spec.replicas` as set in your kubernetes config.
required-contact-point-nr = 2
# Timeout for getting a reply from the service-discovery subsystem
resolve-timeout = 3s
# Does a successful response have to be received by all contact points.
# Used by the LowestAddressJoinDecider
# Can be set to false in environments where old contact points may still be in service discovery
# or when using local discovery and cluster formation is desired without starting all the nodes
# Required-contact-point-nr still needs to be met
contact-with-all-contact-points = true
}
# Configured how we communicate with the contact point once it is discovered
contact-point {
# If no port is discovered along with the host/ip of a contact point this port will be used as fallback
# Also, when no port-name is used and multiple results are returned for a given service with at least one
# port defined, this port is used to disambiguate.
fallback-port = "<fallback-port>"
# by default when no port-name is set only the contact points that contain the fallback-port
# are used for probing. This makes the scenario where each akka node has multiple ports
# returned from service discovery (e.g. management, remoting, front-end HTTP) work without
# having to configure a port-name. If instead service discovery will return only akka management
# ports without specifying a port-name, e.g. management has dynamic ports and its own service
# name, then set this to false to stop the results being filtered
filter-on-fallback-port = true
# If some discovered seed node will keep failing to connect for specified period of time,
# it will initiate rediscovery again instead of keep trying.
probing-failure-timeout = 3s
# Interval at which contact points should be polled
# the effective interval used is this value plus the same value multiplied by the jitter value
probe-interval = 1s
# Max amount of jitter to be added on retries
probe-interval-jitter = 0.2
}
join-decider {
# Implementation of JoinDecider.
# It must extend Akka.Management.Cluster.Bootstrap.IJoinDecider and
# have public constructor with ActorSystem and ClusterBootstrapSettings
# parameters.
class = "Akka.Management.Cluster.Bootstrap.LowestAddressJoinDecider, Akka.Management"
}
}
}
Akka.Management.Cluster.Bootstrap
Akka Cluster Bootstrap helps forming (or joining to) a cluster by using Akka Discovery to discover peer nodes. It is an alternative to configuring static seed-nodes in dynamic deployment environments such as on Kubernetes or AWS.
It builds on the flexibility of Akka Discovery, leveraging a range of discovery mechanisms depending on the environment you want to run your cluster in.
Cluster bootstrap depends on:
- Akka.Discovery to discover other members of the cluster
- Akka.Management to host HTTP endpoints used during the bootstrap process
Usage
Akka management must be started as well as the bootstrap process, this can either be done through Akka.Hosting, HOCON config, or programmatically.
Setting up Cluster.Bootstrap using Akka.Hosting
ClusterBootstrap can be enabled using the Akka.Hosting extension method:
using var host = new HostBuilder()
.ConfigureServices((hostContext, services) =>
{
services.AddAkka(systemName, (builder, provider) =>
{
// Add Akka.Remote support.
builder.WithRemoting(port: 4053);
// Add Akka.Cluster support
builder.WithClustering();
// Add Akka.Management.Cluster.Bootstrap support
builder.WithClusterBootstrap(setup =>
{
setup.ContactPointDiscovery.ServiceName = "clusterbootstrap";
}, autoStart: true);
// Add Akka.Discovery.KubernetesApi support
builder.WithKubernetesDiscovery("app=clusterbootstrap");
});
})
Note that to start ClusterBootstrap, you will also need to provide an Akka.Discovery method to use. In the example above, this is done using the WithKubernetesDiscovery()
method call.
If not set to start automatically, you will need to manually start both Akka.Management and ClusterBootstrap:
builder.AddStartup(async (system, registry) =>
{
await AkkaManagement.Get(system).Start();
await ClusterBootstrap.Get(system).Start();
});
If management or bootstrap configuration is incorrect, the autostart will log an error and terminate the actor system.
Setting Up Cluster.Bootstrap from HOCON Configuration
Listing the ClusterBootstrap extension among the autoloaded akka.extensions in your configuration will cause it to autostart:
# trigger autostart by loading the extension through config
akka.extensions = ["Akka.Management.Cluster.Bootstrap.ClusterBootstrapProvider, Akka.Management"]
If management or bootstrap configuration is incorrect, the autostart will log an error and terminate the actor system.
Setting Up Cluster.Bootstrap Programmatically
// Akka Management hosts the HTTP routes used by bootstrap
await AkkaManagement.Get(system).Start();
// Starting the bootstrap process needs to be done explicitly
await ClusterBootstrap.Get(system).Start();
Ensure that seed-nodes
is not present in configuration and that either autoloading through config or start() is called on every node.
The following configuration is required, more details for each and additional configuration can be found in the reference configuration:
akka.management.cluster.bootstrap.contact-point-discovery.service-name
: a unique name in the deployment environment for this cluster instance which is used to lookup peers in service discovery. If unset, it will be derived from the ActorSystem name.akka.management.cluster.bootstrap.contact-point-discovery.discovery-method
: the intended service discovery mechanism (from what choices Akka Discovery provides). If unset, falls back to the system-wide default from akka.discovery.method.
Exposed Akka.Management REST API Endpoint
Akka.Management.Cluster.Bootstrap
will add a new REST HTTP API endpoint to the Akka.Management
HTTP
server at the address http://{host}:{port}/bootstrap/seed-nodes
. Calling a GET on this endpoint will
return a JSON document containing the Akka cluster address of the node and a list of up to 5 seed nodes
from the that Akka node.
How It Works
- Each node exposes an HTTP endpoint
/bootstrap/seed-nodes
. This is provided by Akka.Management.Cluster.Bootstrap and exposed automatically by starting Akka.Management. /bootstrap/seed-nodes
will query its internal cluster state and returns a list of members that are either in the up, weakly up, or joining state- During bootstrap each node queries service discovery repeatedly to get the initial contact points until at least the number of contact points (and recommended exactly equal) as defined in contact-point-discovery.required-contact-point-nr has been found.
- Each node then probes these found contact points
/bootstrap/seed-nodes
endpoint to see if a cluster has already been formed- If there is an existing cluster, it joins the cluster and bootstrapping is finished.
- If no cluster exists and every node returns an empty list of seed-nodes, The node with the lowest address from the set of contact points forms a new cluster.
- Other nodes will start to see the
/bootstrap/seed-nodes
of the node that self-joined and will join its cluster.
Please see the complete bootstrap process documentation for more information.
Joining Mechanism Precedence
As Akka Cluster allows nodes to join to a cluster using multiple different methods, the precedence of each method is strictly defined and is as follows:
- If akka.cluster.seed-nodes (in your HOCON configuration) are non-empty, those nodes will be joined, and bootstrap will NOT execute even if start() is called or autostart through configuration is enabled, however a warning will be logged.
- If an explicit Cluster.Join or Cluster.JoinSeedNodes is invoked before the bootstrap completes, that joining would take precedence over the bootstrap (but it’s not recommended to do so, see below).
- The Cluster Bootstrap mechanism takes some time to complete, but eventually issues a joinSeednodes.
[!WARNING] It is NOT recommended to mix various joining mechanisms. Pick one mechanism and stick to it in order to avoid any surprises during cluster formation. E.g. do NOT set akka.cluster.seed-nodes and do NOT call Cluster.Join if you are going to be using the Bootstrap mechanism.
Deployment Considerations
Initial deployment
Cluster Bootstrap will always attempt to join an existing cluster if possible. However if no other contact point advertises any seed-nodes
, a new cluster will be formed by the node decided by the JoinDecider
where the default sorts the addresses then picks the lowest.
The HOCON setting akka.management.cluster.bootstrap.new-cluster-enabled
can be used to disable new cluster formation and to only allow the node to join existing clusters.
- On initial deployment use the default akka.management.cluster.bootstrap.new-cluster-enabled=on
- Following the initial deployment it is recommended to set
akka.management.cluster.bootstrap.new-cluster-enabled=off
with an immediate re-deployment once the initial cluster has formed
This can be used to provide additional safety during restarts and redeploys while there is a network partition present. Without new cluster formation disabled, an isolated set of nodes could form a new cluster, creating a split-brain.
Recommended Configuration
When using the bootstrap module, there are some underlying Akka Cluster settings that should be specified to ensure that your deployment is robust.
Since the target environments for this module are dynamic, that is, instances can come and go, failure needs to be considered. The following configuration will result in your application being shut down after 30 seconds if it is unable to join the discovered seed nodes. In this case, the orchestrator (i.e. Kubernetes or Marathon) will restart your node and the operation will (presumably) eventually succeed. You will want to specify the following in your HOCON configuration:
akka.cluster.shutdown-after-unsuccessful-join-seed-nodes = 30s
Rolling Updates
Graceful Shutdown
Akka Cluster can handle hard failures using a downing provider such as the split brain resolver discussed below. However this should not be relied upon for regular rolling redeploys. Features such as ClusterSingletons and ClusterSharding can safely restart actors on new nodes far quicker when it is certain that a node has shutdown rather than crashed.
Graceful leaving will happen with the default settings as it is part of Coordinated Shutdown. Just ensure that a node is sent a SIGTERM and not a SIGKILL. Environments such as Kubernetes will do this, it is important to ensure that if the CLR is wrapped with a script that it forwards the signal.
Upon receiving a SIGTERM Coordinated Shutdown will:
- Perform a Cluster.Get(system).Leave() on itself
- The status of the member will be changed to Exiting while allowing any shards to be shutdown gracefully and ClusterSingletons to be migrated if this was the oldest node. Finally the node is removed from the Akka Cluster membership.
Number of Nodes to Redeploy At Once
Akka bootstrap requires a stable-period
where service discovery returns a stable set of contact points. When doing rolling updates, it is best to wait for a node (or group of nodes) to finish joining the cluster before adding and removing other nodes.
Cluster Singleton
ClusterSingleton
s run on the oldest node in the cluster. To avoid singletons moving during every node deployment, it is advised to start a rolling redeploy starting at the newest node. Then ClusterSingletons only move once. This is the default behaviour for Kubernetes deployments. Cluster Sharding uses a singleton internally so this is important, even if not using singletons directly.
Split Brains and Ungraceful Shutdown
Nodes can crash causing cluster members to become unreachable. This is a tricky problem as it is not possible to distinguish between a network partition and a node failure. To rectify this in an automated manner, make sure you enable the Split Brain Resolver. This module has a number of strategies that can ensure that the cluster continues to function during network partitions and node failures.
Reference Configuration
######################################################
# Akka Cluster Bootstrap Config #
######################################################
akka.management {
# registers bootstrap routes to be included in akka-management's http endpoint
http.routes {
cluster-bootstrap = "Akka.Management.Cluster.Bootstrap.ClusterBootstrapProvider, Akka.Management.Cluster.Bootstrap"
}
cluster.bootstrap {
# Cluster Bootstrap will always attempt to join an existing cluster if possible. However
# if no contact point advertises any seed-nodes a new cluster will be formed by the
# node with the lowest address as decided by [[LowestAddressJoinDecider]].
# Setting `new-cluster-enabled=off` after an initial cluster has formed is recommended to prevent new clusters
# forming during a network partition when nodes are redeployed or restarted.
new-cluster-enabled = on
# Configuration for the first phase of bootstrapping, during which contact points are discovered
# using the configured service discovery mechanism (e.g. DNS records).
contact-point-discovery {
# Define this name to be looked up in service discovery for "neighboring" nodes
# If undefined, the name will be taken from the AKKA__CLUSTER__BOOTSTRAP__SERVICE_NAME
# environment variable or extracted from the ActorSystem name
service-name = "<service-name>"
# The portName passed to discovery. This should be set to the name of the port for Akka Management
# If set to "", `null` is passed to the discovery mechanism.
port-name = ""
# The protocol passed to discovery.
# If set to "" None is passed.
protocol = "tcp"
# Added as suffix to the service-name to build the effective-service name used in the contact-point service lookups
# If undefined, nothing will be appended to the service-name.
#
# Examples, set this to:
# "default.svc.cluster.local" or "my-namespace.svc.cluster.local" for kubernetes clusters.
service-namespace = "<service-namespace>"
# The effective service name is the exact string that will be used to perform service discovery.
#
# Set this value to a specific string to override the default behaviour of building the effective name by
# concatenating the `service-name` with the optional `service-namespace` (e.g. "name.default").
effective-name = "<effective-name>"
# Config path of discovery method to be used to locate the initial contact points.
# It must be a fully qualified config path to the discovery's config section.
#
# By setting this to `akka.discovery` we ride on the configuration mechanisms that akka-discovery has,
# and reuse what is configured for it. You can set it explicitly to something else here, if you want to
# use a different discovery mechanism for the bootstrap than for the rest of the application.
discovery-method = akka.discovery
# Amount of time for which a discovery observation must remain "stable"
# (i.e. discovered contact-points list did not change) before a join decision can be made.
# This is done to decrease the likelihood of performing decisions on fluctuating observations.
#
# This timeout represents a tradeoff between safety and quickness of forming a new cluster.
stable-margin = 5s
# Interval at which service discovery will be polled in search for new contact-points
#
# Note that actual timing of lookups will be the following:
# - perform initial lookup; interval is this base interval
# - await response within resolve-timeout
# (this can be larger than interval, which means interval effectively is resolveTimeout + interval,
# this has been specifically made so, to not hit discovery services with requests while the lookup is being serviced)
# - if failure happens apply backoff to interval (the backoff growth is exponential)
# - if no failure happened, and we receive a resolved list of services, schedule another lookup in interval time
# - if previously failures happened during discovery, a successful lookup resets the interval to `interval` again
# - repeat until stable-margin is reached
interval = 1s
# Adds "noise" to vary the intervals between retries slightly (0.2 means 20% of base value).
# This is important in order to avoid the various nodes performing lookups in the same interval,
# potentially causing a thundering herd effect. Usually there is no need to tweak this parameter.
exponential-backoff-random-factor = 0.2
# Maximum interval to which the exponential backoff is allowed to grow
exponential-backoff-max = 15s
# The smallest number of contact points that need to be discovered before the bootstrap process can start.
# For optimal safety during cluster formation, you may want to set these value to the number of initial
# nodes that you know will participate in the cluster (e.g. the value of `spec.replicas` as set in your kubernetes config.
required-contact-point-nr = 2
# Timeout for getting a reply from the service-discovery subsystem
resolve-timeout = 3s
# Does a successful response have to be received by all contact points.
# Used by the LowestAddressJoinDecider
# Can be set to false in environments where old contact points may still be in service discovery
# or when using local discovery and cluster formation is desired without starting all the nodes
# Required-contact-point-nr still needs to be met
contact-with-all-contact-points = true
}
# Configured how we communicate with the contact point once it is discovered
contact-point {
# If no port is discovered along with the host/ip of a contact point this port will be used as fallback
# Also, when no port-name is used and multiple results are returned for a given service with at least one
# port defined, this port is used to disambiguate.
fallback-port = "<fallback-port>"
# by default when no port-name is set only the contact points that contain the fallback-port
# are used for probing. This makes the scenario where each akka node has multiple ports
# returned from service discovery (e.g. management, remoting, front-end HTTP) work without
# having to configure a port-name. If instead service discovery will return only akka management
# ports without specifying a port-name, e.g. management has dynamic ports and its own service
# name, then set this to false to stop the results being filtered
filter-on-fallback-port = true
# If some discovered seed node will keep failing to connect for specified period of time,
# it will initiate rediscovery again instead of keep trying.
probing-failure-timeout = 3s
# Interval at which contact points should be polled
# the effective interval used is this value plus the same value multiplied by the jitter value
probe-interval = 1s
# Max amount of jitter to be added on retries
probe-interval-jitter = 0.2
}
join-decider {
# Implementation of JoinDecider.
# It must extend Akka.Management.Cluster.Bootstrap.IJoinDecider and
# have public constructor with ActorSystem and ClusterBootstrapSettings
# parameters.
class = "Akka.Management.Cluster.Bootstrap.LowestAddressJoinDecider, Akka.Management"
}
}
}
Product | Versions Compatible and additional computed target framework versions. |
---|---|
.NET | net5.0 was computed. net5.0-windows was computed. net6.0 is compatible. net6.0-android was computed. net6.0-ios was computed. net6.0-maccatalyst was computed. net6.0-macos was computed. net6.0-tvos was computed. net6.0-windows was computed. net7.0 was computed. net7.0-android was computed. net7.0-ios was computed. net7.0-maccatalyst was computed. net7.0-macos was computed. net7.0-tvos was computed. net7.0-windows was computed. net8.0 was computed. net8.0-android was computed. net8.0-browser was computed. net8.0-ios was computed. net8.0-maccatalyst was computed. net8.0-macos was computed. net8.0-tvos was computed. net8.0-windows was computed. |
.NET Core | netcoreapp2.0 was computed. netcoreapp2.1 was computed. netcoreapp2.2 was computed. netcoreapp3.0 was computed. netcoreapp3.1 was computed. |
.NET Standard | netstandard2.0 is compatible. netstandard2.1 was computed. |
.NET Framework | net461 was computed. net462 was computed. net463 was computed. net47 was computed. net471 was computed. net472 was computed. net48 was computed. net481 was computed. |
MonoAndroid | monoandroid was computed. |
MonoMac | monomac was computed. |
MonoTouch | monotouch was computed. |
Tizen | tizen40 was computed. tizen60 was computed. |
Xamarin.iOS | xamarinios was computed. |
Xamarin.Mac | xamarinmac was computed. |
Xamarin.TVOS | xamarintvos was computed. |
Xamarin.WatchOS | xamarinwatchos was computed. |
-
.NETStandard 2.0
- Akka.Cluster (>= 1.5.31)
- Akka.Discovery (>= 1.5.31)
- Akka.Hosting (>= 1.5.31)
- Akka.Http.Shim (>= 1.5.31)
-
net6.0
- Akka.Cluster (>= 1.5.31)
- Akka.Discovery (>= 1.5.31)
- Akka.Hosting (>= 1.5.31)
- Akka.Http.Shim (>= 1.5.31)
NuGet packages (1)
Showing the top 1 NuGet packages that depend on Akka.Management:
Package | Downloads |
---|---|
Akka.Management.Cluster.Bootstrap
Cluster bootstrap module for Akka.NET |
GitHub repositories (1)
Showing the top 1 popular GitHub repositories that depend on Akka.Management:
Repository | Stars |
---|---|
akkadotnet/akka.net
Canonical actor model implementation for .NET with local + distributed actors in C# and F#.
|
Version | Downloads | Last updated |
---|---|---|
1.5.31 | 2,256 | 11/11/2024 |
1.5.30 | 9,083 | 10/3/2024 |
1.5.29 | 870 | 10/1/2024 |
1.5.27 | 24,056 | 7/30/2024 |
1.5.26.1 | 5,544 | 7/16/2024 |
1.5.26 | 143 | 7/15/2024 |
1.5.26-beta3 | 2,807 | 7/2/2024 |
1.5.26-beta2 | 543 | 6/28/2024 |
1.5.26-beta1 | 90 | 6/27/2024 |
1.5.25.1 | 5,228 | 6/25/2024 |
1.5.25 | 2,684 | 6/17/2024 |
1.5.24 | 2,435 | 6/11/2024 |
1.5.19 | 11,971 | 4/17/2024 |
1.5.18-beta2 | 1,277 | 3/27/2024 |
1.5.18-beta1 | 170 | 3/20/2024 |
1.5.17.1 | 10,595 | 3/4/2024 |
1.5.15 | 12,963 | 1/12/2024 |
1.5.7 | 40,978 | 5/23/2023 |
1.5.5 | 1,768 | 5/8/2023 |
1.5.0 | 5,698 | 3/2/2023 |
1.5.0-beta6 | 142 | 3/1/2023 |
1.5.0-alpha4 | 106 | 2/17/2023 |
1.0.3 | 1,077 | 2/13/2023 |
1.0.2 | 455 | 2/8/2023 |
1.0.1 | 862 | 1/31/2023 |
1.0.0 | 624 | 1/18/2023 |
1.0.0-beta2 | 225 | 1/7/2023 |
1.0.0-beta1 | 139 | 1/6/2023 |
0.3.0-beta4 | 1,315 | 12/2/2022 |
0.3.0-beta3 | 727 | 11/7/2022 |
0.3.0-beta2 | 839 | 10/20/2022 |
0.3.0-beta1 | 344 | 10/6/2022 |
0.2.5-beta4 | 1,154 | 8/31/2022 |
0.2.5-beta3 | 8,713 | 8/16/2022 |
0.2.5-beta2 | 200 | 8/8/2022 |
0.2.5-beta1 | 183 | 8/1/2022 |
0.2.4-beta3 | 808 | 5/5/2022 |
0.2.4-beta2 | 255 | 4/14/2022 |
0.2.4-beta1 | 10,738 | 12/2/2021 |
0.2.3-beta2 | 1,415 | 10/5/2021 |
0.2.3-beta1 | 263 | 10/4/2021 |
0.2.2-beta1 | 243 | 9/29/2021 |
0.2.1-beta4 | 220 | 9/3/2021 |
0.2.1-beta3 | 279 | 8/16/2021 |
0.2.1-beta2 | 241 | 8/12/2021 |
0.2.1-beta1 | 256 | 8/6/2021 |
0.2.0-beta1 | 284 | 8/5/2021 |
Update to [Akka.NET v1.5.31](https://github.com/akkadotnet/akka.net/releases/tag/1.5.31)
Update to [Akka.Hosting v1.5.31](https://github.com/akkadotnet/Akka.Hosting/releases/tag/1.5.31)
[Coordination.Azure: Fix LeaseActor bugs](https://github.com/akkadotnet/Akka.Management/pull/2891)
[Coordination.Azure: Fix AzureLease bugs](https://github.com/akkadotnet/Akka.Management/pull/2890)
[Coordination.KubernetesApi: Modernize codes](https://github.com/akkadotnet/Akka.Management/pull/2898)