From 200b2a31cd6ddb724fa3b9ef804a1298cc206684 Mon Sep 17 00:00:00 2001
From: Lisa Durant <85700989+lisadurant@users.noreply.github.com>
Date: Thu, 15 Feb 2024 08:09:38 -0500
Subject: [PATCH] Update README.md
---
nxrm-aws-resiliency/README.md | 205 ++--------------------------------
1 file changed, 11 insertions(+), 194 deletions(-)
diff --git a/nxrm-aws-resiliency/README.md b/nxrm-aws-resiliency/README.md
index 2410bd3..65d1872 100644
--- a/nxrm-aws-resiliency/README.md
+++ b/nxrm-aws-resiliency/README.md
@@ -12,201 +12,18 @@
Eclipse Foundation. All other trademarks are the property of their respective owners.
-->
+# ⚠️ Archive Notice
-# Helm Chart for a Resilient Nexus Repository Deployment in AWS
+As of February 9, 2024, we now provide one [HA/Resiliency Helm Chart](https://github.com/sonatype/nxrm3-ha-repository/tree/main/nxrm-ha) that supports both high availability and resilient deployments in AWS, Azure, or on-premises in a Kubernetes cluster. This is our only supported Helm chart for deploying Sonatype Nexus Repository; it requires a PostgreSQL database and a Pro license.
-This Helm chart configures the Kubernetes resources that are needed for a resilient Nexus Repository deployment on AWS as described in our documented [single-node cloud resilient deployment example using AWS](https://help.sonatype.com/repomanager3/planning-your-implementation/resiliency-and-high-availability/single-node-cloud-resilient-deployment-example-using-aws).
+# Helm Chart Instructions
-Use the checklist below to determine if this Helm chart is suitable for your deployment needs.
+See the [HA/Resiliency Helm Chart in GitHub](https://github.com/sonatype/nxrm3-ha-repository/tree/main/nxrm-ha) for details on the new combined Helm chart.
----
-
-## When to Use This Helm Chart
-Use this Helm chart if you are doing any of the following:
-- Deploying Nexus Repository Pro to an AWS cloud environment with the desire for automatic failover across Availability Zones (AZs) within a single region
-- Planning to configure a single Nexus Repository Pro instance within your Kubernetes/EKS cluster with two or more nodes spread across different AZs within an AWS region
-- Using an external PostgreSQL database
-
-> **Note**: A Nexus Repository Pro license is required for our resilient deployment options. Your Nexus Repository Pro license file must be stored externally as mounted from AWS Secrets AWS (required).
-
----
-
-## Prerequisites for This Chart
-In order to set up an environment like the one illustrated above and described in this section, you will need the following:
-
-- Kubernetes 1.19+
-- [kubectl](https://kubernetes.io/docs/tasks/tools/)
-- [Helm 3](https://helm.sh/docs/intro/install/)
-- A Nexus Repository Pro license
-- An AWS account with permissions for accessing the following AWS services:
- - Elastic Kubernetes Service (EKS)
- - Relational Database Service (RDS) for PostgreSQL
- - Application Load Balancer (ALB)
- - CloudWatch
- - Simple Storage Service (S3)
- - Secrets Manager
-
-You will also need to complete the steps below. See the referenced AWS documentation for detailed configuration steps. Also see [our resiliency documentation](https://help.sonatype.com/repomanager3/planning-your-implementation/resiliency-and-high-availability/single-node-cloud-resilient-deployment-example-using-aws) for more details about why these steps are necessary and how each AWS solution functions within a resilient deployment:
-1. Configure an EKS cluster - [AWS documentation for managed nodes (i.e., EC2)](https://docs.aws.amazon.com/eks/latest/userguide/getting-started-console.html)
-2. Create an Aurora database cluster - [AWS documentation for creating an Aurora database cluster](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/Aurora.CreateInstance.html)
-3. Deploy the AWS Load Balancer Controller (LBC) to your EKS cluster - [AWS documentation for deploying the AWS LBC to your EKS cluster](https://docs.aws.amazon.com/eks/latest/userguide/aws-load-balancer-controller.html)
-4. Install AWS Secrets Store CSI drivers - You need to create an IAM service account using the ```eksctl create iamserviceaccount``` command. Before proceeding, read the points below as they contain important required steps to ensure this helm chart will work for you:
- - **You must include two additional command parameters when running the command**: ```--role-only``` and ```--namespace ```
- - It is important to include the ```--role-only``` option in the ```eksctl create iamserviceaccount``` command so that the helm chart manages the Kubernetes service account.
- - **The namespace you specify to the ```eksctl create iamserviceaccount``` must be the same namespace into which you will deploy the Nexus Repository pod.**
- - Although the namespace does not exist at this point, you must specify it as part of the command. **Do not create that namespace manually beforehand**; the helm chart will create and manage it.
- - You should specify this same namespace as the value of ```nexusNs``` in your values.yaml.
- - Follow the instructions provided in the [AWS Secrets Store CSI drivers documentation](https://github.com/aws/secrets-store-csi-driver-provider-aws/blob/main/README.md) to install the AWS Secrets Store CSI drivers; ensure that you follow the additional instructions in the bullets above when you reach the ```eksctl create iamserviceaccount``` command on that page.
-5. Ensure that your EKS nodes are granted CloudWatchFullAccess and CloudWatchAgentServerPolicy IAM policies. This Helm chart will configure Fluentbit for log externalisation to CloudWatch.
- - [AWS documentation for setting up Fluentbit](https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/deploy-container-insights-EKS.html)
-
----
-
-## External-dns
-
-This helm chart uses [external-dns](https://github.com/kubernetes-sigs/external-dns) to create 'A' records in AWS Route 53 for our [Docker subdomain feature](https://help.sonatype.com/repomanager3/nexus-repository-administration/formats/docker-registry/docker-subdomain-connector).
-
-See the ```external-dns.alpha.kubernetes.io/hostname``` annotation in the dockerIngress resource in the values.yaml.
-
-### Permissions for external-dns
-
-Open a terminal that has connectivity to your EKS cluster and run the following commands:
-```
-
-cat <<'EOF' >> external-dns-r53-policy.json
-{
- "Version": "2012-10-17",
- "Statement": [
- {
- "Effect": "Allow",
- "Action": [
- "route53:ChangeResourceRecordSets"
- ],
- "Resource": [
- "arn:aws:route53:::hostedzone/*"
- ]
- },
- {
- "Effect": "Allow",
- "Action": [
- "route53:ListHostedZones",
- "route53:ListResourceRecordSets"
- ],
- "Resource": [
- "*"
- ]
- }
- ]
-}
-EOF
-
-
-aws iam create-policy --policy-name "AllowExternalDNSUpdates" --policy-document file://external-dns-r53-policy.json
-
-
-POLICY_ARN=$(aws iam list-policies --query 'Policies[?PolicyName==`AllowExternalDNSUpdates`].Arn' --output text)
-
-
-EKS_CLUSTER_NAME=
-
-
-aws eks describe-cluster --name $EKS_CLUSTER_NAME --query "cluster.identity.oidc.issuer" --output text
-
-
-eksctl utils associate-iam-oidc-provider --cluster $EKS_CLUSTER_NAME --approve
-
-ACCOUNT_ID=$(aws sts get-caller-identity --query "Account" --output text)
-OIDC_PROVIDER=$(aws eks describe-cluster --name $EKS_CLUSTER_NAME --query "cluster.identity.oidc.issuer" --output text | sed -e 's|^https://||')
-```
-
-Note: The value you assign to the 'EXTERNALDNS_NS' variable below should be the same as the one you specify in your values.yaml for namespaces.externaldnsNs
-```
-EXTERNALDNS_NS=nexus-externaldns
-
-cat <<-EOF > externaldns-trust.json
-{
- "Version": "2012-10-17",
- "Statement": [
- {
- "Effect": "Allow",
- "Principal": {
- "Federated": "arn:aws:iam::$ACCOUNT_ID:oidc-provider/$OIDC_PROVIDER"
- },
- "Action": "sts:AssumeRoleWithWebIdentity",
- "Condition": {
- "StringEquals": {
- "$OIDC_PROVIDER:sub": "system:serviceaccount:${EXTERNALDNS_NS}:external-dns",
- "$OIDC_PROVIDER:aud": "sts.amazonaws.com"
- }
- }
- }
- ]
-}
-EOF
-
-IRSA_ROLE="nexusrepo-external-dns-irsa-role"
-aws iam create-role --role-name $IRSA_ROLE --assume-role-policy-document file://externaldns-trust.json
-aws iam attach-role-policy --role-name $IRSA_ROLE --policy-arn $POLICY_ARN
-
-ROLE_ARN=$(aws iam get-role --role-name $IRSA_ROLE --query Role.Arn --output text)
-echo $ROLE_ARN
-```
-
-2. Take note of the ROLE_ARN outputted last above and specify it in your values.yaml for serviceAccount.externaldns.role
-
-## Deployment
-1. Add the sonatype repo to your helm:
-```helm repo add sonatype https://sonatype.github.io/helm3-charts/ ```
-2. Ensure you have updated your values.yaml with appropriate values for your environment.
- - Note that you can specify Ingress annotations via the values.yaml.
- - If you wish to add [Labels](https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/), you can do so via kubectl. See the [kubectl Cheat Sheet](https://kubernetes.io/docs/reference/kubectl/cheatsheet/) for specific commands.
-
-3. Install the chart using the following:
-
-```helm install nxrm sonatype/nxrm-aws-resiliency -f values.yaml```
-
-4. Get the Nexus Repository link using the following:
-
-```kubectl get ingresses -n nexusrepo```
-
----
-
-## Health Check
-You can use the following commands to perform various health checks:
-
-See a list of releases:
-
- ```helm list```
-
- Check pods using the following:
-
- ```kubectl get pods -n nexusrepo```
-
-Check the Nexus Repository logs with the following:
-
- ```kubectl logs -n nexusrepo nxrm-app```
-
-Check if the pod is OK by using the following; you shouldn't see any error/warning messages:
-
- ```kubectl describe pod -n nexusrepo```
-
-Check if ingress is OK using the following:
-
- ```kubectl describe ingress -n nexusrepo```
-
-Check that the Fluent Bit pod is sending events to CloudWatch using the following:
-
- ```kubectl logs -n amazon-cloudwatch ```
-
-If the above returns without error, then check CloudWatch for the ```/aws/containerinsights//nexus-logs``` log group, which should contain four log streams.
-
----
-
-## Uninstall
-To uninstall the deployment, use the following:
-
- ```helm uninstall nxrm```
-
-After removing the deployment, ensure that the namespace is deleted and that Nexus Repository is not listed when using the following:
-
- ```helm list```
+Detailed Help instructions are also available at the following locations:
+* [Single-Node Cloud Resilient Example Using AWS] (https://help.sonatype.com/en/single-node-cloud-resilient-deployment-example-using-aws.html)
+* [Single-Node Cloud Resilient Example Using Azure] (https://help.sonatype.com/en/single-node-cloud-resilient-deployment-example-using-azure.html)
+* [Single Data Center On-Premises Resilient Example Using Kubernetes] (https://help.sonatype.com/en/single-data-center-on-premises-deployment-example-using-kubernetes.html)
+* [High Availability Deployment in AWS] (https://help.sonatype.com/en/option-3---high-availability-deployment-in-amazon-web-services--aws-.html)
+* [High Availability Deployment in Azure] (https://help.sonatype.com/en/option-4---high-availability-deployment-in-azure.html)
+* [On-Premises High Availability Deployment Using Kubernetes] (https://help.sonatype.com/en/option-2---on-premises-high-availability-deployment-using-kubernetes.html)