Эх сурвалжийг харах

Remove some TODOs (#6488)

* Remove stale TODOs

* Ban TODO markers from markdown

* Scrub all TODOs from ci/ and book/
Michael Vines 6 жил өмнө
parent
commit
f80a5b8c34

+ 3 - 3
book/src/proposals/vote-signing-to-implement.md

@@ -17,7 +17,7 @@ The following sections outline how this architecture would work:
 
 
    * The keypair is ephemeral. A new keypair is generated on node bootup. A
    * The keypair is ephemeral. A new keypair is generated on node bootup. A
 
 
-     new keypair might also be generated at runtime based on some TBD
+     new keypair might also be generated at runtime based on some to be determined
 
 
      criteria.
      criteria.
 
 
@@ -28,7 +28,7 @@ The following sections outline how this architecture would work:
      signed by a trusted party
      signed by a trusted party
 3. The stakeholder of the node grants ephemeral key permission to use its stake.
 3. The stakeholder of the node grants ephemeral key permission to use its stake.
 
 
-   This process is TBD.
+   This process is to be determined.
 
 
 4. The node's untrusted, non-enclave software calls trusted enclave software
 4. The node's untrusted, non-enclave software calls trusted enclave software
 
 
@@ -40,7 +40,7 @@ The following sections outline how this architecture would work:
 
 
      presented with some verifiable data to check before signing the vote.
      presented with some verifiable data to check before signing the vote.
 
 
-   * The process of generating the verifiable data in untrusted space is TBD
+   * The process of generating the verifiable data in untrusted space is to be determined
 
 
 ### PoH Verification
 ### PoH Verification
 
 

+ 6 - 6
ci/README.md

@@ -20,7 +20,7 @@ Create a new Azure-based "queue=default" agent by running the following command:
 ```
 ```
 $ az vm create \
 $ az vm create \
    --resource-group ci \
    --resource-group ci \
-   --name XXX \
+   --name XYZ \
    --image boilerplate \
    --image boilerplate \
    --admin-username $(whoami) \
    --admin-username $(whoami) \
    --ssh-key-value ~/.ssh/id_rsa.pub
    --ssh-key-value ~/.ssh/id_rsa.pub
@@ -42,11 +42,11 @@ Creating a "queue=cuda" agent follows the same process but additionally:
 1. When ready, ssh into the instance and start a root shell with `sudo -i`.  Then
 1. When ready, ssh into the instance and start a root shell with `sudo -i`.  Then
    prepare it for deallocation by running:
    prepare it for deallocation by running:
    `waagent -deprovision+user; cd /etc; ln -s ../run/systemd/resolve/stub-resolv.conf resolv.conf`
    `waagent -deprovision+user; cd /etc; ln -s ../run/systemd/resolve/stub-resolv.conf resolv.conf`
-1. Run `az vm deallocate --resource-group ci --name XXX`
-1. Run `az vm generalize --resource-group ci --name XXX`
-1. Run `az image create --resource-group ci --source XXX --name boilerplate`
+1. Run `az vm deallocate --resource-group ci --name XYZ`
+1. Run `az vm generalize --resource-group ci --name XYZ`
+1. Run `az image create --resource-group ci --source XYZ --name boilerplate`
 1. Goto the `ci` resource group in the Azure portal and remove all resources
 1. Goto the `ci` resource group in the Azure portal and remove all resources
-   with the XXX name in them
+   with the XYZ name in them
 
 
 ## Reference
 ## Reference
 
 
@@ -95,7 +95,7 @@ VM Instances in each group is manually adjusted.
 Each Instance group has its own disk image, `ci-default-vX` and
 Each Instance group has its own disk image, `ci-default-vX` and
 `ci-cuda-vY`, where *X* and *Y* are incremented each time the image is changed.
 `ci-cuda-vY`, where *X* and *Y* are incremented each time the image is changed.
 
 
-The process to update a disk image is as follows (TODO: make this less manual):
+The manual process to update a disk image is as follows:
 
 
 1. Create a new VM Instance using the disk image to modify.
 1. Create a new VM Instance using the disk image to modify.
 2. Once the VM boots, ssh to it and modify the disk as desired.
 2. Once the VM boots, ssh to it and modify the disk as desired.

+ 0 - 6
ci/buildkite.yml

@@ -29,12 +29,6 @@ steps:
   - command: ". ci/rust-version.sh; ci/docker-run.sh $$rust_nightly_docker_image ci/test-coverage.sh"
   - command: ". ci/rust-version.sh; ci/docker-run.sh $$rust_nightly_docker_image ci/test-coverage.sh"
     name: "coverage"
     name: "coverage"
     timeout_in_minutes: 40
     timeout_in_minutes: 40
-  # TODO: Fix and re-enable test-large-network.sh
-  # - command: "ci/test-large-network.sh || true"
-  #   name: "large-network [ignored]"
-  #   timeout_in_minutes: 20
-  #   agents:
-  #     - "queue=large"
   - wait
   - wait
   - trigger: "solana-secondary"
   - trigger: "solana-secondary"
     branches: "!pull/*"
     branches: "!pull/*"

+ 1 - 1
ci/nits.sh

@@ -57,7 +57,7 @@ declare useGithubIssueInsteadOf=(
   #'TODO'  # TODO: Uncomment this line to disable TODOs
   #'TODO'  # TODO: Uncomment this line to disable TODOs
 )
 )
 
 
-if _ git --no-pager grep -n --max-depth=0 "${useGithubIssueInsteadOf[@]/#/-e }" -- '*.rs' '*.sh'; then
+if _ git --no-pager grep -n --max-depth=0 "${useGithubIssueInsteadOf[@]/#/-e }" -- '*.rs' '*.sh' '*.md'; then
     exit 1
     exit 1
 fi
 fi
 
 

+ 1 - 1
ci/order-crates-for-publishing.py

@@ -46,7 +46,7 @@ def get_packages():
     max_iterations = pow(len(dependency_graph),2)
     max_iterations = pow(len(dependency_graph),2)
     while dependency_graph:
     while dependency_graph:
         if max_iterations == 0:
         if max_iterations == 0:
-            # TODO: Be more helpful and find the actual cycle for the user
+            # One day be more helpful and find the actual cycle for the user...
             sys.exit('Error: Circular dependency suspected between these packages: {}\n'.format(' '.join(dependency_graph.keys())))
             sys.exit('Error: Circular dependency suspected between these packages: {}\n'.format(' '.join(dependency_graph.keys())))
 
 
         max_iterations -= 1
         max_iterations -= 1

+ 2 - 2
ci/publish-crate.sh

@@ -50,8 +50,8 @@ for Cargo_toml in $Cargo_tomls; do
   (
   (
     set -x
     set -x
     crate=$(dirname "$Cargo_toml")
     crate=$(dirname "$Cargo_toml")
-    # TODO: the rocksdb package does not build with the stock rust docker image,
-    # so use the solana rust docker image until this is resolved upstream
+    # The rocksdb package does not build with the stock rust docker image so use
+    # the solana rust docker image
     cargoCommand="cargo publish --token $CRATES_IO_TOKEN"
     cargoCommand="cargo publish --token $CRATES_IO_TOKEN"
     ci/docker-run.sh "$rust_stable_docker_image" bash -exc "cd $crate; $cargoCommand"
     ci/docker-run.sh "$rust_stable_docker_image" bash -exc "cd $crate; $cargoCommand"
   ) || true # <-- Don't fail.  We want to be able to retry the job in cases when a publish fails halfway due to network/cloud issues
   ) || true # <-- Don't fail.  We want to be able to retry the job in cases when a publish fails halfway due to network/cloud issues

+ 0 - 1
ci/publish-metrics-dashboard.sh

@@ -46,7 +46,6 @@ beta)
   ;;
   ;;
 stable)
 stable)
   # Set to whatever branch 'testnet' is on.
   # Set to whatever branch 'testnet' is on.
-  # TODO: Revert to $STABLE_CHANNEL for TdS
   CHANNEL_BRANCH=$BETA_CHANNEL
   CHANNEL_BRANCH=$BETA_CHANNEL
   ;;
   ;;
 *)
 *)

+ 2 - 2
ci/test-bench.sh

@@ -70,9 +70,9 @@ _ cargo +$rust_nightly bench --manifest-path programs/bpf/Cargo.toml ${V:+--verb
 # Run banking bench. Doesn't require nightly, but use since it is already built.
 # Run banking bench. Doesn't require nightly, but use since it is already built.
 _ cargo +$rust_nightly run --release --manifest-path banking_bench/Cargo.toml ${V:+--verbose} | tee -a "$BENCH_FILE"
 _ cargo +$rust_nightly run --release --manifest-path banking_bench/Cargo.toml ${V:+--verbose} | tee -a "$BENCH_FILE"
 
 
-# TODO: debug why solana-upload-perf takes over 30 minutes to complete.
+# `solana-upload-perf` disabled as it can take over 30 minutes to complete for some
+# reason
 exit 0
 exit 0
-
 _ cargo +$rust_nightly run --release --package solana-upload-perf \
 _ cargo +$rust_nightly run --release --package solana-upload-perf \
   -- "$BENCH_FILE" "$TARGET_BRANCH" "$UPLOAD_METRICS" | tee "$BENCH_ARTIFACT"
   -- "$BENCH_FILE" "$TARGET_BRANCH" "$UPLOAD_METRICS" | tee "$BENCH_ARTIFACT"
 
 

+ 0 - 41
ci/test-large-network.sh

@@ -1,41 +0,0 @@
-#!/usr/bin/env bash
-set -e
-
-here=$(dirname "$0")
-cd "$here"/..
-
-source ci/rust-version.sh stable
-
-export RUST_BACKTRACE=1
-
-rm -rf target/perf-libs
-./fetch-perf-libs.sh
-export LD_LIBRARY_PATH=$PWD/target/perf-libs:$LD_LIBRARY_PATH
-
-export RUST_LOG=multinode=info
-
-source scripts/ulimit-n.sh
-
-if [[ $(sysctl -n net.core.rmem_default) -lt 1610612736 ]]; then
-  echo 'Error: rmem_default too small, run "sudo sysctl -w net.core.rmem_default=1610612736" to continue'
-  exit 1
-fi
-
-if [[ $(sysctl -n net.core.rmem_max) -lt 1610612736 ]]; then
-  echo 'Error: rmem_max too small, run "sudo sysctl -w net.core.rmem_max=1610612736" to continue'
-  exit 1
-fi
-
-if [[ $(sysctl -n net.core.wmem_default) -lt 1610612736 ]]; then
-  echo 'Error: rmem_default too small, run "sudo sysctl -w net.core.wmem_default=1610612736" to continue'
-  exit 1
-fi
-
-if [[ $(sysctl -n net.core.wmem_max) -lt 1610612736 ]]; then
-  echo 'Error: rmem_max too small, run "sudo sysctl -w net.core.wmem_max=1610612736" to continue'
-  exit 1
-fi
-
-set -x
-export SOLANA_DYNAMIC_NODES=120
-exec cargo +"$rust_stable" test --release --features=erasure test_multi_node_dynamic_network -- --ignored

+ 1 - 14
ci/testnet-manager.sh

@@ -636,20 +636,7 @@ sanity-or-restart)
   else
   else
     echo "+++ Sanity failed, updating the network"
     echo "+++ Sanity failed, updating the network"
     $metricsWriteDatapoint "testnet-manager sanity-failure=1"
     $metricsWriteDatapoint "testnet-manager sanity-failure=1"
-
-    # TODO: Restore attempt to restart the cluster before recreating it
-    #       See https://github.com/solana-labs/solana/issues/3774
-    if false; then
-      if start; then
-        echo Update successful
-      else
-        echo "+++ Update failed, restarting the network"
-        $metricsWriteDatapoint "testnet-manager update-failure=1"
-        create-and-start
-      fi
-    else
-      create-and-start
-    fi
+    create-and-start
   fi
   fi
   ;;
   ;;
 *)
 *)

+ 1 - 3
multinode-demo/delegate-stake.sh

@@ -88,14 +88,12 @@ if [[ ! -f $vote_keypair_path ]]; then
 fi
 fi
 
 
 if [[ -f $stake_keypair_path ]]; then
 if [[ -f $stake_keypair_path ]]; then
-  # TODO: Add ability to add multiple stakes with this script?
   echo "Error: $stake_keypair_path already exists"
   echo "Error: $stake_keypair_path already exists"
   exit 1
   exit 1
 fi
 fi
 
 
 if ((airdrops_enabled)); then
 if ((airdrops_enabled)); then
-  declare fees=100 # TODO: No hardcoded transaction fees, fetch the current cluster fees
-  $solana_cli "${common_args[@]}" airdrop $((stake_lamports+fees)) lamports
+  $solana_cli "${common_args[@]}" airdrop "$stake_lamports" lamports
 fi
 fi
 
 
 $solana_keygen new -o "$stake_keypair_path"
 $solana_keygen new -o "$stake_keypair_path"

+ 3 - 3
net/remote/remote-node.sh

@@ -377,9 +377,9 @@ EOF
     )
     )
 
 
     if [[ $airdropsEnabled != true ]]; then
     if [[ $airdropsEnabled != true ]]; then
-      echo "TODO: archivers not supported without airdrops"
-      # TODO: need to provide the `--identity` argument to an existing system
-      #       account with lamports in it
+      # If this ever becomes a problem, we need to provide the `--identity`
+      # argument to an existing system account with lamports in it
+      echo "Error: archivers not supported without airdrops"
       exit 1
       exit 1
     fi
     fi
 
 

+ 1 - 5
net/remote/remote-sanity.sh

@@ -66,7 +66,6 @@ local|tar|skip)
   solana_cli=solana
   solana_cli=solana
   solana_gossip=solana-gossip
   solana_gossip=solana-gossip
   solana_install=solana-install
   solana_install=solana-install
-  solana_keygen=solana-keygen
   ;;
   ;;
 *)
 *)
   echo "Unknown deployment method: $deployMethod"
   echo "Unknown deployment method: $deployMethod"
@@ -85,11 +84,8 @@ fi
 
 
 echo "+++ $sanityTargetIp: validators"
 echo "+++ $sanityTargetIp: validators"
 (
 (
-  # Ensure solana-cli has a keypair even though it doesn't really need one...
-  # TODO: Remove when https://github.com/solana-labs/solana/issues/6375 is fixed
-  $solana_keygen new --force -o temp-id.json
   set -x
   set -x
-  $solana_cli --keypair temp-id.json --url http://"$sanityTargetIp":8899 show-validators
+  $solana_cli --url http://"$sanityTargetIp":8899 show-validators
 )
 )
 
 
 echo "+++ $sanityTargetIp: node count ($numSanityNodes expected)"
 echo "+++ $sanityTargetIp: node count ($numSanityNodes expected)"

+ 1 - 1
net/scripts/azure-provider.sh

@@ -126,7 +126,7 @@ cloud_Initialize() {
   declare networkName="$1"
   declare networkName="$1"
   # ec2-provider.sh creates firewall rules programmatically, should do the same
   # ec2-provider.sh creates firewall rules programmatically, should do the same
   # here.
   # here.
-  echo "TODO: create $networkName firewall rules programmatically instead of assuming the 'testnet' tag exists"
+  echo "Note: one day create $networkName firewall rules programmatically instead of assuming the 'testnet' tag exists"
 }
 }
 
 
 #
 #

+ 2 - 2
net/scripts/disable-background-upgrades.sh

@@ -3,8 +3,8 @@ set -ex
 #
 #
 # Prevent background upgrades that block |apt-get|
 # Prevent background upgrades that block |apt-get|
 #
 #
-# TODO: This approach is pretty uncompromising.  An alternative solution that
-#       doesn't involve deleting system files would be welcome.
+# This approach is pretty uncompromising.  An alternative solution that doesn't
+# involve deleting system files would be welcome.
 
 
 [[ $(uname) = Linux ]] || exit 1
 [[ $(uname) = Linux ]] || exit 1
 [[ $USER = root ]] || exit 1
 [[ $USER = root ]] || exit 1

+ 3 - 3
net/scripts/gce-provider.sh

@@ -122,7 +122,7 @@ cloud_Initialize() {
   declare networkName="$1"
   declare networkName="$1"
   # ec2-provider.sh creates firewall rules programmatically, should do the same
   # ec2-provider.sh creates firewall rules programmatically, should do the same
   # here.
   # here.
-  echo "TODO: create $networkName firewall rules programmatically instead of assuming the 'testnet' tag exists"
+  echo "Note: one day create $networkName firewall rules programmatically instead of assuming the 'testnet' tag exists"
 }
 }
 
 
 #
 #
@@ -170,8 +170,8 @@ cloud_CreateInstances() {
   if $enableGpu; then
   if $enableGpu; then
     # Custom Ubuntu 18.04 LTS image with CUDA 9.2 and CUDA 10.0 installed
     # Custom Ubuntu 18.04 LTS image with CUDA 9.2 and CUDA 10.0 installed
     #
     #
-    # TODO: Unfortunately this image is not public.  When this becomes an issue,
-    # use the stock Ubuntu 18.04 image and programmatically install CUDA after the
+    # Unfortunately this image is not public.  When this becomes an issue, use
+    # the stock Ubuntu 18.04 image and programmatically install CUDA after the
     # instance boots
     # instance boots
     #
     #
     imageName="ubuntu-1804-bionic-v20181029-with-cuda-10-and-cuda-9-2"
     imageName="ubuntu-1804-bionic-v20181029-with-cuda-10-and-cuda-9-2"