Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update libs dependency #2764

Open
wants to merge 1 commit into
base: main
Choose a base branch
from
Open

Update libs dependency #2764

wants to merge 1 commit into from

Conversation

vdaas-ci
Copy link
Collaborator

@vdaas-ci vdaas-ci commented Dec 4, 2024

Automated pull request to update Dependencies.

Summary by CodeRabbit

  • New Features

    • Updated version numbers for multiple software components including Go, Rust, Kubernetes, and Helm across various templates and configuration files.
    • Enhanced scheduling capabilities in Kubernetes CronJobs with the addition of affinity configurations.
  • Bug Fixes

    • Updated actions and dependencies to newer versions, potentially addressing bugs and improving functionality.
  • Documentation

    • Revised issue and pull request templates to reflect updated software versions.
  • Chores

    • General maintenance updates including newlines added for proper file formatting in multiple Dockerfiles.

Copy link

cloudflare-workers-and-pages bot commented Dec 4, 2024

Deploying vald with  Cloudflare Pages  Cloudflare Pages

Latest commit: 7fab0e3
Status: ✅  Deploy successful!
Preview URL: https://bbb81500.vald.pages.dev
Branch Preview URL: https://create-pull-request-patch.vald.pages.dev

View logs

Copy link
Contributor

coderabbitai bot commented Dec 4, 2024

📝 Walkthrough

Walkthrough

This pull request primarily updates version numbers across various files related to software components and dependencies. It modifies the .github/ISSUE_TEMPLATE files to reflect new versions for Go, Rust, Kubernetes, and Helm. Additionally, several Dockerfiles are adjusted to include newlines for formatting, while some also remove error-checking comments. The go.mod files are updated to reflect new dependency versions, and Kubernetes configuration files are enhanced with affinity settings for job scheduling.

Changes

File Path Change Summary
.github/ISSUE_TEMPLATE/bug_report.md Updated Go (v1.23.3 to v1.23.4), Rust (v1.82.0 to v1.83.0), Kubernetes (v1.31.2 to v1.31.3), Helm (v3.16.2 to v3.16.3) versions.
.github/ISSUE_TEMPLATE/security_issue_report.md Updated Go (v1.23.3 to v1.23.4), Rust (v1.82.0 to v1.83.0), Kubernetes (v1.31.2 to v1.31.3), Helm (v3.16.2 to v3.16.3) versions.
.github/PULL_REQUEST_TEMPLATE.md Updated Go (v1.23.3 to v1.23.4), Rust (v1.82.0 to v1.83.0), Kubernetes (v1.31.2 to v1.31.3), Helm (v3.16.2 to v3.16.3) versions.
.github/workflows/coverage.yaml Updated Codecov action from v4 to v5.
dockers/*/Dockerfile Added newline at end of various Dockerfiles (e.g., agent, faiss, ngt, sidecar, etc.). Removed error-checking comments in some Dockerfiles (e.g., binfmt, buildkit).
example/client/go.mod Updated dependency versions for google.golang.org/genproto and google.golang.org/protobuf.
go.mod Updated various dependency versions (e.g., cloud.google.com/go/compute, github.com/stretchr/testify, etc.).
k8s/index/job/correction/cronjob.yaml Added affinity section to CronJob configuration.
k8s/index/job/creation/cronjob.yaml Added affinity section to CronJob configuration.
k8s/index/job/deletion/configmap.yaml Removed ConfigMap vald-index-deletion-config.
k8s/index/job/deletion/cronjob.yaml Removed CronJob vald-index-deletion.
k8s/index/job/save/cronjob.yaml Added affinity section to CronJob configuration.
k8s/index/operator/configmap.yaml Updated config.yaml structure in ConfigMap.
k8s/index/operator/deployment.yaml Updated checksum/configmap annotation.
k8s/operator/helm/crds/valdrelease.yaml Added affinity, nodeSelector, and tolerations properties to various sections.
rust/rust-toolchain Updated Rust toolchain version from 1.82.0 to 1.83.0.
versions/* Updated various version files (e.g., BUF_VERSION, CMAKE_VERSION, GO_VERSION, etc.) to new version numbers.

Possibly related PRs

Suggested labels

size/S, area/internal

Suggested reviewers

  • kmrmt
  • vankichi

Thank you for using CodeRabbit. We offer it for free to the OSS community and would appreciate your support in helping us grow. If you find it useful, would you consider giving us a shout-out on your favorite social media?

❤️ Share
🪧 Tips

Chat

There are 3 ways to chat with CodeRabbit:

  • Review comments: Directly reply to a review comment made by CodeRabbit. Example:
    • I pushed a fix in commit <commit_id>, please review it.
    • Generate unit testing code for this file.
    • Open a follow-up GitHub issue for this discussion.
  • Files and specific lines of code (under the "Files changed" tab): Tag @coderabbitai in a new review comment at the desired location with your query. Examples:
    • @coderabbitai generate unit testing code for this file.
    • @coderabbitai modularize this function.
  • PR comments: Tag @coderabbitai in a new PR comment to ask questions about the PR branch. For the best results, please provide a very specific query, as very limited context is provided in this mode. Examples:
    • @coderabbitai gather interesting stats about this repository and render them as a table. Additionally, render a pie chart showing the language distribution in the codebase.
    • @coderabbitai read src/utils.ts and generate unit testing code.
    • @coderabbitai read the files in the src/scheduler package and generate a class diagram using mermaid and a README in the markdown format.
    • @coderabbitai help me debug CodeRabbit configuration file.

Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments.

CodeRabbit Commands (Invoked using PR comments)

  • @coderabbitai pause to pause the reviews on a PR.
  • @coderabbitai resume to resume the paused reviews.
  • @coderabbitai review to trigger an incremental review. This is useful when automatic reviews are disabled for the repository.
  • @coderabbitai full review to do a full review from scratch and review all the files again.
  • @coderabbitai summary to regenerate the summary of the PR.
  • @coderabbitai resolve resolve all the CodeRabbit review comments.
  • @coderabbitai configuration to show the current CodeRabbit configuration for the repository.
  • @coderabbitai help to get help.

Other keywords and placeholders

  • Add @coderabbitai ignore anywhere in the PR description to prevent this PR from being reviewed.
  • Add @coderabbitai summary to generate the high-level summary at a specific location in the PR description.
  • Add @coderabbitai anywhere in the PR title to generate the title automatically.

CodeRabbit Configuration File (.coderabbit.yaml)

  • You can programmatically configure CodeRabbit by adding a .coderabbit.yaml file to the root of your repository.
  • Please see the configuration documentation for more information.
  • If your editor has YAML language server enabled, you can add the path at the top of this file to enable auto-completion and validation: # yaml-language-server: $schema=https://coderabbit.ai/integrations/schema.v2.json

Documentation and Community

  • Visit our Documentation for detailed information on how to use CodeRabbit.
  • Join our Discord Community to get help, request features, and share feedback.
  • Follow us on X/Twitter for updates and announcements.

USER root:root
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ [hadolint] <DL3002> reported by reviewdog 🐶
Last USER should not be root

@vdaas-ci
Copy link
Collaborator Author

vdaas-ci commented Dec 4, 2024

[CHATOPS:HELP] ChatOps commands.

  • 🙆‍♀️ /approve - approve
  • 🍱 /format - format codes and add licenses
  • /gen-test - generate test codes
  • 🏷️ /label - add labels
  • 🔚 2️⃣ 🔚 /label actions/e2e-deploy - run E2E deploy & integration test

Copy link

codecov bot commented Dec 4, 2024

Codecov Report

All modified and coverable lines are covered by tests ✅

Project coverage is 23.86%. Comparing base (630a835) to head (7fab0e3).
Report is 5 commits behind head on main.

Additional details and impacted files
@@            Coverage Diff             @@
##             main    #2764      +/-   ##
==========================================
- Coverage   23.87%   23.86%   -0.01%     
==========================================
  Files         547      547              
  Lines       54660    54660              
==========================================
- Hits        13051    13047       -4     
- Misses      40823    40827       +4     
  Partials      786      786              

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

Copy link
Contributor

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 2

🧹 Outside diff range and nitpick comments (5)
k8s/index/job/save/cronjob.yaml (1)

78-86: Consider defining specific affinity rules.

The added affinity section is currently empty. Consider defining specific rules to:

  1. Ensure optimal pod placement
  2. Prevent resource contention
  3. Improve reliability

Example rules to consider:

  • Spread pods across nodes
  • Avoid co-location with resource-intensive pods
  • Ensure high availability

Here's a suggested implementation:

 affinity:
   nodeAffinity:
-    preferredDuringSchedulingIgnoredDuringExecution: []
+    preferredDuringSchedulingIgnoredDuringExecution:
+    - weight: 100
+      preference:
+        matchExpressions:
+        - key: node-type
+          operator: In
+          values:
+          - high-memory
   podAntiAffinity:
-    preferredDuringSchedulingIgnoredDuringExecution: []
+    preferredDuringSchedulingIgnoredDuringExecution:
+    - weight: 100
+      podAffinityTerm:
+        labelSelector:
+          matchExpressions:
+          - key: app
+            operator: In
+            values:
+            - vald-index-save
+        topologyKey: kubernetes.io/hostname
k8s/index/job/correction/cronjob.yaml (1)

78-86: Consider removing empty affinity configuration or adding meaningful rules

The current affinity configuration contains empty arrays for all affinity types (nodeAffinity, podAffinity, podAntiAffinity). While this provides a structure for future use, empty rules increase configuration verbosity without adding functionality.

Either:

  1. Add meaningful affinity rules to influence pod scheduling, or
  2. Remove the empty configuration until specific scheduling requirements are needed
k8s/index/operator/configmap.yaml (1)

Line range hint 8409-8453: Consider documenting the purpose of empty affinity configurations

Empty affinity rules have been added systematically across multiple job templates. While this provides consistent structure, it increases configuration complexity without current functional impact.

Consider:

  1. Adding documentation explaining the intended use of these affinity settings
  2. Providing example configurations in comments to guide users
  3. Moving the empty affinity template to a shared/reusable configuration if possible

Also applies to: 9085-9087, 13121-13165

k8s/operator/helm/crds/valdrelease.yaml (1)

Line range hint 78-14284: Consider providing affinity configuration examples and documentation

The systematic addition of affinity capabilities across the codebase suggests preparation for advanced scheduling requirements. To make this feature more accessible:

  1. Consider adding a documentation section explaining:
    • Common affinity use cases in the context of Vald
    • Example configurations for typical scenarios
    • Best practices for using affinity rules
  2. Consider providing a default affinity configuration template that users can reference
  3. If possible, implement helper functions/operators to simplify affinity configuration for common use cases
go.mod (1)

Line range hint 3-3: Invalid Go version specified

The specified Go version 1.23.3 is invalid. The latest stable version of Go is 1.22.x.

Apply this diff to update to a valid Go version:

-go 1.23.3
+go 1.22.1
📜 Review details

Configuration used: CodeRabbit UI
Review profile: CHILL

📥 Commits

Reviewing files that changed from the base of the PR and between 9bd9034 and 7fab0e3.

⛔ Files ignored due to path filters (22)
  • apis/grpc/v1/agent/core/agent.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/agent/sidecar/sidecar.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/agent/sidecar/sidecar_vtproto.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/discoverer/discoverer.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/filter/egress/egress_filter.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/filter/ingress/ingress_filter.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/meta/meta.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/mirror/mirror.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/payload/payload.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/rpc/errdetails/error_details.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/filter.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/flush.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/index.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/insert.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/object.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/remove.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/search.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/update.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/upsert.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • example/client/go.sum is excluded by !**/*.sum
  • go.sum is excluded by !**/*.sum
  • rust/Cargo.lock is excluded by !**/*.lock
📒 Files selected for processing (61)
  • .github/ISSUE_TEMPLATE/bug_report.md (1 hunks)
  • .github/ISSUE_TEMPLATE/security_issue_report.md (1 hunks)
  • .github/PULL_REQUEST_TEMPLATE.md (1 hunks)
  • .github/workflows/coverage.yaml (1 hunks)
  • dockers/agent/core/agent/Dockerfile (1 hunks)
  • dockers/agent/core/faiss/Dockerfile (1 hunks)
  • dockers/agent/core/ngt/Dockerfile (1 hunks)
  • dockers/agent/sidecar/Dockerfile (1 hunks)
  • dockers/binfmt/Dockerfile (1 hunks)
  • dockers/buildbase/Dockerfile (1 hunks)
  • dockers/buildkit/Dockerfile (1 hunks)
  • dockers/buildkit/syft/scanner/Dockerfile (1 hunks)
  • dockers/ci/base/Dockerfile (1 hunks)
  • dockers/dev/Dockerfile (1 hunks)
  • dockers/discoverer/k8s/Dockerfile (1 hunks)
  • dockers/example/client/Dockerfile (1 hunks)
  • dockers/gateway/filter/Dockerfile (1 hunks)
  • dockers/gateway/lb/Dockerfile (1 hunks)
  • dockers/gateway/mirror/Dockerfile (1 hunks)
  • dockers/index/job/correction/Dockerfile (1 hunks)
  • dockers/index/job/creation/Dockerfile (1 hunks)
  • dockers/index/job/deletion/Dockerfile (1 hunks)
  • dockers/index/job/readreplica/rotate/Dockerfile (1 hunks)
  • dockers/index/job/save/Dockerfile (1 hunks)
  • dockers/index/operator/Dockerfile (1 hunks)
  • dockers/manager/index/Dockerfile (1 hunks)
  • dockers/operator/helm/Dockerfile (1 hunks)
  • dockers/tools/benchmark/job/Dockerfile (1 hunks)
  • dockers/tools/benchmark/operator/Dockerfile (1 hunks)
  • dockers/tools/cli/loadtest/Dockerfile (1 hunks)
  • example/client/go.mod (2 hunks)
  • go.mod (20 hunks)
  • k8s/index/job/correction/cronjob.yaml (1 hunks)
  • k8s/index/job/creation/cronjob.yaml (1 hunks)
  • k8s/index/job/deletion/configmap.yaml (0 hunks)
  • k8s/index/job/deletion/cronjob.yaml (0 hunks)
  • k8s/index/job/save/cronjob.yaml (1 hunks)
  • k8s/index/operator/configmap.yaml (1 hunks)
  • k8s/index/operator/deployment.yaml (1 hunks)
  • k8s/operator/helm/crds/valdrelease.yaml (9 hunks)
  • rust/rust-toolchain (1 hunks)
  • versions/BUF_VERSION (1 hunks)
  • versions/CMAKE_VERSION (1 hunks)
  • versions/GOLANGCILINT_VERSION (1 hunks)
  • versions/GO_VERSION (1 hunks)
  • versions/HELM_VERSION (1 hunks)
  • versions/KUBECTL_VERSION (1 hunks)
  • versions/OPERATOR_SDK_VERSION (1 hunks)
  • versions/PROMETHEUS_STACK_VERSION (1 hunks)
  • versions/PROTOBUF_VERSION (1 hunks)
  • versions/RUST_VERSION (1 hunks)
  • versions/TELEPRESENCE_VERSION (1 hunks)
  • versions/USEARCH_VERSION (1 hunks)
  • versions/YQ_VERSION (1 hunks)
  • versions/actions/CODECOV_CODECOV_ACTION (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_ANALYZE (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_AUTOBUILD (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_INIT (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_UPLOAD_SARIF (1 hunks)
  • versions/actions/GITHUB_ISSUE_METRICS (1 hunks)
  • versions/actions/REVIEWDOG_ACTION_HADOLINT (1 hunks)
💤 Files with no reviewable changes (2)
  • k8s/index/job/deletion/configmap.yaml
  • k8s/index/job/deletion/cronjob.yaml
✅ Files skipped from review due to trivial changes (51)
  • versions/BUF_VERSION
  • versions/GO_VERSION
  • versions/CMAKE_VERSION
  • versions/USEARCH_VERSION
  • versions/RUST_VERSION
  • versions/TELEPRESENCE_VERSION
  • versions/OPERATOR_SDK_VERSION
  • versions/HELM_VERSION
  • versions/PROTOBUF_VERSION
  • versions/actions/GITHUB_CODEQL_ACTION_AUTOBUILD
  • versions/actions/GITHUB_CODEQL_ACTION_ANALYZE
  • versions/actions/GITHUB_CODEQL_ACTION_INIT
  • versions/KUBECTL_VERSION
  • versions/actions/GITHUB_CODEQL_ACTION_UPLOAD_SARIF
  • versions/YQ_VERSION
  • dockers/buildkit/Dockerfile
  • versions/PROMETHEUS_STACK_VERSION
  • versions/actions/CODECOV_CODECOV_ACTION
  • versions/actions/REVIEWDOG_ACTION_HADOLINT
  • dockers/buildbase/Dockerfile
  • dockers/binfmt/Dockerfile
  • dockers/buildkit/syft/scanner/Dockerfile
  • rust/rust-toolchain
  • versions/actions/GITHUB_ISSUE_METRICS
  • dockers/agent/core/agent/Dockerfile
  • dockers/tools/cli/loadtest/Dockerfile
  • dockers/index/job/correction/Dockerfile
  • dockers/agent/core/ngt/Dockerfile
  • dockers/agent/core/faiss/Dockerfile
  • dockers/ci/base/Dockerfile
  • versions/GOLANGCILINT_VERSION
  • dockers/index/operator/Dockerfile
  • dockers/index/job/save/Dockerfile
  • dockers/index/job/deletion/Dockerfile
  • dockers/manager/index/Dockerfile
  • dockers/operator/helm/Dockerfile
  • dockers/gateway/filter/Dockerfile
  • dockers/example/client/Dockerfile
  • dockers/tools/benchmark/job/Dockerfile
  • dockers/dev/Dockerfile
  • dockers/gateway/mirror/Dockerfile
  • dockers/gateway/lb/Dockerfile
  • .github/ISSUE_TEMPLATE/bug_report.md
  • .github/ISSUE_TEMPLATE/security_issue_report.md
  • dockers/discoverer/k8s/Dockerfile
  • dockers/index/job/readreplica/rotate/Dockerfile
  • dockers/index/job/creation/Dockerfile
  • dockers/tools/benchmark/operator/Dockerfile
  • dockers/agent/sidecar/Dockerfile
  • .github/PULL_REQUEST_TEMPLATE.md
  • example/client/go.mod
🔇 Additional comments (7)
k8s/index/operator/deployment.yaml (1)

49-49: ConfigMap checksum update looks correct.

The checksum update indicates changes in the associated ConfigMap. This will trigger a rolling update of the deployment, which is expected.

Let's verify the ConfigMap changes that triggered this checksum update:

✅ Verification successful

Let me generate additional verification scripts to check the changes in the ConfigMap.


Let me check the changes in the other ConfigMap file to ensure all related changes are verified.


ConfigMap changes verified - only security enhancements added

After reviewing the changes in the ConfigMap, the update adds security context configurations to the job templates, which is a security enhancement. The checksum update in the deployment is correctly reflecting these changes. The changes include:

  • Added security context for containers with non-root user settings
  • Added pod-level security context configurations
  • No functional changes to the core configuration
🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Verify the ConfigMap changes
# Compare the ConfigMap content with its previous version

# Get the ConfigMap content
cat k8s/index/operator/configmap.yaml

# Check git history for the ConfigMap changes
git log -p k8s/index/operator/configmap.yaml

Length of output: 27881


Script:

#!/bin/bash
# Get the git diff of the ConfigMap to see what changed
git diff HEAD~1 k8s/index/operator/configmap.yaml

# Check if there are any other related files changed in this PR
git diff --name-only HEAD~1 HEAD | grep -i configmap

Length of output: 52934


Script:

#!/bin/bash
# Check changes in the deletion job's ConfigMap
git diff HEAD~1 k8s/index/job/deletion/configmap.yaml

Length of output: 273

k8s/index/job/creation/cronjob.yaml (2)

78-86: Consider defining specific affinity rules.

Similar to the save job, the affinity section should have specific rules defined for optimal scheduling.


Line range hint 30-30: Review the schedule frequency.

The current schedule "* * * * *" runs the job every minute, which might be:

  1. Resource intensive
  2. Unnecessary for index creation
  3. Potentially causing job queue buildup due to concurrencyPolicy: Forbid

Consider adjusting the schedule based on your actual index creation needs.

Let's check for potential job congestion:

k8s/operator/helm/crds/valdrelease.yaml (1)

Line range hint 8409-14284: LGTM: Well-structured affinity schema definition

The CRD schema for affinity settings is thorough and follows Kubernetes API conventions:

  • Properly defines all standard Kubernetes affinity fields
  • Uses x-kubernetes-preserve-unknown-fields appropriately for complex objects
  • Maintains consistency with Kubernetes API structure
go.mod (3)

8-8: Verify changelog for cloud provider dependency updates

Multiple cloud provider dependencies have been updated. Please verify the changelog for any breaking changes:

  • cloud.google.com/go/compute: v1.28.3 → v1.29.0
  • cloud.google.com/go/pubsub: v1.45.1 → v1.45.2
  • cloud.google.com/go/storage: v1.46.0 → v1.47.0
  • github.com/aws/aws-sdk-go-v2: v1.32.4 → v1.32.6

Run the following script to check the release notes:

Also applies to: 14-14, 16-16, 48-48

✅ Verification successful

Cloud provider dependency updates are safe to proceed

Based on the changelog verification:

  • cloud.google.com/go/compute v1.29.0: Only adds new features with API updates, no breaking changes
  • cloud.google.com/go/pubsub v1.45.2: Contains only bug fixes and documentation updates
  • cloud.google.com/go/storage v1.47.0: Adds a new feature and a minor auth dependency bump, no breaking changes

Let me verify the AWS SDK changelog with a different approach:


Cloud provider dependency updates verified as safe

The Google Cloud dependencies (compute, pubsub, storage) show only minor version increments with non-breaking changes:

  • Feature additions in compute and storage
  • Bug fixes and documentation updates in pubsub
  • No breaking changes reported in any changelog

For AWS SDK update (v1.32.4 → v1.32.6), being a patch version increment (x.y.Z), it should only contain backwards-compatible bug fixes as per semantic versioning.

🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Check release notes for breaking changes

# Check Google Cloud compute changelog
gh api repos/googleapis/google-cloud-go/releases/tags/compute/v1.29.0 --jq .body

# Check Google Cloud pubsub changelog
gh api repos/googleapis/google-cloud-go/releases/tags/pubsub/v1.45.2 --jq .body

# Check Google Cloud storage changelog
gh api repos/googleapis/google-cloud-go/releases/tags/storage/v1.47.0 --jq .body

# Check AWS SDK changelog
gh api repos/aws/aws-sdk-go-v2/releases/tags/v1.32.6 --jq .body

Length of output: 2531


Script:

#!/bin/bash
# Try alternative endpoint for AWS SDK changelog
gh api repos/aws/aws-sdk-go-v2/releases --jq '.[0] | select(.tag_name=="v1.32.6") | .body'

Length of output: 92


228-228: Verify test compatibility with updated testing frameworks

Testing frameworks have been updated:

  • github.com/stretchr/testify: v1.9.0 → v1.10.0
  • github.com/onsi/ginkgo/v2: v2.21.0 → v2.22.0

Please ensure all tests pass with the new versions.

Run the following script to check for any test-related issues:

Also applies to: 268-268

✅ Verification successful

No compatibility concerns with updated testing frameworks

Based on the analysis of test files:

  • All tests use testify for assertions via require package
  • No usage of deprecated test patterns or test suites
  • No direct usage of Ginkgo found in the codebase
  • Simple test patterns are used (standard Go tests with testify assertions)

The minor version updates in both frameworks are backward compatible and the codebase uses stable patterns that are unlikely to be affected.

🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Search for test files and potential compatibility issues

# Find all test files
echo "Searching for test files..."
fd -e go -E vendor --exec grep -l "github.com/stretchr/testify\|github.com/onsi/ginkgo"

# Check for deprecated test patterns
echo "Checking for deprecated test patterns..."
rg -l "suite\.Run|suite\.SetupTest|BeforeSuite|AfterSuite" --type go

Length of output: 797


Script:

#!/bin/bash
# Let's check the actual test patterns being used in these files

# Check testify usage patterns
echo "=== Checking testify usage patterns ==="
rg "github.com/stretchr/testify" -A 5 -B 5 --type go

# Check ginkgo usage patterns
echo "=== Checking ginkgo usage patterns ==="
rg "github.com/onsi/ginkgo" -A 5 -B 5 --type go

# Check test suite implementations
echo "=== Checking test suite implementations ==="
ast-grep --pattern 'type $_ struct {
    $$$
    suite.Suite
    $$$
}'

Length of output: 9465


324-324: Verify API compatibility with updated core dependencies

Core infrastructure dependencies have been updated:

  • google.golang.org/protobuf: v1.35.1 → v1.35.2
  • google.golang.org/grpc: Updated to v1.68.0

These updates could affect API compatibility.

Run the following script to check for potential API compatibility issues:

Also applies to: 412-412

✅ Verification successful

Let me gather more specific information about the API changes by checking the proto definitions and their implementations.


Based on the analysis of the shell script results, I can now provide a final response regarding the API compatibility concerns:

No API compatibility issues with protobuf and gRPC updates

The version updates are minor:

  • protobuf: v1.35.1 → v1.35.2 (patch version)
  • gRPC: Updated to v1.68.0 (maintains backward compatibility)

The codebase inspection shows:

  • All protobuf usage follows standard patterns through generated code
  • No deprecated API usage found
  • No breaking changes in service definitions
  • Core message/service structures remain unchanged
🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Check for potential API compatibility issues

# Find all proto service definitions
echo "Searching for proto service definitions..."
fd -e proto --exec grep -l "service"

# Find all gRPC service implementations
echo "Checking gRPC service implementations..."
ast-grep --pattern 'type $service struct {
  $$$
  grpc.$Server
  $$$
}'

Length of output: 987


Script:

#!/bin/bash
# Check for specific protobuf and gRPC usage patterns

# Look for gRPC service implementations in Go files
echo "Searching for gRPC service implementations..."
rg "google.golang.org/grpc" -A 3 -g "*.go"

# Check proto service definitions for potential breaking changes
echo -e "\nChecking proto service definitions..."
for proto in $(fd -e proto); do
  echo -e "\n=== $proto ==="
  grep -A 3 "^service\|^rpc\|^message" "$proto"
done

# Look for protobuf specific API usage
echo -e "\nChecking protobuf API usage..."
rg "proto\." -g "*.go" -A 2

Length of output: 101583

@@ -54,7 +54,7 @@ jobs:
run: |
make coverage
- name: Upload coverage report to Codecov
uses: codecov/codecov-action@v4
uses: codecov/codecov-action@v5
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codebase verification

⚠️ Update required: Codecov Action v5 configuration needs changes

The workflow needs to be updated to align with v5's breaking changes:

  • The file parameter is deprecated in favor of files
  • Current workflow uses the deprecated file parameter

Required changes:

  • Replace file: ./coverage.out with files: ./coverage.out
🔗 Analysis chain

Verify Codecov Action v5 compatibility

The update from v4 to v5 aligns with the PR objective of updating dependencies. However, as this is a major version bump, please ensure compatibility.

Let's verify the Codecov Action v5 requirements and changes:

🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Check Codecov Action v5 release notes and requirements

# Get the latest release information
gh api \
  -H "Accept: application/vnd.github+json" \
  repos/codecov/codecov-action/releases | \
  jq -r '.[] | select(.tag_name == "v5.0.7") | .body'

# Check if the workflow uses any deprecated inputs
rg -l 'codecov/codecov-action@v[0-9]' | \
  xargs rg 'uses: codecov/codecov-action@v[0-9]' -A 5 -B 5

Length of output: 556


Script:

#!/bin/bash
# Let's check the changelog between v4 and v5 by getting the v5.0.0 release notes
gh api \
  -H "Accept: application/vnd.github+json" \
  repos/codecov/codecov-action/releases | \
  jq -r '.[] | select(.tag_name == "v5.0.0") | .body'

# Let's also check our current workflow configuration
cat .github/workflows/coverage.yaml

Length of output: 8069

@@ -349,7 +349,7 @@ replace (
)

require (
buf.build/gen/go/bufbuild/protovalidate/protocolbuffers/go v1.35.1-20240920164238-5a7b106cbb87.1
buf.build/gen/go/bufbuild/protovalidate/protocolbuffers/go v1.35.2-20241127180247-a33202765966.1
cloud.google.com/go/storage v1.43.0
code.cloudfoundry.org/bytefmt v0.0.0-20190710193110-1eb035ffe2b6
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue

Resolve placeholder dependency versions

Several dependencies are using placeholder versions (v0.0.0-00010101000000-000000000000):

  • github.com/akrylysov/pogreb
  • github.com/go-redis/redis/v8
  • github.com/gocraft/dbr/v2
  • github.com/gorilla/mux
  • github.com/grafana/pyroscope-go/godeltaprof
  • github.com/hashicorp/go-version
  • github.com/kpango/fuid
  • github.com/kpango/gache/v2

This could lead to build instability. Please specify explicit versions for these dependencies.

Also applies to: 358-358, 359-359, 361-361, 362-362, 363-363, 364-364, 365-365

Copy link
Contributor

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 3

🧹 Outside diff range and nitpick comments (3)
.github/workflows/coverage.yaml (1)

57-60: Consider pinning to specific version for better security.

While using @v5 works, pinning to a specific version (e.g., @v5.0.7) provides better security and reproducibility by preventing automatic updates to newer patches that might introduce issues.

-        uses: codecov/codecov-action@v5
+        uses: codecov/[email protected]
         with:
           token: ${{secrets.CODECOV_TOKEN}}
           file: ./coverage.out
k8s/index/job/correction/cronjob.yaml (1)

78-86: Consider documenting affinity configuration usage

The addition of empty affinity configurations provides a good foundation for customizing pod scheduling. Consider adding documentation or comments explaining:

  • How to configure node affinity for specific hardware requirements
  • When to use pod affinity for co-location
  • How to leverage pod anti-affinity for high availability
k8s/operator/helm/crds/valdrelease.yaml (1)

Line range hint 78-86: Consider documenting scheduling strategy recommendations

While the affinity configuration implementation is solid, consider adding documentation that covers:

  1. Recommended node affinity settings for different deployment scenarios
  2. Pod affinity strategies for optimizing data locality
  3. Anti-affinity patterns for high availability
  4. Node selector usage guidelines for resource optimization

This would help users effectively utilize these scheduling capabilities.

Also applies to: 8409-8453, 9085-9087, 9778-9822, 13121-13165, 13599-13601

📜 Review details

Configuration used: CodeRabbit UI
Review profile: CHILL

📥 Commits

Reviewing files that changed from the base of the PR and between 9bd9034 and 7fab0e3.

⛔ Files ignored due to path filters (22)
  • apis/grpc/v1/agent/core/agent.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/agent/sidecar/sidecar.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/agent/sidecar/sidecar_vtproto.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/discoverer/discoverer.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/filter/egress/egress_filter.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/filter/ingress/ingress_filter.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/meta/meta.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/mirror/mirror.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/payload/payload.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/rpc/errdetails/error_details.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/filter.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/flush.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/index.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/insert.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/object.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/remove.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/search.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/update.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/upsert.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • example/client/go.sum is excluded by !**/*.sum
  • go.sum is excluded by !**/*.sum
  • rust/Cargo.lock is excluded by !**/*.lock
📒 Files selected for processing (61)
  • .github/ISSUE_TEMPLATE/bug_report.md (1 hunks)
  • .github/ISSUE_TEMPLATE/security_issue_report.md (1 hunks)
  • .github/PULL_REQUEST_TEMPLATE.md (1 hunks)
  • .github/workflows/coverage.yaml (1 hunks)
  • dockers/agent/core/agent/Dockerfile (1 hunks)
  • dockers/agent/core/faiss/Dockerfile (1 hunks)
  • dockers/agent/core/ngt/Dockerfile (1 hunks)
  • dockers/agent/sidecar/Dockerfile (1 hunks)
  • dockers/binfmt/Dockerfile (1 hunks)
  • dockers/buildbase/Dockerfile (1 hunks)
  • dockers/buildkit/Dockerfile (1 hunks)
  • dockers/buildkit/syft/scanner/Dockerfile (1 hunks)
  • dockers/ci/base/Dockerfile (1 hunks)
  • dockers/dev/Dockerfile (1 hunks)
  • dockers/discoverer/k8s/Dockerfile (1 hunks)
  • dockers/example/client/Dockerfile (1 hunks)
  • dockers/gateway/filter/Dockerfile (1 hunks)
  • dockers/gateway/lb/Dockerfile (1 hunks)
  • dockers/gateway/mirror/Dockerfile (1 hunks)
  • dockers/index/job/correction/Dockerfile (1 hunks)
  • dockers/index/job/creation/Dockerfile (1 hunks)
  • dockers/index/job/deletion/Dockerfile (1 hunks)
  • dockers/index/job/readreplica/rotate/Dockerfile (1 hunks)
  • dockers/index/job/save/Dockerfile (1 hunks)
  • dockers/index/operator/Dockerfile (1 hunks)
  • dockers/manager/index/Dockerfile (1 hunks)
  • dockers/operator/helm/Dockerfile (1 hunks)
  • dockers/tools/benchmark/job/Dockerfile (1 hunks)
  • dockers/tools/benchmark/operator/Dockerfile (1 hunks)
  • dockers/tools/cli/loadtest/Dockerfile (1 hunks)
  • example/client/go.mod (2 hunks)
  • go.mod (20 hunks)
  • k8s/index/job/correction/cronjob.yaml (1 hunks)
  • k8s/index/job/creation/cronjob.yaml (1 hunks)
  • k8s/index/job/deletion/configmap.yaml (0 hunks)
  • k8s/index/job/deletion/cronjob.yaml (0 hunks)
  • k8s/index/job/save/cronjob.yaml (1 hunks)
  • k8s/index/operator/configmap.yaml (1 hunks)
  • k8s/index/operator/deployment.yaml (1 hunks)
  • k8s/operator/helm/crds/valdrelease.yaml (9 hunks)
  • rust/rust-toolchain (1 hunks)
  • versions/BUF_VERSION (1 hunks)
  • versions/CMAKE_VERSION (1 hunks)
  • versions/GOLANGCILINT_VERSION (1 hunks)
  • versions/GO_VERSION (1 hunks)
  • versions/HELM_VERSION (1 hunks)
  • versions/KUBECTL_VERSION (1 hunks)
  • versions/OPERATOR_SDK_VERSION (1 hunks)
  • versions/PROMETHEUS_STACK_VERSION (1 hunks)
  • versions/PROTOBUF_VERSION (1 hunks)
  • versions/RUST_VERSION (1 hunks)
  • versions/TELEPRESENCE_VERSION (1 hunks)
  • versions/USEARCH_VERSION (1 hunks)
  • versions/YQ_VERSION (1 hunks)
  • versions/actions/CODECOV_CODECOV_ACTION (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_ANALYZE (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_AUTOBUILD (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_INIT (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_UPLOAD_SARIF (1 hunks)
  • versions/actions/GITHUB_ISSUE_METRICS (1 hunks)
  • versions/actions/REVIEWDOG_ACTION_HADOLINT (1 hunks)
💤 Files with no reviewable changes (2)
  • k8s/index/job/deletion/configmap.yaml
  • k8s/index/job/deletion/cronjob.yaml
✅ Files skipped from review due to trivial changes (50)
  • versions/RUST_VERSION
  • versions/HELM_VERSION
  • versions/BUF_VERSION
  • versions/GO_VERSION
  • versions/GOLANGCILINT_VERSION
  • versions/CMAKE_VERSION
  • versions/USEARCH_VERSION
  • versions/YQ_VERSION
  • versions/KUBECTL_VERSION
  • versions/PROTOBUF_VERSION
  • versions/actions/GITHUB_CODEQL_ACTION_UPLOAD_SARIF
  • versions/actions/GITHUB_CODEQL_ACTION_INIT
  • versions/actions/REVIEWDOG_ACTION_HADOLINT
  • versions/OPERATOR_SDK_VERSION
  • versions/actions/GITHUB_CODEQL_ACTION_ANALYZE
  • rust/rust-toolchain
  • versions/actions/GITHUB_CODEQL_ACTION_AUTOBUILD
  • versions/actions/CODECOV_CODECOV_ACTION
  • dockers/example/client/Dockerfile
  • versions/PROMETHEUS_STACK_VERSION
  • dockers/manager/index/Dockerfile
  • versions/actions/GITHUB_ISSUE_METRICS
  • dockers/operator/helm/Dockerfile
  • dockers/buildkit/syft/scanner/Dockerfile
  • dockers/binfmt/Dockerfile
  • dockers/agent/sidecar/Dockerfile
  • dockers/tools/cli/loadtest/Dockerfile
  • dockers/gateway/lb/Dockerfile
  • dockers/index/job/deletion/Dockerfile
  • dockers/buildbase/Dockerfile
  • dockers/index/job/correction/Dockerfile
  • dockers/index/operator/Dockerfile
  • dockers/agent/core/faiss/Dockerfile
  • dockers/buildkit/Dockerfile
  • dockers/agent/core/ngt/Dockerfile
  • dockers/ci/base/Dockerfile
  • dockers/tools/benchmark/job/Dockerfile
  • dockers/index/job/readreplica/rotate/Dockerfile
  • dockers/agent/core/agent/Dockerfile
  • dockers/gateway/filter/Dockerfile
  • dockers/discoverer/k8s/Dockerfile
  • dockers/gateway/mirror/Dockerfile
  • dockers/tools/benchmark/operator/Dockerfile
  • dockers/dev/Dockerfile
  • .github/ISSUE_TEMPLATE/security_issue_report.md
  • .github/ISSUE_TEMPLATE/bug_report.md
  • dockers/index/job/save/Dockerfile
  • .github/PULL_REQUEST_TEMPLATE.md
  • versions/TELEPRESENCE_VERSION
  • dockers/index/job/creation/Dockerfile
🔇 Additional comments (10)
k8s/index/operator/deployment.yaml (1)

49-49: ConfigMap checksum update looks correct.

The checksum update reflects changes in the underlying ConfigMap. This is consistent with the addition of affinity configurations in the job templates.

Let's verify the ConfigMap changes:

✅ Verification successful

ConfigMap checksum update is consistent with affinity configurations

The ConfigMap content shows that affinity rules are properly configured in all job templates (creation, save, and correction jobs), with empty arrays for node affinity, pod affinity, and pod anti-affinity preferences and requirements. This confirms that the checksum update in the deployment reflects these affinity configurations.

🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Verify the ConfigMap changes that triggered this checksum update
# Compare the old and new ConfigMap content

# Get the ConfigMap content
kubectl get configmap vald-index-operator-config -o yaml 2>/dev/null || \
  gh api /repos/vdaas/vald/contents/k8s/index/operator/configmap.yaml | jq -r '.content' | base64 -d

Length of output: 254


Script:

#!/bin/bash
# Let's try a different approach to verify the ConfigMap changes
# First, let's check if the ConfigMap file exists and view its contents
fd configmap.yaml --type f k8s/index/operator/ --exec cat {}

Length of output: 27906


Script:

#!/bin/bash
# Let's check the git history of the ConfigMap to see what changed
git log -p k8s/index/operator/configmap.yaml | head -n 100

Length of output: 0


Script:

#!/bin/bash
# Let's check for affinity configurations in the job templates
rg -A 5 -B 5 "affinity:" k8s/index/operator/configmap.yaml

Length of output: 27136

k8s/index/operator/configmap.yaml (1)

Line range hint 8409-8453: LGTM: Consistent affinity configuration approach

The affinity configuration structure is consistently applied across job templates, providing a uniform way to customize pod scheduling when needed. The empty arrays allow for future configuration without structural changes.

Also applies to: 13121-13165

k8s/operator/helm/crds/valdrelease.yaml (1)

8409-8453: LGTM: Well-structured affinity and node selector schema definitions

The CRD schema properly defines:

  • Complete affinity configuration structure matching Kubernetes API
  • Proper type definitions and validation
  • Consistent implementation across different components

Also applies to: 9085-9087, 9778-9822, 13121-13165, 13599-13601

example/client/go.mod (2)

14-16: Coordinated update of google.golang.org/genproto packages.

The update of all genproto-related packages to the same version (v0.0.0-20241202173237-19429a94021a) maintains consistency and prevents potential compatibility issues.


18-18: Verify compatibility between updated protobuf versions.

The google.golang.org/protobuf version has been updated to v1.35.2 in both replace and require directives. This consistency is good, but we should verify compatibility with other dependencies.

Also applies to: 40-41

✅ Verification successful

Based on the results, I can generate the final response as we have sufficient information. The search results show:

  1. The protobuf version v1.35.2 is consistently used across the codebase (go.mod, go.sum)
  2. The protobuf imports are primarily used in generated code (*.pb.go files)
  3. The imports use standard protobuf packages that are stable across minor version updates
  4. No custom protobuf usage patterns that might be version-sensitive

Protobuf v1.35.2 update is safe to proceed.

The update appears safe as:

  • All protobuf imports are using standard, stable APIs
  • The code primarily uses generated protobuf files which maintain compatibility
  • No custom or advanced protobuf features are being used that might be affected by this minor version update
🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Check for potential compatibility issues with protobuf version

# Search for protobuf version requirements in go.mod files
rg -A 1 "google.golang.org/protobuf" | grep -v "replace\|require"

# Look for any protobuf-related import statements
ast-grep --pattern 'import $$$"google.golang.org/protobuf/$$$"'

Length of output: 19264

go.mod (5)

8-8: Coordinated update of Google Cloud packages.

The updates to Cloud SDK packages are consistent:

  • compute v1.29.0
  • pubsub v1.45.2
  • storage v1.47.0

This coordinated update helps maintain compatibility between Google Cloud components.

Also applies to: 14-14, 16-16


48-69: Comprehensive update of AWS SDK components.

The AWS SDK components have been updated consistently:

  • aws-sdk-go-v2 v1.32.6
  • All related service packages updated to compatible versions

This coordinated update ensures compatibility across AWS services.


317-317: Coordinated update of Google API and Protobuf packages.

The updates maintain consistency across related packages:

  • google.golang.org/api v0.209.0
  • google.golang.org/genproto and related packages to v0.0.0-20241202173237-19429a94021a
  • google.golang.org/protobuf v1.35.2

This coordinated update helps maintain compatibility between Google API components.

Also applies to: 319-321, 324-324


381-381: New dependencies added.

New dependencies have been introduced:

  • github.com/quic-go/quic-go v0.48.2
  • github.com/unum-cloud/usearch/golang

Ensure these new dependencies are necessary and their versions are stable.

Also applies to: 384-384

✅ Verification successful

Let me gather more information to verify the actual usage of these dependencies.


Dependencies update looks valid and actively used

The changes are minor version updates for actively used dependencies:

  • quic-go: v0.48.1 -> v0.48.2 (used in QUIC networking implementation)
  • usearch/golang: Patch update for the USearch algorithm implementation

Both dependencies are properly integrated into the codebase:

  • QUIC is used in network-related functionality (internal/net/quic/*)
  • USearch is used as one of the core algorithms (internal/core/algorithm/usearch/*)
🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Verify usage of new dependencies

# Check for QUIC usage
ast-grep --pattern 'import $$$"github.com/quic-go/quic-go$$$"'

# Check for USearch usage
ast-grep --pattern 'import $$$"github.com/unum-cloud/usearch/golang$$$"'

Length of output: 17459


Script:

#!/bin/bash
# Search for actual usage of quic-go and usearch packages in the codebase
# beyond just imports

# Search for QUIC usage patterns
rg -l "quic\." 

# Search for USearch usage patterns
rg -l "usearch\."

# Check go.mod history for these dependencies
git diff HEAD~1 go.mod | grep -E "(quic-go|usearch)"

Length of output: 443


228-229: Testing framework updates.

The Ginkgo testing framework has been updated:

  • ginkgo/v2 v2.22.0
  • gomega v1.36.0

Ensure that existing tests are compatible with the new versions.

@@ -54,7 +54,7 @@ jobs:
run: |
make coverage
- name: Upload coverage report to Codecov
uses: codecov/codecov-action@v4
uses: codecov/codecov-action@v5
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codebase verification

⚠️ Update Required: Breaking Changes in Codecov Action v5

The upgrade to codecov-action@v5 requires modifications to the workflow file due to breaking changes:

  • The file parameter in .github/workflows/coverage.yaml needs to be updated to files
  • Several known issues are currently open with v5, including problems with PR comments and tokenless uploads

Recommend either:

  1. Stay on v4 until the reported issues are resolved, or
  2. Update the workflow file to use the new parameter name:
with:
  files: ./coverage.out  # Changed from 'file' to 'files'
🔗 Analysis chain

LGTM! Verify Codecov action compatibility.

The upgrade to codecov-action@v5 is a good practice for staying current. However, as this is a major version bump, let's verify the compatibility.


⚠️ Caution: Breaking Changes in Codecov Action v5

Based on the release notes, v5 introduces several breaking changes that require attention:

  1. The file parameter is deprecated in favor of files
  2. The plugin parameter is deprecated in favor of plugins
  3. Token handling for public repositories has changed with a new opt-out feature

Additionally, there are active issues reported with v5:

🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Verify codecov-action v5 compatibility and check for breaking changes

# Check the changelog and release notes for v5
gh api repos/codecov/codecov-action/releases/tags/v5.0.0 --jq '.body'

# Check if there are any open issues related to v5 migration
gh api search/issues -X GET -f q='repo:codecov/codecov-action is:issue is:open label:bug v5'

Length of output: 36411


Script:

#!/bin/bash
# Check current usage of deprecated parameters
rg -l "file:|plugin:" .github/workflows/

Length of output: 112


Script:

#!/bin/bash
# Check the specific usage of deprecated parameters in the workflow files
rg -A 5 "file:|plugin:" .github/workflows/coverage.yaml .github/workflows/codeql-analysis.yml

Length of output: 971

Comment on lines +78 to +86
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
podAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
podAntiAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

🛠️ Refactor suggestion

Consider adding default pod anti-affinity rules.

While the addition of affinity configuration is good, empty rules don't provide any scheduling benefits. Consider adding default pod anti-affinity rules to avoid scheduling multiple save jobs on the same node, which could help with:

  • Better resource distribution
  • Improved fault tolerance
  • Reduced resource contention

Here's a suggested configuration:

           affinity:
             nodeAffinity:
               preferredDuringSchedulingIgnoredDuringExecution: []
             podAffinity:
               preferredDuringSchedulingIgnoredDuringExecution: []
               requiredDuringSchedulingIgnoredDuringExecution: []
             podAntiAffinity:
-              preferredDuringSchedulingIgnoredDuringExecution: []
+              preferredDuringSchedulingIgnoredDuringExecution:
+                - weight: 100
+                  podAffinityTerm:
+                    labelSelector:
+                      matchExpressions:
+                        - key: app
+                          operator: In
+                          values:
+                            - vald-index-save
+                    topologyKey: kubernetes.io/hostname
               requiredDuringSchedulingIgnoredDuringExecution: []
📝 Committable suggestion

‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.

Suggested change
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
podAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
podAntiAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
podAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
podAntiAffinity:
preferredDuringSchedulingIgnoredDuringExecution:
- weight: 100
podAffinityTerm:
labelSelector:
matchExpressions:
- key: app
operator: In
values:
- vald-index-save
topologyKey: kubernetes.io/hostname
requiredDuringSchedulingIgnoredDuringExecution: []

Comment on lines +78 to +86
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
podAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
podAntiAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

🛠️ Refactor suggestion

Apply consistent affinity rules across jobs.

Similar to the vald-index-save job, consider adding default pod anti-affinity rules. This is particularly important for this job as it runs every minute, making proper pod distribution crucial for system stability.

Apply the same configuration as suggested for vald-index-save, adjusted for this job:

           affinity:
             nodeAffinity:
               preferredDuringSchedulingIgnoredDuringExecution: []
             podAffinity:
               preferredDuringSchedulingIgnoredDuringExecution: []
               requiredDuringSchedulingIgnoredDuringExecution: []
             podAntiAffinity:
-              preferredDuringSchedulingIgnoredDuringExecution: []
+              preferredDuringSchedulingIgnoredDuringExecution:
+                - weight: 100
+                  podAffinityTerm:
+                    labelSelector:
+                      matchExpressions:
+                        - key: app
+                          operator: In
+                          values:
+                            - vald-index-creation
+                    topologyKey: kubernetes.io/hostname
               requiredDuringSchedulingIgnoredDuringExecution: []

Consider implementing these affinity rules as a common template to maintain consistency across all Vald index jobs. This would help ensure uniform scheduling behavior and make future maintenance easier.

📝 Committable suggestion

‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.

Suggested change
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
podAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
podAntiAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
podAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
podAntiAffinity:
preferredDuringSchedulingIgnoredDuringExecution:
- weight: 100
podAffinityTerm:
labelSelector:
matchExpressions:
- key: app
operator: In
values:
- vald-index-creation
topologyKey: kubernetes.io/hostname
requiredDuringSchedulingIgnoredDuringExecution: []

Copy link
Contributor

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 3

🧹 Outside diff range and nitpick comments (4)
k8s/operator/helm/crds/valdrelease.yaml (1)

8409-8453: Scheduling configuration improvements added for Vald components

The changes add standard Kubernetes scheduling configurations (affinity, nodeSelector, tolerations) to the corrector and saver components. This enhances pod scheduling flexibility and control by allowing:

  • Node affinity rules for pod-to-node scheduling preferences
  • Pod affinity/anti-affinity rules for pod-to-pod scheduling preferences
  • Node selector for node label-based scheduling
  • Tolerations for scheduling on tainted nodes

These additions align with Kubernetes best practices for workload placement control.

Consider documenting recommended affinity/anti-affinity patterns for different deployment scenarios, such as:

  • High availability: Using pod anti-affinity to spread across nodes
  • Performance: Using node affinity for specialized hardware
  • Cost optimization: Using node selectors for spot/preemptible instances

Also applies to: 9085-9087, 13121-13165, 13599-13601, 14280-14284

k8s/index/operator/configmap.yaml (2)

Line range hint 350-357: Consider defining specific affinity rules

The affinity configurations are currently empty placeholders. While this provides flexibility, consider defining specific rules based on your deployment requirements:

  • Node affinity rules for hardware requirements
  • Pod anti-affinity rules for high availability
  • Pod affinity rules for co-location needs

Example configuration:

affinity:
  nodeAffinity:
    preferredDuringSchedulingIgnoredDuringExecution:
    - weight: 1
      preference:
        matchExpressions:
        - key: node-type
          operator: In
          values:
          - high-memory
  podAntiAffinity:
    preferredDuringSchedulingIgnoredDuringExecution:
    - weight: 100
      podAffinityTerm:
        labelSelector:
          matchExpressions:
          - key: app
            operator: In
            values:
            - vald-index-creation
        topologyKey: kubernetes.io/hostname

Also applies to: 577-584, 804-811


Line range hint 309-334: Verify init container health check logic

The init containers use a simple HTTP status check for readiness. Consider enhancing the health check logic:

  1. Add timeout to the wget command
  2. Implement exponential backoff for retries
  3. Add a maximum retry limit

Example enhancement:

 command:
   - /bin/sh
   - -e
   - -c
   - |
-    until [ "$(wget --server-response --spider --quiet http://vald-agent.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')" == "200" ]; do
+    max_attempts=30
+    attempt=0
+    until [ "$(wget --timeout=5 --server-response --spider --quiet http://vald-agent.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')" == "200" ] || [ $attempt -ge $max_attempts ]; do
       echo "waiting for agent to be ready..."
-      sleep 2;
+      attempt=$((attempt + 1))
+      sleep $((2 ** (attempt / 3)));
     done
+    if [ $attempt -ge $max_attempts ]; then
+      echo "Timeout waiting for agent to be ready"
+      exit 1
+    fi

Also applies to: 536-561, 763-788

go.mod (1)

Line range hint 352-413: Review direct dependency version constraints

Several direct dependencies are using commit hash constraints (000000000000). This might cause issues with reproducible builds and should be replaced with proper version tags where available.

Consider updating these dependencies to use proper version tags instead of commit hashes. For example:

  • github.com/akrylysov/pogreb
  • github.com/go-redis/redis/v8
  • github.com/gocraft/dbr/v2
  • github.com/gorilla/mux
📜 Review details

Configuration used: CodeRabbit UI
Review profile: CHILL

📥 Commits

Reviewing files that changed from the base of the PR and between 9bd9034 and 7fab0e3.

⛔ Files ignored due to path filters (22)
  • apis/grpc/v1/agent/core/agent.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/agent/sidecar/sidecar.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/agent/sidecar/sidecar_vtproto.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/discoverer/discoverer.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/filter/egress/egress_filter.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/filter/ingress/ingress_filter.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/meta/meta.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/mirror/mirror.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/payload/payload.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/rpc/errdetails/error_details.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/filter.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/flush.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/index.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/insert.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/object.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/remove.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/search.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/update.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/upsert.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • example/client/go.sum is excluded by !**/*.sum
  • go.sum is excluded by !**/*.sum
  • rust/Cargo.lock is excluded by !**/*.lock
📒 Files selected for processing (61)
  • .github/ISSUE_TEMPLATE/bug_report.md (1 hunks)
  • .github/ISSUE_TEMPLATE/security_issue_report.md (1 hunks)
  • .github/PULL_REQUEST_TEMPLATE.md (1 hunks)
  • .github/workflows/coverage.yaml (1 hunks)
  • dockers/agent/core/agent/Dockerfile (1 hunks)
  • dockers/agent/core/faiss/Dockerfile (1 hunks)
  • dockers/agent/core/ngt/Dockerfile (1 hunks)
  • dockers/agent/sidecar/Dockerfile (1 hunks)
  • dockers/binfmt/Dockerfile (1 hunks)
  • dockers/buildbase/Dockerfile (1 hunks)
  • dockers/buildkit/Dockerfile (1 hunks)
  • dockers/buildkit/syft/scanner/Dockerfile (1 hunks)
  • dockers/ci/base/Dockerfile (1 hunks)
  • dockers/dev/Dockerfile (1 hunks)
  • dockers/discoverer/k8s/Dockerfile (1 hunks)
  • dockers/example/client/Dockerfile (1 hunks)
  • dockers/gateway/filter/Dockerfile (1 hunks)
  • dockers/gateway/lb/Dockerfile (1 hunks)
  • dockers/gateway/mirror/Dockerfile (1 hunks)
  • dockers/index/job/correction/Dockerfile (1 hunks)
  • dockers/index/job/creation/Dockerfile (1 hunks)
  • dockers/index/job/deletion/Dockerfile (1 hunks)
  • dockers/index/job/readreplica/rotate/Dockerfile (1 hunks)
  • dockers/index/job/save/Dockerfile (1 hunks)
  • dockers/index/operator/Dockerfile (1 hunks)
  • dockers/manager/index/Dockerfile (1 hunks)
  • dockers/operator/helm/Dockerfile (1 hunks)
  • dockers/tools/benchmark/job/Dockerfile (1 hunks)
  • dockers/tools/benchmark/operator/Dockerfile (1 hunks)
  • dockers/tools/cli/loadtest/Dockerfile (1 hunks)
  • example/client/go.mod (2 hunks)
  • go.mod (20 hunks)
  • k8s/index/job/correction/cronjob.yaml (1 hunks)
  • k8s/index/job/creation/cronjob.yaml (1 hunks)
  • k8s/index/job/deletion/configmap.yaml (0 hunks)
  • k8s/index/job/deletion/cronjob.yaml (0 hunks)
  • k8s/index/job/save/cronjob.yaml (1 hunks)
  • k8s/index/operator/configmap.yaml (1 hunks)
  • k8s/index/operator/deployment.yaml (1 hunks)
  • k8s/operator/helm/crds/valdrelease.yaml (9 hunks)
  • rust/rust-toolchain (1 hunks)
  • versions/BUF_VERSION (1 hunks)
  • versions/CMAKE_VERSION (1 hunks)
  • versions/GOLANGCILINT_VERSION (1 hunks)
  • versions/GO_VERSION (1 hunks)
  • versions/HELM_VERSION (1 hunks)
  • versions/KUBECTL_VERSION (1 hunks)
  • versions/OPERATOR_SDK_VERSION (1 hunks)
  • versions/PROMETHEUS_STACK_VERSION (1 hunks)
  • versions/PROTOBUF_VERSION (1 hunks)
  • versions/RUST_VERSION (1 hunks)
  • versions/TELEPRESENCE_VERSION (1 hunks)
  • versions/USEARCH_VERSION (1 hunks)
  • versions/YQ_VERSION (1 hunks)
  • versions/actions/CODECOV_CODECOV_ACTION (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_ANALYZE (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_AUTOBUILD (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_INIT (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_UPLOAD_SARIF (1 hunks)
  • versions/actions/GITHUB_ISSUE_METRICS (1 hunks)
  • versions/actions/REVIEWDOG_ACTION_HADOLINT (1 hunks)
💤 Files with no reviewable changes (2)
  • k8s/index/job/deletion/cronjob.yaml
  • k8s/index/job/deletion/configmap.yaml
✅ Files skipped from review due to trivial changes (51)
  • versions/BUF_VERSION
  • versions/GO_VERSION
  • versions/USEARCH_VERSION
  • versions/HELM_VERSION
  • versions/PROTOBUF_VERSION
  • versions/RUST_VERSION
  • versions/actions/GITHUB_CODEQL_ACTION_UPLOAD_SARIF
  • versions/PROMETHEUS_STACK_VERSION
  • versions/actions/REVIEWDOG_ACTION_HADOLINT
  • rust/rust-toolchain
  • versions/actions/GITHUB_CODEQL_ACTION_INIT
  • versions/KUBECTL_VERSION
  • versions/actions/CODECOV_CODECOV_ACTION
  • dockers/example/client/Dockerfile
  • dockers/agent/core/faiss/Dockerfile
  • dockers/manager/index/Dockerfile
  • versions/GOLANGCILINT_VERSION
  • dockers/gateway/lb/Dockerfile
  • versions/CMAKE_VERSION
  • versions/actions/GITHUB_CODEQL_ACTION_ANALYZE
  • dockers/agent/sidecar/Dockerfile
  • dockers/index/job/readreplica/rotate/Dockerfile
  • versions/actions/GITHUB_ISSUE_METRICS
  • dockers/index/operator/Dockerfile
  • versions/OPERATOR_SDK_VERSION
  • dockers/index/job/correction/Dockerfile
  • dockers/binfmt/Dockerfile
  • versions/actions/GITHUB_CODEQL_ACTION_AUTOBUILD
  • dockers/buildbase/Dockerfile
  • dockers/gateway/mirror/Dockerfile
  • dockers/operator/helm/Dockerfile
  • dockers/discoverer/k8s/Dockerfile
  • dockers/dev/Dockerfile
  • dockers/buildkit/syft/scanner/Dockerfile
  • dockers/gateway/filter/Dockerfile
  • dockers/buildkit/Dockerfile
  • versions/TELEPRESENCE_VERSION
  • dockers/index/job/save/Dockerfile
  • versions/YQ_VERSION
  • dockers/tools/benchmark/operator/Dockerfile
  • dockers/ci/base/Dockerfile
  • dockers/agent/core/ngt/Dockerfile
  • dockers/tools/benchmark/job/Dockerfile
  • .github/ISSUE_TEMPLATE/bug_report.md
  • dockers/index/job/deletion/Dockerfile
  • dockers/index/job/creation/Dockerfile
  • dockers/agent/core/agent/Dockerfile
  • .github/ISSUE_TEMPLATE/security_issue_report.md
  • dockers/tools/cli/loadtest/Dockerfile
  • .github/PULL_REQUEST_TEMPLATE.md
  • k8s/index/operator/deployment.yaml
🔇 Additional comments (10)
.github/workflows/coverage.yaml (1)

57-57: Verify Codecov Action v5 compatibility

The upgrade from codecov/codecov-action@v4 to v5 is a major version change that might include breaking changes.

Please ensure:

  1. All required parameters for v5 are properly configured
  2. The token and file parameters are still valid in v5
  3. No additional parameters are required for v5
✅ Verification successful

Let me gather more information about the v5 migration and configuration requirements.


Based on the gathered information, I can now generate the final response:

Codecov Action v5 upgrade is compatible

The upgrade from v4 to v5 is safe to proceed as:

  • No breaking changes are reported in the latest release notes
  • No open issues related to v5 migration
  • The current configuration using token and file parameters remains valid in v5
  • The workflow file shows a standard configuration pattern that is compatible with v5
🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Verify Codecov Action v5 compatibility and requirements

# Check the latest release notes and changes
gh api repos/codecov/codecov-action/releases/latest --jq '.body'

# Check if there are any open issues related to v5 migration
gh api search/issues -X GET -f q='repo:codecov/codecov-action is:issue is:open label:breaking-change v5' --jq '.items[].title'

# Verify if other repositories are using similar configuration with v5
rg -A 3 "codecov/codecov-action@v5" .github/

Length of output: 811


Script:

#!/bin/bash
# Check the workflow file content to understand the complete configuration
cat .github/workflows/coverage.yaml

# Search for any migration guides or documentation in the repository
fd README.md --exec cat {}

# Check for any deprecated parameters in the workflow
rg "deprecated|warning" .github/workflows/coverage.yaml

Length of output: 67840

k8s/operator/helm/crds/valdrelease.yaml (1)

8409-8453: Schema structure follows Kubernetes conventions

The affinity configuration schema correctly follows Kubernetes API conventions with:

  • Proper nesting of nodeAffinity, podAffinity, podAntiAffinity
  • Correct use of preferredDuringSchedulingIgnoredDuringExecution and requiredDuringSchedulingIgnoredDuringExecution
  • Consistent use of x-kubernetes-preserve-unknown-fields for extensibility

Also applies to: 13121-13165

k8s/index/job/save/cronjob.yaml (1)

78-86: Consider validating the empty affinity configuration

While the structure for affinity rules is in place, all arrays are currently empty, which means these affinity rules have no effect on pod scheduling. If this is intended as a template for future customization, consider:

  1. Adding a comment explaining the purpose of these empty rules
  2. Documenting examples of recommended affinity rules in the configuration

Let's check if this is a common pattern in the codebase:

✅ Verification successful

Let me check the documentation to see if there's any explanation for this pattern.


Empty affinity configuration is intentional and documented

The empty affinity configuration is actually a documented pattern in the capacity planning guide. According to the documentation, these empty arrays serve as customizable templates for users to configure their own affinity rules based on their specific deployment needs. The documentation explicitly mentions:

  1. Affinity settings are meant to be customized in the values.yaml file
  2. The documentation provides examples of how to configure these settings
  3. There's a specific warning about Agent's affinity being significant for the Vald cluster
🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Search for similar empty affinity configurations
rg -U "affinity:\s*\n\s*nodeAffinity:\s*\n\s*preferredDuringSchedulingIgnoredDuringExecution:\s*\[\]" .

Length of output: 2254


Script:

#!/bin/bash
# Check the capacity planning documentation for affinity explanation
cat ./docs/user-guides/capacity-planning.md

Length of output: 6789

k8s/index/operator/configmap.yaml (3)

28-28: LGTM: ConfigMap structure and metadata are well-defined

The ConfigMap is properly structured with appropriate metadata, labels, and version information.


Line range hint 385-397: LGTM: Comprehensive security context configuration

The security context configuration follows best practices:

  • Non-root user execution
  • Read-only root filesystem
  • Dropped capabilities
  • Proper filesystem group settings

Also applies to: 612-624, 839-851


Line range hint 398-411: LGTM: Well-configured environment variables

The environment variables are properly configured using Kubernetes downward API for pod metadata.

Also applies to: 625-638, 852-865

example/client/go.mod (1)

14-16: Verify the impact of genproto updates

The update of Google API proto packages to v0.0.0-20241202173237-19429a94021a appears consistent. However, since these are development snapshots, we should verify there are no breaking changes.

✅ Verification successful

The previous attempt didn't provide clear information about breaking changes. This new script will help us understand:

  1. How the genproto packages are actually used in the code
  2. The exact version changes in go.mod
  3. Specific usage of critical API components that are more likely to be affected by breaking changes

Genproto update appears safe to proceed

Based on the analysis of the codebase:

  1. The genproto packages are primarily used for generated protobuf code and annotations
  2. The main usage is limited to well-established RPC types (status, code) which are stable
  3. The version update is a minor development snapshot change (from 20241118233622 to 20241202173237)
  4. All three genproto packages (core, api, rpc) are being updated to the same consistent version

The usage patterns show that the code primarily depends on stable RPC interfaces that rarely have breaking changes, and the version bump is relatively small (about 2 weeks). The risk of breaking changes is minimal.

🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Check for any breaking changes in the updated genproto packages

# Search for any API usage that might be affected
rg -l "google.golang.org/genproto" | xargs -I {} rg -A 5 "google.golang.org/genproto" {}

# Check commit messages between versions for breaking changes
gh api repos/googleapis/go-genproto/commits \
  --jq '.[] | select(.commit.message | contains("BREAKING"))' \
  --template '{{range .}}{{.commit.message}}{{end}}'

Length of output: 11504


Script:

#!/bin/bash
# Let's try a different approach to check for potential breaking changes

# Check the actual usage of genproto packages in the code
rg -B 2 -A 2 "google\.golang\.org/genproto" --type go

# Check the difference between the old and new versions in go.mod
rg -A 1 "google\.golang\.org/genproto.*v0\.0\.0-\d+" example/client/go.mod

# Look for any error-related imports that might be affected
rg "google\.golang\.org/genproto/googleapis/(api|rpc)/(status|errdetails|code)" --type go

Length of output: 9643

go.mod (3)

319-324: Verify Google API and Protobuf compatibility

The updates to Google APIs and Protobuf are significant:

  • genproto packages updated to latest development snapshot
  • gRPC updated to v1.68.0
  • protobuf updated to v1.35.2

These changes appear consistent across related packages.


228-229: Verify test framework compatibility

The update of Ginkgo (v2.22.0) and Gomega (v1.36.0) test frameworks might require test adjustments.


48-68: Review AWS SDK updates for breaking changes

Multiple AWS SDK packages have been updated to newer versions. The update pattern looks consistent across related packages, but we should verify the impact of these changes.

✅ Verification successful

The previous attempt didn't provide enough context about the actual SDK usage in the codebase. These new commands will help us understand how the AWS SDK is being used and potentially identify any breaking changes that might affect the codebase.


Based on the search results, I can now generate the final response:

AWS SDK updates appear safe to proceed

The codebase shows limited and well-structured AWS SDK usage, primarily focused on S3 operations in the internal/db/storage/blob/s3 package. The updates are minor version bumps (no major version changes) and the SDK's core interfaces for S3 operations remain stable in these versions.

  • The main AWS SDK usage is for S3 operations using GetObjectWithContext
  • The code uses stable AWS SDK patterns (e.g., aws.String, aws.Context)
  • No usage of deprecated or removed features was found in the codebase
🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Check AWS SDK usage and potential breaking changes

# Search for AWS SDK usage patterns
rg -l "github.com/aws/aws-sdk-go-v2" | xargs -I {} rg -A 5 "aws-sdk-go-v2" {}

# Check AWS SDK migration guides and release notes
gh api repos/aws/aws-sdk-go-v2/releases \
  --jq '.[] | select(.body | contains("BREAKING CHANGES"))' \
  --template '{{range .}}{{.body}}{{end}}'

Length of output: 10954


Script:

#!/bin/bash
# Let's search for actual AWS SDK usage patterns in the code to understand the impact

# Search for AWS service client initializations and usage
rg -A 5 "New(Client|From)" --type go

# Search for specific AWS service operations that might be affected
rg -A 3 "\.(Get|Put|Create|Delete|Update|List|Describe)" --type go | grep -i "aws"

# Look for AWS SDK version imports in Go files
rg "github.com/aws/aws-sdk-go-v2" --type go -B 2 -A 2

Length of output: 12350

Comment on lines +78 to +86
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
podAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
podAntiAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

🛠️ Refactor suggestion

Review scheduling frequency with affinity rules

The CronJob runs every minute (* * * * *) which could create scheduling pressure. Consider:

  1. Implementing pod anti-affinity rules to ensure better distribution of load across nodes
  2. Adding node affinity rules to target nodes with appropriate resources

Example configuration:

           affinity:
             nodeAffinity:
-              preferredDuringSchedulingIgnoredDuringExecution: []
+              preferredDuringSchedulingIgnoredDuringExecution:
+              - weight: 100
+                preference:
+                  matchExpressions:
+                  - key: node-role.kubernetes.io/worker
+                    operator: In
+                    values:
+                    - "true"
             podAntiAffinity:
-              preferredDuringSchedulingIgnoredDuringExecution: []
+              preferredDuringSchedulingIgnoredDuringExecution:
+              - weight: 100
+                podAffinityTerm:
+                  labelSelector:
+                    matchLabels:
+                      app: vald-index-creation
+                  topologyKey: kubernetes.io/hostname

Committable suggestion skipped: line range outside the PR's diff.

Comment on lines +78 to +86
affinity:
nodeAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
podAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
podAntiAffinity:
preferredDuringSchedulingIgnoredDuringExecution: []
requiredDuringSchedulingIgnoredDuringExecution: []
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codebase verification

Based on the documentation findings, I can now generate the final response:

Define pod anti-affinity rules for index jobs to ensure optimal scheduling

The documentation reveals that pod anti-affinity rules are considered significant for Vald components, with specific guidelines and default configurations for agents and discoverers. For index jobs:

  • Add preferredDuringSchedulingIgnoredDuringExecution pod anti-affinity rules to spread index jobs across nodes:
affinity:
  podAntiAffinity:
    preferredDuringSchedulingIgnoredDuringExecution:
    - weight: 100
      podAffinityTerm:
        labelSelector:
          matchExpressions:
          - key: app
            operator: In
            values: ["vald-index"]
        topologyKey: "kubernetes.io/hostname"
  • This follows the same pattern used by other Vald components while ensuring index jobs don't compete for resources on the same node.
🔗 Analysis chain

Consider standardizing affinity rules across index jobs

The empty affinity configuration is consistent across all index-related CronJobs. To improve the overall system reliability:

  1. Define a standard set of affinity rules that works for all index jobs
  2. Document the rationale for affinity settings in the configuration
  3. Consider the relationship between these jobs when defining pod affinity/anti-affinity rules

Let's check for any existing affinity documentation or standards:


Let's examine the content of these documentation files to understand if there are any existing affinity standards or guidelines.

🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Search for affinity-related documentation
rg -l "affinity" --type md

Length of output: 197


Script:

#!/bin/bash
# Check content of relevant documentation files
echo "=== Configuration Guide ==="
rg -A 5 "affinity" docs/user-guides/configuration.md

echo -e "\n=== Capacity Planning Guide ==="
rg -A 5 "affinity" docs/user-guides/capacity-planning.md

echo -e "\n=== Main Chart README ==="
rg -A 5 "affinity" charts/vald/README.md

Length of output: 66493

google.golang.org/grpc => google.golang.org/grpc v1.68.0
google.golang.org/protobuf => google.golang.org/protobuf v1.35.1
google.golang.org/protobuf => google.golang.org/protobuf v1.35.2
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codebase verification

Update protobuf version to address security vulnerabilities

The current version v1.35.2 is vulnerable to a moderate severity issue (infinite loop in protojson.Unmarshal). This was patched in v1.33.0.

  • Update the protobuf version to at least v1.33.0 to address the security vulnerability
  • Consider updating to the latest stable version for additional fixes and improvements
🔗 Analysis chain

Verify protobuf version compatibility

The update to google.golang.org/protobuf v1.35.2 is a minor version bump. This version should be compatible, but let's verify its usage and check for any security advisories.

Also applies to: 41-41

🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Check protobuf usage and security advisories

# Search for protobuf message definitions that might be affected
rg -l "proto.Message" | xargs -I {} rg -A 5 "proto.Message" {}

# Check for any security advisories
gh api graphql -f query='
{
  securityVulnerabilities(first: 5, ecosystem: GO, package: "google.golang.org/protobuf") {
    nodes {
      advisory {
        summary
        severity
        publishedAt
      }
      vulnerableVersionRange
      firstPatchedVersion {
        identifier
      }
    }
  }
}'

Length of output: 42249

Copy link
Contributor

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 3

🧹 Outside diff range and nitpick comments (5)
k8s/index/operator/configmap.yaml (2)

28-28: Consider optimizing metrics server configuration

The metrics server configuration could be improved:

  1. TCP settings (cork, no_delay, quick_ack) might not be optimal for metrics endpoints
  2. The write_timeout of 1m might be too long for metrics endpoints

Consider applying this configuration update:

      metrics_servers:
        - name: pprof
          http:
-           write_timeout: 1m
+           write_timeout: 30s
          socket_option:
-           tcp_cork: true
+           tcp_cork: false
-           tcp_no_delay: false
+           tcp_no_delay: true

28-28: Review job template configurations

The job templates for index operations (creation, save, correction) include appropriate:

  • TTL settings (86400s = 24h)
  • Resource limits
  • Security contexts
  • Init containers for dependency checks
  • Proper service account bindings

Consider implementing:

  1. Backoff limits for failed jobs
  2. Pod disruption budgets for critical operations
k8s/index/job/save/cronjob.yaml (1)

78-86: Consider adding specific affinity rules or remove if not needed

The empty affinity configuration doesn't provide any scheduling constraints. Consider either:

  1. Removing these empty rules if pod scheduling requirements aren't defined yet
  2. Adding specific rules to optimize pod placement, such as:
    • Using nodeAffinity to schedule on nodes with specific resources
    • Using podAntiAffinity to spread index jobs across nodes

Here's an example of meaningful affinity rules:

 affinity:
   nodeAffinity:
-    preferredDuringSchedulingIgnoredDuringExecution: []
+    preferredDuringSchedulingIgnoredDuringExecution:
+    - weight: 100
+      preference:
+        matchExpressions:
+        - key: node-type
+          operator: In
+          values:
+          - high-memory
   podAntiAffinity:
-    preferredDuringSchedulingIgnoredDuringExecution: []
+    preferredDuringSchedulingIgnoredDuringExecution:
+    - weight: 100
+      podAffinityTerm:
+        labelSelector:
+          matchExpressions:
+          - key: app
+            operator: In
+            values:
+            - vald-index-save
+        topologyKey: kubernetes.io/hostname
k8s/index/job/creation/cronjob.yaml (1)

Based on the configuration files examined, I can now provide a definitive response about the scheduling concern.

Running every minute is likely not intentional and should be adjusted

The aggressive * * * * * schedule combined with a long startingDeadlineSeconds: 43200 (12 hours) could lead to resource contention issues. This configuration suggests:

  1. The job will attempt to run every minute
  2. Failed jobs can be retried for up to 12 hours after their scheduled time
  3. While concurrencyPolicy: Forbid prevents parallel runs, the frequent scheduling could still cause unnecessary load

Recommendations:

  • Adjust the schedule based on actual index creation needs (e.g., */15 * * * * for every 15 minutes or 0 */1 * * * for hourly)
  • Consider reducing startingDeadlineSeconds to match the intended schedule frequency
  • Document the intended schedule in comments or documentation
🔗 Analysis chain

Line range hint 32-32: Verify if running every minute is intentional

The job is scheduled to run every minute (* * * * *). This aggressive schedule might:

  1. Create unnecessary load on the cluster
  2. Potentially overlap with previous runs despite concurrencyPolicy: Forbid
  3. Consume more resources than necessary

Consider adjusting the schedule based on actual index creation needs.


The previous attempt with kubectl failed as expected since we're in a sandbox environment. Let's examine the actual configuration files and documentation to understand the scheduling configuration and its rationale.

🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Check for any existing index creation jobs and their completion times
kubectl get jobs -l app=vald-index-creation --sort-by=.metadata.creationTimestamp -o json | \
  jq -r '.items[] | select(.status.completionTime != null) | {name: .metadata.name, startTime: .status.startTime, completionTime: .status.completionTime}'

Length of output: 344


Script:

#!/bin/bash
# Let's examine the cronjob configuration and related files
rg -A 5 -B 5 "startingDeadlineSeconds|concurrencyPolicy|schedule" k8s/index/job/creation/cronjob.yaml

# Also check for any documentation or configuration that might explain the scheduling decision
fd README.md --exec cat {} \; 

Length of output: 79446

k8s/operator/helm/crds/valdrelease.yaml (1)

Line range hint 8409-14284: Consider adding schema validation examples

While the schema is technically correct, it would be helpful to add examples in the CRD documentation to demonstrate proper usage of these scheduling configurations.

📜 Review details

Configuration used: CodeRabbit UI
Review profile: CHILL

📥 Commits

Reviewing files that changed from the base of the PR and between 9bd9034 and 7fab0e3.

⛔ Files ignored due to path filters (22)
  • apis/grpc/v1/agent/core/agent.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/agent/sidecar/sidecar.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/agent/sidecar/sidecar_vtproto.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/discoverer/discoverer.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/filter/egress/egress_filter.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/filter/ingress/ingress_filter.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/meta/meta.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/mirror/mirror.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/payload/payload.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/rpc/errdetails/error_details.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/filter.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/flush.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/index.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/insert.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/object.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/remove.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/search.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/update.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • apis/grpc/v1/vald/upsert.pb.go is excluded by !**/*.pb.go, !**/*.pb.go
  • example/client/go.sum is excluded by !**/*.sum
  • go.sum is excluded by !**/*.sum
  • rust/Cargo.lock is excluded by !**/*.lock
📒 Files selected for processing (61)
  • .github/ISSUE_TEMPLATE/bug_report.md (1 hunks)
  • .github/ISSUE_TEMPLATE/security_issue_report.md (1 hunks)
  • .github/PULL_REQUEST_TEMPLATE.md (1 hunks)
  • .github/workflows/coverage.yaml (1 hunks)
  • dockers/agent/core/agent/Dockerfile (1 hunks)
  • dockers/agent/core/faiss/Dockerfile (1 hunks)
  • dockers/agent/core/ngt/Dockerfile (1 hunks)
  • dockers/agent/sidecar/Dockerfile (1 hunks)
  • dockers/binfmt/Dockerfile (1 hunks)
  • dockers/buildbase/Dockerfile (1 hunks)
  • dockers/buildkit/Dockerfile (1 hunks)
  • dockers/buildkit/syft/scanner/Dockerfile (1 hunks)
  • dockers/ci/base/Dockerfile (1 hunks)
  • dockers/dev/Dockerfile (1 hunks)
  • dockers/discoverer/k8s/Dockerfile (1 hunks)
  • dockers/example/client/Dockerfile (1 hunks)
  • dockers/gateway/filter/Dockerfile (1 hunks)
  • dockers/gateway/lb/Dockerfile (1 hunks)
  • dockers/gateway/mirror/Dockerfile (1 hunks)
  • dockers/index/job/correction/Dockerfile (1 hunks)
  • dockers/index/job/creation/Dockerfile (1 hunks)
  • dockers/index/job/deletion/Dockerfile (1 hunks)
  • dockers/index/job/readreplica/rotate/Dockerfile (1 hunks)
  • dockers/index/job/save/Dockerfile (1 hunks)
  • dockers/index/operator/Dockerfile (1 hunks)
  • dockers/manager/index/Dockerfile (1 hunks)
  • dockers/operator/helm/Dockerfile (1 hunks)
  • dockers/tools/benchmark/job/Dockerfile (1 hunks)
  • dockers/tools/benchmark/operator/Dockerfile (1 hunks)
  • dockers/tools/cli/loadtest/Dockerfile (1 hunks)
  • example/client/go.mod (2 hunks)
  • go.mod (20 hunks)
  • k8s/index/job/correction/cronjob.yaml (1 hunks)
  • k8s/index/job/creation/cronjob.yaml (1 hunks)
  • k8s/index/job/deletion/configmap.yaml (0 hunks)
  • k8s/index/job/deletion/cronjob.yaml (0 hunks)
  • k8s/index/job/save/cronjob.yaml (1 hunks)
  • k8s/index/operator/configmap.yaml (1 hunks)
  • k8s/index/operator/deployment.yaml (1 hunks)
  • k8s/operator/helm/crds/valdrelease.yaml (9 hunks)
  • rust/rust-toolchain (1 hunks)
  • versions/BUF_VERSION (1 hunks)
  • versions/CMAKE_VERSION (1 hunks)
  • versions/GOLANGCILINT_VERSION (1 hunks)
  • versions/GO_VERSION (1 hunks)
  • versions/HELM_VERSION (1 hunks)
  • versions/KUBECTL_VERSION (1 hunks)
  • versions/OPERATOR_SDK_VERSION (1 hunks)
  • versions/PROMETHEUS_STACK_VERSION (1 hunks)
  • versions/PROTOBUF_VERSION (1 hunks)
  • versions/RUST_VERSION (1 hunks)
  • versions/TELEPRESENCE_VERSION (1 hunks)
  • versions/USEARCH_VERSION (1 hunks)
  • versions/YQ_VERSION (1 hunks)
  • versions/actions/CODECOV_CODECOV_ACTION (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_ANALYZE (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_AUTOBUILD (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_INIT (1 hunks)
  • versions/actions/GITHUB_CODEQL_ACTION_UPLOAD_SARIF (1 hunks)
  • versions/actions/GITHUB_ISSUE_METRICS (1 hunks)
  • versions/actions/REVIEWDOG_ACTION_HADOLINT (1 hunks)
💤 Files with no reviewable changes (2)
  • k8s/index/job/deletion/configmap.yaml
  • k8s/index/job/deletion/cronjob.yaml
✅ Files skipped from review due to trivial changes (50)
  • versions/BUF_VERSION
  • versions/GO_VERSION
  • versions/OPERATOR_SDK_VERSION
  • versions/HELM_VERSION
  • versions/RUST_VERSION
  • versions/actions/GITHUB_ISSUE_METRICS
  • versions/KUBECTL_VERSION
  • versions/PROTOBUF_VERSION
  • versions/actions/GITHUB_CODEQL_ACTION_AUTOBUILD
  • versions/USEARCH_VERSION
  • versions/YQ_VERSION
  • versions/PROMETHEUS_STACK_VERSION
  • versions/actions/GITHUB_CODEQL_ACTION_UPLOAD_SARIF
  • versions/actions/GITHUB_CODEQL_ACTION_INIT
  • rust/rust-toolchain
  • versions/TELEPRESENCE_VERSION
  • versions/CMAKE_VERSION
  • versions/GOLANGCILINT_VERSION
  • versions/actions/CODECOV_CODECOV_ACTION
  • versions/actions/REVIEWDOG_ACTION_HADOLINT
  • dockers/operator/helm/Dockerfile
  • dockers/gateway/filter/Dockerfile
  • dockers/buildbase/Dockerfile
  • dockers/index/job/readreplica/rotate/Dockerfile
  • dockers/binfmt/Dockerfile
  • dockers/index/operator/Dockerfile
  • dockers/tools/cli/loadtest/Dockerfile
  • dockers/agent/core/ngt/Dockerfile
  • dockers/buildkit/Dockerfile
  • dockers/buildkit/syft/scanner/Dockerfile
  • dockers/index/job/correction/Dockerfile
  • dockers/agent/core/agent/Dockerfile
  • versions/actions/GITHUB_CODEQL_ACTION_ANALYZE
  • dockers/discoverer/k8s/Dockerfile
  • dockers/ci/base/Dockerfile
  • .github/ISSUE_TEMPLATE/security_issue_report.md
  • dockers/tools/benchmark/job/Dockerfile
  • dockers/example/client/Dockerfile
  • dockers/tools/benchmark/operator/Dockerfile
  • dockers/agent/core/faiss/Dockerfile
  • dockers/gateway/lb/Dockerfile
  • dockers/index/job/save/Dockerfile
  • dockers/index/job/deletion/Dockerfile
  • .github/ISSUE_TEMPLATE/bug_report.md
  • dockers/manager/index/Dockerfile
  • dockers/agent/sidecar/Dockerfile
  • dockers/index/job/creation/Dockerfile
  • dockers/dev/Dockerfile
  • dockers/gateway/mirror/Dockerfile
  • .github/PULL_REQUEST_TEMPLATE.md
🔇 Additional comments (15)
k8s/index/operator/deployment.yaml (1)

49-49: LGTM: ConfigMap checksum update

The checksum update is consistent with the changes made to the ConfigMap content.

k8s/index/operator/configmap.yaml (2)

28-28: Review health check configuration

The health check configuration appears robust with:

  • Separate liveness and readiness probes
  • Appropriate timeouts and failure thresholds
  • HTTP/2 disabled for health check endpoints (good for simplicity)
  • TCP socket options optimized for quick health checks

28-28: Verify the security settings in the configuration

The configuration includes several security-related settings that should be reviewed:

  1. TLS is disabled by default
  2. All containers run as non-root (user 65532)
  3. Read-only root filesystem is enabled
  4. All Linux capabilities are dropped
✅ Verification successful

Security settings are properly configured across all components

The security settings in the configuration have been verified and are properly configured:

  • All containers run as non-root user (UID: 65532)
  • Read-only root filesystem is enabled
  • All Linux capabilities are dropped
  • Privilege escalation is disabled
  • Pod security context is properly set with fsGroup and runAsNonRoot
  • TLS configuration is present but disabled by default, which is acceptable for initial setup
🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Verify security settings across all job templates
# Check for security context configurations
rg -A 10 "securityContext:" k8s/index/operator/

Length of output: 28389

k8s/index/job/creation/cronjob.yaml (1)

78-86: Reference previous affinity configuration comment

The empty affinity configuration is identical to k8s/index/job/save/cronjob.yaml. Please refer to the previous comment about adding specific affinity rules.

k8s/index/job/correction/cronjob.yaml (1)

78-86: Reference previous affinity configuration comment

The empty affinity configuration is identical to k8s/index/job/save/cronjob.yaml. Please refer to the previous comment about adding specific affinity rules.

k8s/operator/helm/crds/valdrelease.yaml (3)

8409-8453: LGTM: Well-structured affinity configuration

The affinity configuration follows Kubernetes best practices by:

  • Including all standard affinity types (node/pod/anti-pod)
  • Using proper schema validation
  • Maintaining flexibility with preserve-unknown-fields

9085-9087: LGTM: Standard nodeSelector implementation

The nodeSelector configuration follows the standard Kubernetes pattern and preserves unknown fields for flexibility.


14280-14284: LGTM: Proper tolerations schema

The tolerations array is well-structured and follows Kubernetes conventions:

  • Allows multiple toleration entries
  • Preserves unknown fields for maximum compatibility
  • Enables flexible pod scheduling
example/client/go.mod (2)

14-16: Consistent update of google.golang.org/genproto packages

The update of genproto and its subpackages to the same pseudo-version v0.0.0-20241202173237-19429a94021a maintains consistency across related packages.


18-18: Verified protobuf version update

The update of google.golang.org/protobuf from v1.35.1 to v1.35.2 is a patch version increment, suggesting backward compatibility is maintained.

Also applies to: 40-41

✅ Verification successful

Protobuf version 1.35.2 is vulnerable to a known security issue

The security scan reveals that versions below 1.33.0 are vulnerable to a moderate severity issue where protojson.Unmarshal can enter an infinite loop with invalid JSON (published March 2024). The current update to v1.35.2 should be revised to use at least v1.33.0 or newer to patch this vulnerability.

  • The vulnerability affects all versions < 1.33.0
  • First patched version: 1.33.0
  • Current version in go.mod: 1.35.2
  • Issue: Infinite loop in protojson.Unmarshal with invalid JSON input
🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Verify if the new protobuf version has any known security advisories

# Check for security advisories
gh api graphql -f query='
{
  securityVulnerabilities(first: 5, ecosystem: GO, package: "google.golang.org/protobuf") {
    nodes {
      advisory {
        summary
        severity
        publishedAt
      }
      vulnerableVersionRange
      firstPatchedVersion {
        identifier
      }
    }
  }
}'

Length of output: 874

go.mod (5)

8-8: Verified Cloud provider SDK updates

Updates to Cloud provider SDKs:

  • cloud.google.com/go/compute v1.29.0
  • cloud.google.com/go/pubsub v1.45.2
  • cloud.google.com/go/storage v1.47.0
    These are minor version updates that should maintain backward compatibility.

Also applies to: 14-14, 16-16, 18-18


317-317: Google API and protobuf updates

Coordinated updates of Google-related packages:

  • google.golang.org/api v0.209.0
  • google.golang.org/genproto and related packages to v0.0.0-20241202173237-19429a94021a
  • google.golang.org/protobuf v1.35.2

These updates are consistent with the changes in example/client/go.mod.

Also applies to: 319-321, 324-324


48-69: Comprehensive AWS SDK update

The AWS SDK components have been updated consistently across all packages. This coordinated update helps maintain compatibility between AWS SDK components.

✅ Verification successful

Let me gather more information to verify the version compatibility across all AWS SDK components.


AWS SDK v2 components are properly aligned and compatible

The verification shows that all AWS SDK Go v2 components are properly updated and aligned:

  • Core SDK is at v1.32.6
  • All dependent packages have matching version numbers in both go.mod and go.sum
  • The smithy-go dependency is at v1.22.1, which is the correct version for this SDK release
  • Internal packages and service-specific packages have compatible version numbers that correspond to the core SDK version
🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Verify AWS SDK version compatibility

# Check if all AWS SDK packages are at compatible versions
rg -A 1 'github.com/aws/aws-sdk-go-v2.*v1.32' | sort

Length of output: 567


Script:

#!/bin/bash
# Check versions of all AWS SDK components to ensure they are compatible
# Looking at both go.mod and go.sum for completeness

echo "=== AWS SDK Components in go.mod ==="
rg '^[[:space:]]*github.com/aws/' go.mod | sort

echo -e "\n=== Corresponding entries in go.sum ==="
rg '^github.com/aws/.*v1\.' go.sum | sort | uniq

Length of output: 6570


228-229: Testing framework updates

Update of testing frameworks:

  • github.com/onsi/ginkgo/v2 v2.22.0
  • github.com/onsi/gomega v1.36.0
✅ Verification successful

Based on the results from our verification attempts, I can now generate the final response:

Minor version updates of testing frameworks appear safe

The updates from Ginkgo v2.21.0 to v2.22.0 and Gomega v1.35.1 to v1.36.0 are minor version bumps. While the codebase contains numerous test files, our analysis shows no direct usage of these testing frameworks (no imports or framework-specific syntax found), suggesting these are likely transitive dependencies. Minor version updates typically maintain backward compatibility, making these changes low risk.

🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Check for any breaking changes in test files due to framework updates

# Search for test files using these frameworks
rg -l 'github.com/onsi/(ginkgo|gomega)' --type go

Length of output: 51


Script:

#!/bin/bash
# Let's try a different approach to find test files and framework usage

# 1. Find all test files
echo "=== Go test files ==="
fd -e go -t f "_test\.go$"

# 2. Check go.mod for framework versions before the update
echo -e "\n=== Previous versions in go.mod ==="
git show HEAD^:go.mod | grep -A 1 "github.com/onsi/ginkgo/v2"

Length of output: 22505


Script:

#!/bin/bash
# Let's check for any breaking changes in the test files by looking for specific Ginkgo/Gomega syntax patterns

# 1. Check for Ginkgo syntax that might have changed between v2.21.0 and v2.22.0
echo "=== Checking Ginkgo syntax patterns ==="
rg -l "ginkgo\." --type go

# 2. Check for Gomega syntax that might have changed between v1.35.1 and v1.36.0
echo -e "\n=== Checking Gomega syntax patterns ==="
rg -l "gomega\." --type go

# 3. Check imports to see how widely these frameworks are used
echo -e "\n=== Checking framework imports ==="
rg "^import.*github.com/onsi/(ginkgo|gomega)" --type go

Length of output: 384


381-381: Notable dependency updates requiring attention

Updates to critical dependencies:

  • github.com/quic-go/quic-go v0.48.2
  • github.com/stretchr/testify v1.9.0
  • github.com/unum-cloud/usearch/golang updated to latest version

These updates include major version changes that might introduce breaking changes.

Also applies to: 383-384

✅ Verification successful

The previous results show QUIC usage in the codebase, but we need to understand the exact version changes and potential breaking changes. This will help us provide a more accurate assessment of the impact.


Based on the git diff results, I can see that this is a minor version update of dependencies:

  • quic-go from v0.48.1 to v0.48.2 (patch version)
  • testify from v1.9.0 to v1.10.0 (minor version)
  • usearch/golang from 20241104182314-dccdd8e4152d to 20241121085649-2de3e34eea92 (commit hash update)

The code search shows QUIC usage is limited to connection handling in:

  • internal/net/quic/conn.go
  • internal/net/quic/listener.go
  • internal/net/dialer.go

The changes appear to be minor version updates that shouldn't introduce breaking changes. The QUIC usage in the codebase is focused on basic connection handling which is unlikely to be affected by a patch version update.

Dependencies updated with no breaking changes detected

The version changes are minor/patch updates and the codebase usage patterns appear compatible with the updates.

🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Check for potential breaking changes in usage

# Search for QUIC usage patterns that might be affected
rg -A 5 'quic\.(Config|Dial|Listen)' --type go

# Search for testify assertions that might need updates
rg -A 3 'assert\.(Equal|NotEqual|True|False)' --type go

Length of output: 1621


Script:

#!/bin/bash
# Let's check the git diff of go.mod to see the exact version changes
git diff origin/main -- go.mod | grep -A 1 -B 1 'quic-go\|testify\|usearch'

# Check for any QUIC-specific type usage that might be affected
rg 'quic\.(Stream|Connection|Transport|Session)' --type go

# Let's also check the changelog or release notes in vendor directory
fd -t f CHANGELOG -e md | xargs cat

Length of output: 66808

@@ -54,7 +54,7 @@ jobs:
run: |
make coverage
- name: Upload coverage report to Codecov
uses: codecov/codecov-action@v4
uses: codecov/codecov-action@v5
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

🛠️ Refactor suggestion

Consider pinning the action to a specific SHA

For security reasons, it's recommended to pin GitHub Actions to specific commit SHAs instead of using major version tags.

-        uses: codecov/codecov-action@v5
+        uses: codecov/codecov-action@e0b68c0bf36cfa71c264a216c38a96ce5d89165f  # v5.0.7
📝 Committable suggestion

‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.

Suggested change
uses: codecov/codecov-action@v5
uses: codecov/codecov-action@e0b68c0bf36cfa71c264a216c38a96ce5d89165f # v5.0.7

💡 Codebase verification

Action required: Update deprecated parameters in Codecov Action v5

The workflow is using the deprecated file parameter which needs to be updated to files as per v5 breaking changes. Here are the required changes:

  • Replace file: ./coverage.out with files: ./coverage.out
  • Consider pinning to a specific SHA for security: codecov/[email protected]

Additional notes:

  • The workflow is using a token which is good practice even with the new tokenless feature
  • No other deprecated parameters are in use
🔗 Analysis chain

Verify compatibility with Codecov Action v5

The update from v4 to v5 of codecov/codecov-action is a major version change that might include breaking changes.


Update workflow to handle breaking changes in Codecov Action v5

The update to v5 introduces several breaking changes that need to be addressed:

  • The file parameter has been deprecated in favor of files
  • The plugin parameter has been deprecated in favor of plugins
  • New tokenless upload feature for public repositories is now available

Let's verify the current usage:

🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Check for breaking changes between v4 and v5 of codecov/codecov-action

# Get the changelog or release notes for v5
gh api repos/codecov/codecov-action/releases | jq '.[] | select(.tag_name | startswith("v5.")) | {tag_name, body}'

# Check current usage in other workflow files
rg -l "codecov/codecov-action@" .github/workflows/

Length of output: 14443


Script:

#!/bin/bash
# Check current parameters used with codecov action
cat .github/workflows/coverage.yaml

Length of output: 2147

@@ -25,4 +25,4 @@ metadata:
app.kubernetes.io/version: v1.7.14
app.kubernetes.io/component: index-operator
data:
config.yaml: "---\nversion: v0.0.0\ntime_zone: UTC\nlogging:\n format: raw\n level: debug\n logger: glg\nserver_config:\n servers:\n - name: grpc\n host: 0.0.0.0\n port: 8081\n grpc:\n bidirectional_stream_concurrency: 20\n connection_timeout: \"\"\n enable_admin: true\n enable_channelz: true\n enable_reflection: true\n header_table_size: 0\n initial_conn_window_size: 2097152\n initial_window_size: 1048576\n interceptors:\n - RecoverInterceptor\n keepalive:\n max_conn_age: \"\"\n max_conn_age_grace: \"\"\n max_conn_idle: \"\"\n min_time: 10m\n permit_without_stream: false\n time: 3h\n timeout: 60s\n max_concurrent_streams: 0\n max_header_list_size: 0\n max_receive_message_size: 0\n max_send_message_size: 0\n num_stream_workers: 0\n read_buffer_size: 0\n shared_write_buffer: false\n wait_for_handlers: true\n write_buffer_size: 0\n mode: GRPC\n network: tcp\n probe_wait_time: 3s\n restart: true\n socket_option:\n ip_recover_destination_addr: false\n ip_transparent: false\n reuse_addr: true\n reuse_port: true\n tcp_cork: false\n tcp_defer_accept: false\n tcp_fast_open: false\n tcp_no_delay: false\n tcp_quick_ack: false\n socket_path: \"\"\n health_check_servers:\n - name: liveness\n host: 0.0.0.0\n port: 3000\n http:\n handler_timeout: \"\"\n http2:\n enabled: false\n handler_limit: 0\n max_concurrent_streams: 0\n max_decoder_header_table_size: 4096\n max_encoder_header_table_size: 4096\n max_read_frame_size: 0\n max_upload_buffer_per_connection: 0\n max_upload_buffer_per_stream: 0\n permit_prohibited_cipher_suites: true\n idle_timeout: \"\"\n read_header_timeout: \"\"\n read_timeout: \"\"\n shutdown_duration: 5s\n write_timeout: \"\"\n mode: REST\n network: tcp\n probe_wait_time: 3s\n restart: true\n socket_option:\n ip_recover_destination_addr: false\n ip_transparent: false\n reuse_addr: true\n reuse_port: true\n tcp_cork: false\n tcp_defer_accept: false\n tcp_fast_open: true\n tcp_no_delay: true\n tcp_quick_ack: true\n socket_path: \"\"\n - name: readiness\n host: 0.0.0.0\n port: 3001\n http:\n handler_timeout: \"\"\n http2:\n enabled: false\n handler_limit: 0\n max_concurrent_streams: 0\n max_decoder_header_table_size: 4096\n max_encoder_header_table_size: 4096\n max_read_frame_size: 0\n max_upload_buffer_per_connection: 0\n max_upload_buffer_per_stream: 0\n permit_prohibited_cipher_suites: true\n idle_timeout: \"\"\n read_header_timeout: \"\"\n read_timeout: \"\"\n shutdown_duration: 0s\n write_timeout: \"\"\n mode: REST\n network: tcp\n probe_wait_time: 3s\n restart: true\n socket_option:\n ip_recover_destination_addr: false\n ip_transparent: false\n reuse_addr: true\n reuse_port: true\n tcp_cork: false\n tcp_defer_accept: false\n tcp_fast_open: true\n tcp_no_delay: true\n tcp_quick_ack: true\n socket_path: \"\"\n metrics_servers:\n - name: pprof\n host: 0.0.0.0\n port: 6060\n http:\n handler_timeout: 5s\n http2:\n enabled: false\n handler_limit: 0\n max_concurrent_streams: 0\n max_decoder_header_table_size: 4096\n max_encoder_header_table_size: 4096\n max_read_frame_size: 0\n max_upload_buffer_per_connection: 0\n max_upload_buffer_per_stream: 0\n permit_prohibited_cipher_suites: true\n idle_timeout: 2s\n read_header_timeout: 1s\n read_timeout: 1s\n shutdown_duration: 5s\n write_timeout: 1m\n mode: REST\n network: tcp\n probe_wait_time: 3s\n restart: true\n socket_option:\n ip_recover_destination_addr: false\n ip_transparent: false\n reuse_addr: true\n reuse_port: true\n tcp_cork: true\n tcp_defer_accept: false\n tcp_fast_open: false\n tcp_no_delay: false\n tcp_quick_ack: false\n socket_path: \"\"\n startup_strategy:\n - liveness\n - pprof\n - grpc\n - readiness\n shutdown_strategy:\n - readiness\n - grpc\n - pprof\n - liveness\n full_shutdown_duration: 600s\n tls:\n ca: /path/to/ca\n cert: /path/to/cert\n enabled: false\n insecure_skip_verify: false\n key: /path/to/key\nobservability:\n enabled: false\n otlp:\n collector_endpoint: \"\"\n trace_batch_timeout: \"1s\"\n trace_export_timeout: \"1m\"\n trace_max_export_batch_size: 1024\n trace_max_queue_size: 256\n metrics_export_interval: \"1s\"\n metrics_export_timeout: \"1m\"\n attribute:\n namespace: \"_MY_POD_NAMESPACE_\"\n pod_name: \"_MY_POD_NAME_\"\n node_name: \"_MY_NODE_NAME_\"\n service_name: \"vald-index-operator\"\n metrics:\n enable_cgo: true\n enable_goroutine: true\n enable_memory: true\n enable_version_info: true\n version_info_labels:\n - vald_version\n - server_name\n - git_commit\n - build_time\n - go_version\n - go_os\n - go_arch\n - algorithm_info\n trace:\n enabled: false\noperator:\n namespace: _MY_POD_NAMESPACE_\n agent_name: vald-agent\n agent_namespace: \n rotator_name: vald-readreplica-rotate\n target_read_replica_id_annotations_key: vald.vdaas.org/target-read-replica-id\n rotation_job_concurrency: 2\n read_replica_enabled: false\n read_replica_label_key: vald-readreplica-id\n job_templates:\n rotate:\n apiVersion: batch/v1\n kind: Job\n metadata:\n name: vald-readreplica-rotate\n labels:\n app: vald-readreplica-rotate\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-readreplica-rotate\n app.kubernetes.io/version: v1.7.14\n spec:\n ttlSecondsAfterFinished: 86400\n template:\n metadata:\n labels:\n app: vald-readreplica-rotate\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-readreplica-rotate\n app.kubernetes.io/version: v1.7.14\n annotations:\n pyroscope.io/scrape: \"true\"\n pyroscope.io/application-name: vald-readreplica-rotate\n pyroscope.io/profile-cpu-enabled: \"true\"\n pyroscope.io/profile-mem-enabled: \"true\"\n pyroscope.io/port: \"6060\"\n spec:\n containers:\n - name: vald-readreplica-rotate\n image: \"vdaas/vald-readreplica-rotate:nightly\"\n imagePullPolicy: Always\n volumeMounts:\n - name: vald-readreplica-rotate-config\n mountPath: /etc/server/\n livenessProbe:\n failureThreshold: 2\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n readinessProbe:\n failureThreshold: 2\n httpGet:\n path: /readiness\n port: readiness\n scheme: HTTP\n initialDelaySeconds: 10\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n startupProbe:\n failureThreshold: 30\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 5\n successThreshold: 1\n timeoutSeconds: 2\n ports:\n - name: liveness\n protocol: TCP\n containerPort: 3000\n - name: readiness\n protocol: TCP\n containerPort: 3001\n - name: grpc\n protocol: TCP\n containerPort: 8081\n - name: pprof\n protocol: TCP\n containerPort: 6060\n securityContext:\n allowPrivilegeEscalation: false\n capabilities:\n drop:\n - ALL\n privileged: false\n readOnlyRootFilesystem: true\n runAsGroup: 65532\n runAsNonRoot: true\n runAsUser: 65532\n env:\n - name: MY_NODE_NAME\n valueFrom:\n fieldRef:\n fieldPath: spec.nodeName\n - name: MY_POD_NAME\n valueFrom:\n fieldRef:\n fieldPath: metadata.name\n - name: MY_POD_NAMESPACE\n valueFrom:\n fieldRef:\n fieldPath: metadata.namespace\n - name: TARGET_READREPLICA_ID_RELEASE_NAME_DEFAULT_VALD\n valueFrom:\n fieldRef:\n fieldPath: metadata.annotations['vald.vdaas.org/target-read-replica-id']\n securityContext:\n fsGroup: 65532\n fsGroupChangePolicy: OnRootMismatch\n runAsGroup: 65532\n runAsNonRoot: true\n runAsUser: 65532\n restartPolicy: OnFailure\n volumes:\n - name: vald-readreplica-rotate-config\n configMap:\n defaultMode: 420\n name: vald-readreplica-rotate-config\n serviceAccountName: vald-readreplica-rotate\n creation:\n apiVersion: batch/v1\n kind: Job\n metadata:\n name: vald-index-creation\n labels:\n app: vald-index-creation\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-index-creation\n app.kubernetes.io/version: v1.7.14\n spec:\n ttlSecondsAfterFinished: 86400\n template:\n metadata:\n labels:\n app: vald-index-creation\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-index-creation\n app.kubernetes.io/version: v1.7.14\n annotations:\n pyroscope.io/scrape: \"true\"\n pyroscope.io/application-name: vald-index-creation\n pyroscope.io/profile-cpu-enabled: \"true\"\n pyroscope.io/profile-mem-enabled: \"true\"\n pyroscope.io/port: \"6060\"\n spec:\n initContainers:\n - name: wait-for-agent\n image: busybox:stable\n imagePullPolicy: Always\n command:\n - /bin/sh\n - -e\n - -c\n - |\n until [ \"$(wget --server-response --spider --quiet http://vald-agent.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')\" == \"200\" ]; do\n echo \"waiting for agent to be ready...\"\n sleep 2;\n done\n - name: wait-for-discoverer\n image: busybox:stable\n imagePullPolicy: Always\n command:\n - /bin/sh\n - -e\n - -c\n - |\n until [ \"$(wget --server-response --spider --quiet http://vald-discoverer.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')\" == \"200\" ]; do\n echo \"waiting for discoverer to be ready...\"\n sleep 2;\n done\n containers:\n - name: vald-index-creation\n image: \"vdaas/vald-index-creation:nightly\"\n imagePullPolicy: Always\n volumeMounts:\n - name: vald-index-creation-config\n mountPath: /etc/server/\n livenessProbe:\n failureThreshold: 2\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n readinessProbe:\n failureThreshold: 2\n httpGet:\n path: /readiness\n port: readiness\n scheme: HTTP\n initialDelaySeconds: 10\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n startupProbe:\n failureThreshold: 30\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 5\n successThreshold: 1\n timeoutSeconds: 2\n ports:\n - name: liveness\n protocol: TCP\n containerPort: 3000\n - name: readiness\n protocol: TCP\n containerPort: 3001\n - name: grpc\n protocol: TCP\n containerPort: 8081\n - name: pprof\n protocol: TCP\n containerPort: 6060\n env:\n - name: MY_NODE_NAME\n valueFrom:\n fieldRef:\n fieldPath: spec.nodeName\n - name: MY_POD_NAME\n valueFrom:\n fieldRef:\n fieldPath: metadata.name\n - name: MY_POD_NAMESPACE\n valueFrom:\n fieldRef:\n fieldPath: metadata.namespace\n restartPolicy: OnFailure\n volumes:\n - name: vald-index-creation-config\n configMap:\n defaultMode: 420\n name: vald-index-creation-config\n save:\n apiVersion: batch/v1\n kind: Job\n metadata:\n name: vald-index-save\n labels:\n app: vald-index-save\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-index-save\n app.kubernetes.io/version: v1.7.14\n spec:\n ttlSecondsAfterFinished: 86400\n template:\n metadata:\n labels:\n app: vald-index-save\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-index-save\n app.kubernetes.io/version: v1.7.14\n annotations:\n pyroscope.io/scrape: \"true\"\n pyroscope.io/application-name: vald-index-save\n pyroscope.io/profile-cpu-enabled: \"true\"\n pyroscope.io/profile-mem-enabled: \"true\"\n pyroscope.io/port: \"6060\"\n spec:\n initContainers:\n - name: wait-for-agent\n image: busybox:stable\n imagePullPolicy: Always\n command:\n - /bin/sh\n - -e\n - -c\n - |\n until [ \"$(wget --server-response --spider --quiet http://vald-agent.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')\" == \"200\" ]; do\n echo \"waiting for agent to be ready...\"\n sleep 2;\n done\n - name: wait-for-discoverer\n image: busybox:stable\n imagePullPolicy: Always\n command:\n - /bin/sh\n - -e\n - -c\n - |\n until [ \"$(wget --server-response --spider --quiet http://vald-discoverer.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')\" == \"200\" ]; do\n echo \"waiting for discoverer to be ready...\"\n sleep 2;\n done\n containers:\n - name: vald-index-save\n image: \"vdaas/vald-index-save:nightly\"\n imagePullPolicy: Always\n volumeMounts:\n - name: vald-index-save-config\n mountPath: /etc/server/\n livenessProbe:\n failureThreshold: 2\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n readinessProbe:\n failureThreshold: 2\n httpGet:\n path: /readiness\n port: readiness\n scheme: HTTP\n initialDelaySeconds: 10\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n startupProbe:\n failureThreshold: 30\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 5\n successThreshold: 1\n timeoutSeconds: 2\n ports:\n - name: liveness\n protocol: TCP\n containerPort: 3000\n - name: readiness\n protocol: TCP\n containerPort: 3001\n - name: grpc\n protocol: TCP\n containerPort: 8081\n - name: pprof\n protocol: TCP\n containerPort: 6060\n env:\n - name: MY_NODE_NAME\n valueFrom:\n fieldRef:\n fieldPath: spec.nodeName\n - name: MY_POD_NAME\n valueFrom:\n fieldRef:\n fieldPath: metadata.name\n - name: MY_POD_NAMESPACE\n valueFrom:\n fieldRef:\n fieldPath: metadata.namespace\n restartPolicy: OnFailure\n volumes:\n - name: vald-index-save-config\n configMap:\n defaultMode: 420\n name: vald-index-save-config\n correction:\n apiVersion: batch/v1\n kind: Job\n metadata:\n name: vald-index-correction\n labels:\n app: vald-index-correction\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-index-correction\n app.kubernetes.io/version: v1.7.14\n spec:\n ttlSecondsAfterFinished: 86400\n template:\n metadata:\n labels:\n app: vald-index-correction\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-index-correction\n app.kubernetes.io/version: v1.7.14\n annotations:\n pyroscope.io/scrape: \"true\"\n pyroscope.io/application-name: vald-index-correction\n pyroscope.io/profile-cpu-enabled: \"true\"\n pyroscope.io/profile-mem-enabled: \"true\"\n pyroscope.io/port: \"6060\"\n spec:\n initContainers:\n - name: wait-for-agent\n image: busybox:stable\n imagePullPolicy: Always\n command:\n - /bin/sh\n - -e\n - -c\n - |\n until [ \"$(wget --server-response --spider --quiet http://vald-agent.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')\" == \"200\" ]; do\n echo \"waiting for agent to be ready...\"\n sleep 2;\n done\n - name: wait-for-discoverer\n image: busybox:stable\n imagePullPolicy: Always\n command:\n - /bin/sh\n - -e\n - -c\n - |\n until [ \"$(wget --server-response --spider --quiet http://vald-discoverer.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')\" == \"200\" ]; do\n echo \"waiting for discoverer to be ready...\"\n sleep 2;\n done\n containers:\n - name: vald-index-correction\n image: \"vdaas/vald-index-correction:nightly\"\n imagePullPolicy: Always\n volumeMounts:\n - name: vald-index-correction-config\n mountPath: /etc/server/\n livenessProbe:\n failureThreshold: 2\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n readinessProbe:\n failureThreshold: 2\n httpGet:\n path: /readiness\n port: readiness\n scheme: HTTP\n initialDelaySeconds: 10\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n startupProbe:\n failureThreshold: 30\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 5\n successThreshold: 1\n timeoutSeconds: 2\n ports:\n - name: liveness\n protocol: TCP\n containerPort: 3000\n - name: readiness\n protocol: TCP\n containerPort: 3001\n - name: grpc\n protocol: TCP\n containerPort: 8081\n - name: pprof\n protocol: TCP\n containerPort: 6060\n env:\n - name: MY_NODE_NAME\n valueFrom:\n fieldRef:\n fieldPath: spec.nodeName\n - name: MY_POD_NAME\n valueFrom:\n fieldRef:\n fieldPath: metadata.name\n - name: MY_POD_NAMESPACE\n valueFrom:\n fieldRef:\n fieldPath: metadata.namespace\n restartPolicy: OnFailure\n volumes:\n - name: vald-index-correction-config\n configMap:\n defaultMode: 420\n name: vald-index-correction-config\n"
config.yaml: "---\nversion: v0.0.0\ntime_zone: UTC\nlogging:\n format: raw\n level: debug\n logger: glg\nserver_config:\n servers:\n - name: grpc\n host: 0.0.0.0\n port: 8081\n grpc:\n bidirectional_stream_concurrency: 20\n connection_timeout: \"\"\n enable_admin: true\n enable_channelz: true\n enable_reflection: true\n header_table_size: 0\n initial_conn_window_size: 2097152\n initial_window_size: 1048576\n interceptors:\n - RecoverInterceptor\n keepalive:\n max_conn_age: \"\"\n max_conn_age_grace: \"\"\n max_conn_idle: \"\"\n min_time: 10m\n permit_without_stream: false\n time: 3h\n timeout: 60s\n max_concurrent_streams: 0\n max_header_list_size: 0\n max_receive_message_size: 0\n max_send_message_size: 0\n num_stream_workers: 0\n read_buffer_size: 0\n shared_write_buffer: false\n wait_for_handlers: true\n write_buffer_size: 0\n mode: GRPC\n network: tcp\n probe_wait_time: 3s\n restart: true\n socket_option:\n ip_recover_destination_addr: false\n ip_transparent: false\n reuse_addr: true\n reuse_port: true\n tcp_cork: false\n tcp_defer_accept: false\n tcp_fast_open: false\n tcp_no_delay: false\n tcp_quick_ack: false\n socket_path: \"\"\n health_check_servers:\n - name: liveness\n host: 0.0.0.0\n port: 3000\n http:\n handler_timeout: \"\"\n http2:\n enabled: false\n handler_limit: 0\n max_concurrent_streams: 0\n max_decoder_header_table_size: 4096\n max_encoder_header_table_size: 4096\n max_read_frame_size: 0\n max_upload_buffer_per_connection: 0\n max_upload_buffer_per_stream: 0\n permit_prohibited_cipher_suites: true\n idle_timeout: \"\"\n read_header_timeout: \"\"\n read_timeout: \"\"\n shutdown_duration: 5s\n write_timeout: \"\"\n mode: REST\n network: tcp\n probe_wait_time: 3s\n restart: true\n socket_option:\n ip_recover_destination_addr: false\n ip_transparent: false\n reuse_addr: true\n reuse_port: true\n tcp_cork: false\n tcp_defer_accept: false\n tcp_fast_open: true\n tcp_no_delay: true\n tcp_quick_ack: true\n socket_path: \"\"\n - name: readiness\n host: 0.0.0.0\n port: 3001\n http:\n handler_timeout: \"\"\n http2:\n enabled: false\n handler_limit: 0\n max_concurrent_streams: 0\n max_decoder_header_table_size: 4096\n max_encoder_header_table_size: 4096\n max_read_frame_size: 0\n max_upload_buffer_per_connection: 0\n max_upload_buffer_per_stream: 0\n permit_prohibited_cipher_suites: true\n idle_timeout: \"\"\n read_header_timeout: \"\"\n read_timeout: \"\"\n shutdown_duration: 0s\n write_timeout: \"\"\n mode: REST\n network: tcp\n probe_wait_time: 3s\n restart: true\n socket_option:\n ip_recover_destination_addr: false\n ip_transparent: false\n reuse_addr: true\n reuse_port: true\n tcp_cork: false\n tcp_defer_accept: false\n tcp_fast_open: true\n tcp_no_delay: true\n tcp_quick_ack: true\n socket_path: \"\"\n metrics_servers:\n - name: pprof\n host: 0.0.0.0\n port: 6060\n http:\n handler_timeout: 5s\n http2:\n enabled: false\n handler_limit: 0\n max_concurrent_streams: 0\n max_decoder_header_table_size: 4096\n max_encoder_header_table_size: 4096\n max_read_frame_size: 0\n max_upload_buffer_per_connection: 0\n max_upload_buffer_per_stream: 0\n permit_prohibited_cipher_suites: true\n idle_timeout: 2s\n read_header_timeout: 1s\n read_timeout: 1s\n shutdown_duration: 5s\n write_timeout: 1m\n mode: REST\n network: tcp\n probe_wait_time: 3s\n restart: true\n socket_option:\n ip_recover_destination_addr: false\n ip_transparent: false\n reuse_addr: true\n reuse_port: true\n tcp_cork: true\n tcp_defer_accept: false\n tcp_fast_open: false\n tcp_no_delay: false\n tcp_quick_ack: false\n socket_path: \"\"\n startup_strategy:\n - liveness\n - pprof\n - grpc\n - readiness\n shutdown_strategy:\n - readiness\n - grpc\n - pprof\n - liveness\n full_shutdown_duration: 600s\n tls:\n ca: /path/to/ca\n cert: /path/to/cert\n enabled: false\n insecure_skip_verify: false\n key: /path/to/key\nobservability:\n enabled: false\n otlp:\n collector_endpoint: \"\"\n trace_batch_timeout: \"1s\"\n trace_export_timeout: \"1m\"\n trace_max_export_batch_size: 1024\n trace_max_queue_size: 256\n metrics_export_interval: \"1s\"\n metrics_export_timeout: \"1m\"\n attribute:\n namespace: \"_MY_POD_NAMESPACE_\"\n pod_name: \"_MY_POD_NAME_\"\n node_name: \"_MY_NODE_NAME_\"\n service_name: \"vald-index-operator\"\n metrics:\n enable_cgo: true\n enable_goroutine: true\n enable_memory: true\n enable_version_info: true\n version_info_labels:\n - vald_version\n - server_name\n - git_commit\n - build_time\n - go_version\n - go_os\n - go_arch\n - algorithm_info\n trace:\n enabled: false\noperator:\n namespace: _MY_POD_NAMESPACE_\n agent_name: vald-agent\n agent_namespace: \n rotator_name: vald-readreplica-rotate\n target_read_replica_id_annotations_key: vald.vdaas.org/target-read-replica-id\n rotation_job_concurrency: 2\n read_replica_enabled: false\n read_replica_label_key: vald-readreplica-id\n job_templates:\n rotate:\n apiVersion: batch/v1\n kind: Job\n metadata:\n name: vald-readreplica-rotate\n labels:\n app: vald-readreplica-rotate\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-readreplica-rotate\n app.kubernetes.io/version: v1.7.14\n spec:\n ttlSecondsAfterFinished: 86400\n template:\n metadata:\n labels:\n app: vald-readreplica-rotate\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-readreplica-rotate\n app.kubernetes.io/version: v1.7.14\n annotations:\n pyroscope.io/scrape: \"true\"\n pyroscope.io/application-name: vald-readreplica-rotate\n pyroscope.io/profile-cpu-enabled: \"true\"\n pyroscope.io/profile-mem-enabled: \"true\"\n pyroscope.io/port: \"6060\"\n spec:\n containers:\n - name: vald-readreplica-rotate\n image: \"vdaas/vald-readreplica-rotate:nightly\"\n imagePullPolicy: Always\n volumeMounts:\n - name: vald-readreplica-rotate-config\n mountPath: /etc/server/\n livenessProbe:\n failureThreshold: 2\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n readinessProbe:\n failureThreshold: 2\n httpGet:\n path: /readiness\n port: readiness\n scheme: HTTP\n initialDelaySeconds: 10\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n startupProbe:\n failureThreshold: 30\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 5\n successThreshold: 1\n timeoutSeconds: 2\n ports:\n - name: liveness\n protocol: TCP\n containerPort: 3000\n - name: readiness\n protocol: TCP\n containerPort: 3001\n - name: grpc\n protocol: TCP\n containerPort: 8081\n - name: pprof\n protocol: TCP\n containerPort: 6060\n securityContext:\n allowPrivilegeEscalation: false\n capabilities:\n drop:\n - ALL\n privileged: false\n readOnlyRootFilesystem: true\n runAsGroup: 65532\n runAsNonRoot: true\n runAsUser: 65532\n env:\n - name: MY_NODE_NAME\n valueFrom:\n fieldRef:\n fieldPath: spec.nodeName\n - name: MY_POD_NAME\n valueFrom:\n fieldRef:\n fieldPath: metadata.name\n - name: MY_POD_NAMESPACE\n valueFrom:\n fieldRef:\n fieldPath: metadata.namespace\n - name: TARGET_READREPLICA_ID_RELEASE_NAME_DEFAULT_VALD\n valueFrom:\n fieldRef:\n fieldPath: metadata.annotations['vald.vdaas.org/target-read-replica-id']\n securityContext:\n fsGroup: 65532\n fsGroupChangePolicy: OnRootMismatch\n runAsGroup: 65532\n runAsNonRoot: true\n runAsUser: 65532\n restartPolicy: OnFailure\n volumes:\n - name: vald-readreplica-rotate-config\n configMap:\n defaultMode: 420\n name: vald-readreplica-rotate-config\n serviceAccountName: vald-readreplica-rotate\n creation:\n apiVersion: batch/v1\n kind: Job\n metadata:\n name: vald-index-creation\n labels:\n app: vald-index-creation\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-index-creation\n app.kubernetes.io/version: v1.7.14\n spec:\n ttlSecondsAfterFinished: 86400\n template:\n metadata:\n labels:\n app: vald-index-creation\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-index-creation\n app.kubernetes.io/version: v1.7.14\n annotations:\n pyroscope.io/scrape: \"true\"\n pyroscope.io/application-name: vald-index-creation\n pyroscope.io/profile-cpu-enabled: \"true\"\n pyroscope.io/profile-mem-enabled: \"true\"\n pyroscope.io/port: \"6060\"\n spec:\n initContainers:\n - name: wait-for-agent\n image: busybox:stable\n imagePullPolicy: Always\n command:\n - /bin/sh\n - -e\n - -c\n - |\n until [ \"$(wget --server-response --spider --quiet http://vald-agent.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')\" == \"200\" ]; do\n echo \"waiting for agent to be ready...\"\n sleep 2;\n done\n - name: wait-for-discoverer\n image: busybox:stable\n imagePullPolicy: Always\n command:\n - /bin/sh\n - -e\n - -c\n - |\n until [ \"$(wget --server-response --spider --quiet http://vald-discoverer.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')\" == \"200\" ]; do\n echo \"waiting for discoverer to be ready...\"\n sleep 2;\n done\n affinity:\n nodeAffinity:\n preferredDuringSchedulingIgnoredDuringExecution: []\n podAffinity:\n preferredDuringSchedulingIgnoredDuringExecution: []\n requiredDuringSchedulingIgnoredDuringExecution: []\n podAntiAffinity:\n preferredDuringSchedulingIgnoredDuringExecution: []\n requiredDuringSchedulingIgnoredDuringExecution: []\n containers:\n - name: vald-index-creation\n image: \"vdaas/vald-index-creation:nightly\"\n imagePullPolicy: Always\n volumeMounts:\n - name: vald-index-creation-config\n mountPath: /etc/server/\n livenessProbe:\n failureThreshold: 2\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n readinessProbe:\n failureThreshold: 2\n httpGet:\n path: /readiness\n port: readiness\n scheme: HTTP\n initialDelaySeconds: 10\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n startupProbe:\n failureThreshold: 30\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 5\n successThreshold: 1\n timeoutSeconds: 2\n ports:\n - name: liveness\n protocol: TCP\n containerPort: 3000\n - name: readiness\n protocol: TCP\n containerPort: 3001\n - name: grpc\n protocol: TCP\n containerPort: 8081\n - name: pprof\n protocol: TCP\n containerPort: 6060\n env:\n - name: MY_NODE_NAME\n valueFrom:\n fieldRef:\n fieldPath: spec.nodeName\n - name: MY_POD_NAME\n valueFrom:\n fieldRef:\n fieldPath: metadata.name\n - name: MY_POD_NAMESPACE\n valueFrom:\n fieldRef:\n fieldPath: metadata.namespace\n restartPolicy: OnFailure\n volumes:\n - name: vald-index-creation-config\n configMap:\n defaultMode: 420\n name: vald-index-creation-config\n save:\n apiVersion: batch/v1\n kind: Job\n metadata:\n name: vald-index-save\n labels:\n app: vald-index-save\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-index-save\n app.kubernetes.io/version: v1.7.14\n spec:\n ttlSecondsAfterFinished: 86400\n template:\n metadata:\n labels:\n app: vald-index-save\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-index-save\n app.kubernetes.io/version: v1.7.14\n annotations:\n pyroscope.io/scrape: \"true\"\n pyroscope.io/application-name: vald-index-save\n pyroscope.io/profile-cpu-enabled: \"true\"\n pyroscope.io/profile-mem-enabled: \"true\"\n pyroscope.io/port: \"6060\"\n spec:\n initContainers:\n - name: wait-for-agent\n image: busybox:stable\n imagePullPolicy: Always\n command:\n - /bin/sh\n - -e\n - -c\n - |\n until [ \"$(wget --server-response --spider --quiet http://vald-agent.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')\" == \"200\" ]; do\n echo \"waiting for agent to be ready...\"\n sleep 2;\n done\n - name: wait-for-discoverer\n image: busybox:stable\n imagePullPolicy: Always\n command:\n - /bin/sh\n - -e\n - -c\n - |\n until [ \"$(wget --server-response --spider --quiet http://vald-discoverer.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')\" == \"200\" ]; do\n echo \"waiting for discoverer to be ready...\"\n sleep 2;\n done\n affinity:\n nodeAffinity:\n preferredDuringSchedulingIgnoredDuringExecution: []\n podAffinity:\n preferredDuringSchedulingIgnoredDuringExecution: []\n requiredDuringSchedulingIgnoredDuringExecution: []\n podAntiAffinity:\n preferredDuringSchedulingIgnoredDuringExecution: []\n requiredDuringSchedulingIgnoredDuringExecution: []\n containers:\n - name: vald-index-save\n image: \"vdaas/vald-index-save:nightly\"\n imagePullPolicy: Always\n volumeMounts:\n - name: vald-index-save-config\n mountPath: /etc/server/\n livenessProbe:\n failureThreshold: 2\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n readinessProbe:\n failureThreshold: 2\n httpGet:\n path: /readiness\n port: readiness\n scheme: HTTP\n initialDelaySeconds: 10\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n startupProbe:\n failureThreshold: 30\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 5\n successThreshold: 1\n timeoutSeconds: 2\n ports:\n - name: liveness\n protocol: TCP\n containerPort: 3000\n - name: readiness\n protocol: TCP\n containerPort: 3001\n - name: grpc\n protocol: TCP\n containerPort: 8081\n - name: pprof\n protocol: TCP\n containerPort: 6060\n env:\n - name: MY_NODE_NAME\n valueFrom:\n fieldRef:\n fieldPath: spec.nodeName\n - name: MY_POD_NAME\n valueFrom:\n fieldRef:\n fieldPath: metadata.name\n - name: MY_POD_NAMESPACE\n valueFrom:\n fieldRef:\n fieldPath: metadata.namespace\n restartPolicy: OnFailure\n volumes:\n - name: vald-index-save-config\n configMap:\n defaultMode: 420\n name: vald-index-save-config\n correction:\n apiVersion: batch/v1\n kind: Job\n metadata:\n name: vald-index-correction\n labels:\n app: vald-index-correction\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-index-correction\n app.kubernetes.io/version: v1.7.14\n spec:\n ttlSecondsAfterFinished: 86400\n template:\n metadata:\n labels:\n app: vald-index-correction\n app.kubernetes.io/name: vald\n helm.sh/chart: vald-v1.7.14\n app.kubernetes.io/managed-by: Helm\n app.kubernetes.io/instance: release-name\n app.kubernetes.io/component: vald-index-correction\n app.kubernetes.io/version: v1.7.14\n annotations:\n pyroscope.io/scrape: \"true\"\n pyroscope.io/application-name: vald-index-correction\n pyroscope.io/profile-cpu-enabled: \"true\"\n pyroscope.io/profile-mem-enabled: \"true\"\n pyroscope.io/port: \"6060\"\n spec:\n initContainers:\n - name: wait-for-agent\n image: busybox:stable\n imagePullPolicy: Always\n command:\n - /bin/sh\n - -e\n - -c\n - |\n until [ \"$(wget --server-response --spider --quiet http://vald-agent.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')\" == \"200\" ]; do\n echo \"waiting for agent to be ready...\"\n sleep 2;\n done\n - name: wait-for-discoverer\n image: busybox:stable\n imagePullPolicy: Always\n command:\n - /bin/sh\n - -e\n - -c\n - |\n until [ \"$(wget --server-response --spider --quiet http://vald-discoverer.default.svc.cluster.local:3001/readiness 2>&1 | awk 'NR==1{print $2}')\" == \"200\" ]; do\n echo \"waiting for discoverer to be ready...\"\n sleep 2;\n done\n affinity:\n nodeAffinity:\n preferredDuringSchedulingIgnoredDuringExecution: []\n podAffinity:\n preferredDuringSchedulingIgnoredDuringExecution: []\n requiredDuringSchedulingIgnoredDuringExecution: []\n podAntiAffinity:\n preferredDuringSchedulingIgnoredDuringExecution: []\n requiredDuringSchedulingIgnoredDuringExecution: []\n containers:\n - name: vald-index-correction\n image: \"vdaas/vald-index-correction:nightly\"\n imagePullPolicy: Always\n volumeMounts:\n - name: vald-index-correction-config\n mountPath: /etc/server/\n livenessProbe:\n failureThreshold: 2\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n readinessProbe:\n failureThreshold: 2\n httpGet:\n path: /readiness\n port: readiness\n scheme: HTTP\n initialDelaySeconds: 10\n periodSeconds: 3\n successThreshold: 1\n timeoutSeconds: 2\n startupProbe:\n failureThreshold: 30\n httpGet:\n path: /liveness\n port: liveness\n scheme: HTTP\n initialDelaySeconds: 5\n periodSeconds: 5\n successThreshold: 1\n timeoutSeconds: 2\n ports:\n - name: liveness\n protocol: TCP\n containerPort: 3000\n - name: readiness\n protocol: TCP\n containerPort: 3001\n - name: grpc\n protocol: TCP\n containerPort: 8081\n - name: pprof\n protocol: TCP\n containerPort: 6060\n env:\n - name: MY_NODE_NAME\n valueFrom:\n fieldRef:\n fieldPath: spec.nodeName\n - name: MY_POD_NAME\n valueFrom:\n fieldRef:\n fieldPath: metadata.name\n - name: MY_POD_NAMESPACE\n valueFrom:\n fieldRef:\n fieldPath: metadata.namespace\n restartPolicy: OnFailure\n volumes:\n - name: vald-index-correction-config\n configMap:\n defaultMode: 420\n name: vald-index-correction-config\n"
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codebase verification

Services exist but may require namespace configuration

The services vald-agent and vald-discoverer are properly defined in the Kubernetes manifests, but the configuration assumes they are in the default namespace. The initContainer's health checks in the configmap use hardcoded URLs with .default.svc.cluster.local, but the operator's namespace configuration is dynamic using _MY_POD_NAMESPACE_.

  • The services exist with correct ports (3001 for readiness)
  • The operator config uses dynamic namespace but initContainers assume default namespace
  • This mismatch could cause issues if deployed in non-default namespace
🔗 Analysis chain

Verify service dependencies

The configuration assumes the existence of several services:

  • vald-agent.default.svc.cluster.local
  • vald-discoverer.default.svc.cluster.local
🏁 Scripts executed

The following scripts were executed for the analysis:

Script:

#!/bin/bash
# Description: Verify service dependencies
# Check for service definitions
rg -l "kind:\s+Service.*name:\s+(vald-agent|vald-discoverer)" k8s/

Length of output: 68


Script:

#!/bin/bash
# Let's try a different approach to find service definitions
# Search for service definitions in yaml files
fd -e yaml -e yml . k8s/ -x rg -l "kind:\s+Service" {}

# Also check for references to these services in the codebase
rg -l "vald-(agent|discoverer).*svc\.cluster\.local"

Length of output: 2062


Script:

#!/bin/bash
# Let's examine the service definitions for vald-agent and vald-discoverer
cat k8s/agent/svc.yaml k8s/discoverer/svc.yaml

# Also check their deployment configurations
cat k8s/agent/deployment.yaml k8s/discoverer/deployment.yaml 2>/dev/null || echo "Deployment files not found"

Length of output: 8332

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment