You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Describe the bug
Tracetest default helm chart values resulting in the following error
"could not execute migrations: could not get driver from postgres connection: pq: password authentication failed for user "tracetest"
To Reproduce
Steps to reproduce the behavior:
Run helm install tracetest kubeshop/tracetest --values values.yaml (only changes postgres password and collector configs)
See error
Expected behavior
TraceTest running properly
Version / Cluster
Tracetest helm chart version 0.2.47
kubernetes v.1.25.4 AKS
Screenshots
Additional context
My values.yaml
# Section for configuring the postgres database that will be used by Tracetest
postgresql:
# For now, this is required to be enabled, otherwise tracetest will not
# be able to function properly.
enabled: true
architecture: standalone
image:
tag: 14.6.0-debian-11-r13
# credentials for accessing the database
auth:
database: "tracetest"
username: "tracetest"
password: test123
existingSecret: ""
# Provisioning allows to define initial settings to be loaded into the database on first run.
# These will only be applied when running tracetest against an empty database. If tracetest has already
# been configured, the provisioning settings are ignored.
provisioning: |
---
# Datastore is where your application stores its traces. You can define several different datastores with
# different names, however, only one is used by Tracetest.
# You can see all available datastore configurations at https://kubeshop.github.io/tracetest/supported-backends/
type: DataStore
spec:
name: otlp
# Indicates that this datastore is a jaeger instance
type: otlp
# Configures how tracetest connects to jaeger.
otlp:
type: otlp
---
type: Config
spec:
analyticsEnabled: false
---
type: PollingProfile
spec:
name: Custom Profile
strategy: periodic
default: true
periodic:
timeout: 2m
retryDelay: 3s
# This section configures the strategy for pooling traces from the trace backend
poolingConfig:
# How long tracetest will wait for a complete trace before timing out
# If you have long-running operations that will generate part of your trace, you have
# to change this attribute to be greater than the execution time of your operation.
maxWaitTimeForTrace: 30s
# How long tracetest will wait to retry fetching the trace. If you define it as 5s it means that
# tracetest will retrieve the operation trace every 5 seconds and check if it's complete. It will
# do that until the operation times out.
retryDelay: 5s
# Section for anonymous analytics. If it's enabled, tracetest will collect anonymous analytics data
# to help us improving our project. If you don't want to send analytics data, set enabled as false.
analytics:
enabled: true
# Section to configure how telemetry works in Tracetest.
telemetry:
# exporters are opentelemetry exporters. It configures how tracetest generates and send telemetry to
# a datastore.
exporters:
# This is an exporter called collector, but it could be named anything else.
collector:
# configures the service.name attribute in all trace spans generated by tracetest
serviceName: tracetest
# indicates the percentage of traces that would be sent to the datastore. 100 = 100%
sampling: 100
# configures the exporter
exporter:
# Tracetest sends data using the opentelemetry collector. For now there is no other
# alternative. But the collector is flexible enough to send traces to any other tracing backend
# you might need.
type: collector
collector:
# endpoint to send traces to the collector
endpoint: otel-collector.open-telemetry.svc.cluster.local:4317
# Configures the server
server:
# Indicates the port that Tracetest will run ons
httpPort: 11633
# Indicates which telemetry components will be used by Tracetest
telemetry:
# The exporter that tracetest will use to send telemetry
# exporter: collector
# The exporter that tracetest will send the triggering transaction span. This is optional. If you
# want to have the trigger transaction span in your trace, you have to configure this field to
# send the span to the same place your application sends its telemetry.
# applicationExporter: collector
replicaCount: 1
# You don't have to change anything bellow this line.
image:
repository: kubeshop/tracetest
pullPolicy: IfNotPresent
# Overrides the image tag whose default is the chart appVersion.
tag: ""
imagePullSecrets: []
nameOverride: ""
fullnameOverride: ""
serviceAccount:
# Specifies whether a service account should be created
create: false
# Annotations to add to the service account
annotations: {}
# The name of the service account to use.
# If not set and create is true, a name is generated using the fullname template
name: ""
podAnnotations: {}
podSecurityContext: {}
# fsGroup: 2000
securityContext: {}
# capabilities:
# drop:
# - ALL
# readOnlyRootFilesystem: true
# runAsNonRoot: true
# runAsUser: 1000
service:
type: ClusterIP
port: 11633
annotations: {}
ingress:
enabled: false
className: ""
annotations: {}
# kubernetes.io/ingress.class: nginx
# kubernetes.io/tls-acme: "true"
hosts:
- host: chart-example.local
paths:
- path: /
pathType: ImplementationSpecific
tls: []
# - secretName: chart-example-tls
# hosts:
# - chart-example.local
resources: {}
# We usually recommend not to specify default resources and to leave this as a conscious
# choice for the user. This also increases chances charts run on environments with little
# resources, such as Minikube. If you do want to specify resources, uncomment the following
# lines, adjust them as necessary, and remove the curly braces after 'resources:'.
# limits:
# cpu: 100m
# memory: 128Mi
# requests:
# cpu: 100m
# memory: 128Mi
autoscaling:
enabled: false
minReplicas: 1
maxReplicas: 100
targetCPUUtilizationPercentage: 80
# targetMemoryUtilizationPercentage: 80
nodeSelector: {}
tolerations: []
affinity: {}
The text was updated successfully, but these errors were encountered:
Describe the bug
Tracetest default helm chart values resulting in the following error
"could not execute migrations: could not get driver from postgres connection: pq: password authentication failed for user "tracetest"
To Reproduce
Steps to reproduce the behavior:
Expected behavior
TraceTest running properly
Version / Cluster
Screenshots
Additional context
My values.yaml
The text was updated successfully, but these errors were encountered: