kube-prometheus | Time Series Database library
kandi X-RAY | kube-prometheus Summary
kandi X-RAY | kube-prometheus Summary
kube-prometheus
Support
Quality
Security
License
Reuse
Top functions reviewed by kandi - BETA
Currently covering the most popular Java, JavaScript and Python libraries. See a Sample of kube-prometheus
kube-prometheus Key Features
kube-prometheus Examples and Code Snippets
Community Discussions
Trending Discussions on kube-prometheus
QUESTION
I would like to deploy the kube-prometheus-stack
helm chart on a GKE cluster using the following values:
ANSWER
Answered 2022-Apr-04 at 05:47The ClusterIssuer
has no bearing on whether the Grafana ingress is created or not. If your ingress is not getting an IP, you have no ingress controller configured on your cluster.
Alternatively do a kubectl describe
on the ingress and look at the events section at the bottom of the output. Sometimes, there will be messages there that might tell you where you need to look.
QUESTION
I have a metrics-server and a horizontal pod autoscaler using this server, running on my cluster.
This works perfectly fine, until i inject linkerd-proxies into the deployments of the namespace where my application is running. Running kubectl top pod
in that namespace results in a error: Metrics not available for pod
error. However, nothing appears in the metrics-server pod's logs.
The metrics-server clearly works fine in other namespaces, because top works in every namespace but the meshed one.
At first i thought it could be because the proxies' resource requests/limits weren't set, but after running the injection with them (kubectl get -n deploy -o yaml | linkerd inject - --proxy-cpu-request "10m" --proxy-cpu-limit "1" --proxy-memory-request "64Mi" --proxy-memory-limit "256Mi" | kubectl apply -f -
), the issue stays the same.
Is this a known problem, are there any possible solutions?
PS: I have a kube-prometheus-stack running in a different namespace, and this seems to be able to scrape the pod metrics from the meshed pods just fine
...ANSWER
Answered 2022-Mar-04 at 01:18I'm able to use kubectl top
on pods that have linkerd injected:
QUESTION
Basically, I had installed Prometheues-Grafana from the kube-prometheus-stack using the provided helm chart repo prometheus-community
...ANSWER
Answered 2022-Feb-24 at 15:46It's not enough to simply install them, you need to integrate prometheus
with thanos
.
Below I'll describe all steps you need to perform to get the result.
First short theory. The most common approach to integrate them is to use thanos sidecar
container for prometheus
pod. You can read more here.
How this is done:
(considering that installation is clean, it can be easily deleted and reinstalled from the scratch).
- Get
thanos sidecar
added to theprometheus
pod.
Pull kube-prometheus-stack
chart:
QUESTION
I am using kube-prometheus-stack to monitor my system in gcp. Due to new requirements all my ingress need to be secured with tls. As a first step I wanted to make the grafana webpage available via https.
I created a tls secret and updated my values.yaml. After helm upgrade everything seems to work fine but page is still available via http only.
Hope you can support me here.
...ANSWER
Answered 2022-Feb-24 at 08:12WORKING NOW WITH FOLLOWING CONFIG
QUESTION
I'm trying to deploy Prometheus using Prometheus operator. I have used the documentation and helm charts from https://github.com/prometheus-operator/prometheus-operator. Since I need the charts for future reference, rather then directly installing the charts from repository I made a Chart.yaml file and added the repository as dependency.
...ANSWER
Answered 2022-Feb-23 at 07:15an operator pod acts as a controller that listens to events regarding specific custom resources. if you only deploy the operator, you have to seperately deploy the custom resource you wish to be created.
with the prometeus-operator, that would be a custom resource of kind "prometheus". if the helm chart you choose is capable to also deploy this (or not) should be indicated in the charts values.yaml and documented on their github page.
you can also use the examples from the prometheus-operator repo to create prometheus instances. check out these files to do so: https://github.com/prometheus-operator/prometheus-operator/tree/main/example/rbac/prometheus
QUESTION
For the kube-prometheus-stack we added more and more dashboards config to /grafana/dashboards
folder to have more and more provisioned dashboards.
And then in one day we've done this:
...ANSWER
Answered 2022-Feb-19 at 18:36Secret ... is invalid: data: Too long: must have at most 1048576 bytes
This is a well known limitation of Kubernetes secrets (version 1.23 at the moment). The official k8s documentation says:
Individual secrets are limited to 1MiB in size. This is to discourage creation of very large secrets which would exhaust the API server and kubelet memory. However, creation of many smaller secrets could also exhaust memory. More comprehensive limits on memory usage due to secrets is a planned feature.
So, first of all, check if some unnecessary files/dirs
are stored in your chart directories and remove
them. I am sure you have already removed all unnecessary files.
Using such a storage backend is particularly useful if your release information weighs more than 1MB (in which case, it can't be stored in Secrets because of internal limits in Kubernetes).
To enable the SQL backend, you'll need to deploy a SQL database and set the environmental variable HELM_DRIVER
to sql
. The DB details are set with the environmental variable HELM_DRIVER_SQL_CONNECTION_STRING
.
You can set it in a shell as follows:
QUESTION
I had installed kube-prometheus-stack
from the helm chart repo prometheus-community
ANSWER
Answered 2022-Feb-03 at 08:21Enable persistence if you want to make it stateful. However I did not see an option to make Grafana a statefulset in the chart you mentioned.
Usually you will see persistence enable option if the corresponding Helm chart support it. For example: you can enable persistence in this grafana helm chart. You may generate template out of it and make use of it in your repo.
QUESTION
I have a kube-prometheus deployed to multiple environments using kustomize.
kube-prometheus is a base and each environment is an overlay. Let's say I want to deploy dashboards to overlays, which means I need to deploy the same ConfigMaps and the same patch to each overlay.
Ideally, I want to avoid changing the base as it is declared outside of my repo and to keep things DRY and not to copy the same configs all over the place.
Is there a way to achieve this?
Folder structure:
...ANSWER
Answered 2021-Dec-04 at 22:58In your main kustomization, or some top-level overlay, you should be able to call for a common folder or repository.
Have you tried something like this:
QUESTION
I try to install prometheus-community/kube-prometheus-stack
using below command
helm install stable prometheus-community/kube-prometheus-stack -n monitoring --debug
But it failed with issue:
helm.go:88: [debug] failed pre-install: timed out waiting for the condition
I tried to figured out by googled it but cannot find the solution.
below full debug statement:
...ANSWER
Answered 2022-Jan-19 at 14:08I fix it. Need to use not latest version. Just include tag in you image which is not the latest. I use version 18
QUESTION
Hell All- i have Prometheus installed using Helm
...ANSWER
Answered 2022-Jan-19 at 10:46You can check for another instance of Prometheus running on the same cluster:
Community Discussions, Code Snippets contain sources that include Stack Exchange Network
Vulnerabilities
No vulnerabilities reported
Install kube-prometheus
Support
Reuse Trending Solutions
Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items
Find more librariesStay Updated
Subscribe to our newsletter for trending solutions and developer bootcamps
Share this Page