-
Notifications
You must be signed in to change notification settings - Fork 2.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Loki-distributed] query error open /var/loki/chunks/ #1111
Comments
This started working after using s3 storage with the following loki-distributed config:
|
I'm experiencing the same thing with a very similar config to yours but using azure blob storage. |
Thanks @danielserrao |
I have the same problem after restarting some components. Anyone have a solution on how to fix it?
But the file exists and all perminision is OK
|
When I use loki-simple-scalabel, and I use nfs of storageClass, open /var/loki/chunks/fake/755005aa5e414340/MTgxMTNjOGM5MGI6MTgxMTQzNmE2NTI6M2RkYjQzYmQ=: no such file or directory when I enter the write pod ,the file is exists! This error occurs sometimes and sometimes not |
This is mentioned in the chart README I think:
Using filesystem storage in the multi pod setup would require multiple pods to access the same volume, so data is only queryable as long as it's cached in memory. |
I could get the things working by configuring the volumes:
and I created this folder with permissions to the pods to write on them. Of course, this settings are for local directories, not for volumes on GCS or S3, for example. |
@aberenshtein hi, have you solved this issue? i meet the same issue. I don't use object storage and just use filesystem(lvm-localpv) |
yes, but I see that the references I put for the value files are outdated. |
I'm getting this error when there's high traffic in the cluster. I managed to duplicate by running the benchmark tool - wrk. It seems that when Any solution for this? UPDATE: I'm running the following
|
For me, the problem was solved by removing the default configuration here is the snippet that removes the extra # values.yaml
loki:
annotations: {}
...
storageConfig:
boltdb_shipper:
shared_store: s3
aws:
s3: s3://${cluster_region}
bucketnames: ${bucket_name}
filesystem: null Notice the latest # generated configMap
apiVersion: v1
data:
config.yaml: |
auth_enabled: false
...
storage_config:
aws:
bucketnames: bucket-for-logs
s3: s3://${region}
boltdb_shipper:
active_index_directory: /var/loki/index
cache_location: /var/loki/cache
cache_ttl: 168h
shared_store: s3
- filesystem:
- directory: /var/loki/chunks
|
I have distributed micro services working in one cluster, but in production facing issues after couple of weeks. I added pvc to grafana and restarted the same and now I am not able to get labels in grafana UI with "failed to call resource " |
i still have problem when use distributed to query log |
As pointed in a previous comment, the querier and the ingester needs to have access to the same dir I got working and this is by creating a pvc out of the helm chart because the helm chart is hard coding the access mode of the data pvc, and therefore I had to create a RWX pvc
|
I have Grafana with the Loki datasource pointing to the loki querier-frontend but I get the following error when making queries:
Sometimes it is working and then it gets the same error for some reason that is not clear to me.
On the logs of the querier-frontend pod I can see:
caller=logging.go:72 traceID=5c8361c04594c7a2 orgID=fake msg="GET /loki/api/v1/query_range?direction=BACKWARD&limit=1000&query=%7Bjob%3D%22fbit_k8s%22%7D&start=1647619419284000000&end=1647630219285000000&step=5 (500) 53.767877ms Response: \"open /var/loki/chunks/ZmFrZS9kOGU4OGYwOTg3ZTM0NWUyOjE3ZjllMTk0NmE4OjE3ZjllMTk1NmVkOmMwMWFiYmNm: no such file or directory\\n\" ws: false; Accept: application/json, text/plain, */*; Accept-Encoding: gzip, deflate, br; Accept-Language: en-GB,en;q=0.9,en-US;q=0.8; Sec-Ch-Ua: \" Not A;Brand\";v=\"99\", \"Chromium\";v=\"99\", \"Microsoft Edge\";v=\"99\"; Sec-Ch-Ua-Mobile: ?0; Sec-Ch-Ua-Platform: \"Windows\"; Sec-Fetch-Dest: empty; Sec-Fetch-Mode: cors; Sec-Fetch-Site: same-origin; User-Agent: Grafana/8.3.5; X-Forwarded-For: 127.0.0.1, 127.0.0.1; X-Grafana-Org-Id: 1; "
When doing "helm template", the K8s manifest (which is applied) is the following:
test.txt
I already tried multiple types of configurations, but I always get this annoying error.
Some help would be very appreciated.
The text was updated successfully, but these errors were encountered: