I am running a single binary Loki 3.2.1 on a filesystem TSDB store, where my logs gradually fade out (in a few hours, mostly determined by max_chunk_age) when they are flushing off the ingester. I can no longer access them through Grafana dashboard, despite no error logged.
I came across this issue which seems to be the exact thing I am experiencing. Is there any workaround currently available, apart from downgrading to 2.8.4?
I’d say double check and make sure chunks are actually written into your permanent storage (and make sure /var/lib/loki is from outside of container).
LXC Container is similar to VM. Everything is persistent. Loki is not running in a one-off containers like Docker.
If your logs are disappearing after flushing then it’s not written to your chunk storage, or it’s written to the wrong place.
I could find the index file and the chunk file in the expected place, and the directory size is actually increasing. Unless there is a separated path settings in the querier I’d say it behaves in the expected way.
After tackling with the config file here and there, followed by a hard crash of the entire VM, the problem seems to be fixed.
It looks like the culprit is the instance_addr / instance_interface_names in the common section, which I shouldn’t touch upon when using inmemory ring store.
Also
Does your Loki container produce any error from querier or query-frontend when querying older logs?
There is no error indicated. There is lots of debug log about mock.goGet - deadline exceeded.