You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
There are cases where a particular MV/sink is consuming a lot of resources due to some unknown problem at the moment. No matter it can be perfectly solved or not eventually, users may often want to remove this particular one and bring the cluster back to normal for the moment. Showing how much memory is being used by each MV can help easily locate the problematic MV/sinks. (Don't know if there is an existing to locate the problem? I may miss a few things)
There are cases where too many MVs/sinks exist in the cluster and the workload runs out of memory as a whole, i.e. everything works normally but the cluster is simply under too much stress. However, it is unclear to users how many is too many. Showing how much memory is being used by each MV looks very intuitive. Users will have reasonable expectations of RW and are more likely to be convinced.
The text was updated successfully, but these errors were encountered:
"risingwave_dev_dashboard -> Streaming Actors -> Executor Cache Memory Usage of Materialized Views" can monitor the memory usage of the executor's cache.
But the memtable's memory usage is hard to monitor More info here #11442 c.c. @fuyufjh
Just checked the code of Executor Cache Memory Usage of Materialized Views or Executor Cache Memory Usage, and want to confirm that:
they are both estimated, and could be inaccurate, e.g. versus Jemalloc (which I suppose is accurate? but also impossible to get the current metric via Jemalloc?) (I think inaccuracy could be a minor problem that can be ignored right now, after all we just want to have a big picture)
still need to do the sum over multiple tables that belong to the same query manually to calculate the total memory usage of that query (but this is awkward to do within Grafana as the query-tables relationship is not there)
they are both estimated, and could be inaccurate, e.g. versus Jemalloc (which I suppose is accurate? but also impossible to get the current metric via Jemalloc?) (I think inaccuracy could be a minor problem that can be ignored right now, after all we just want to have a big picture)
jemalloc does not support multi-tenancy(I am not sure if this word is proper...) memory statistics. It can only give all memory usage in the process
still need to do the sum over multiple tables that belong to the same query manually to calculate the total memory usage of that query (but this is awkward to do within Grafana as the query-tables relationship is not there)
IIRC "Executor Cache Memory Usage of Materialized Views" has sum all the tables with the promQL
The motivation is two-fold:
There are cases where a particular MV/sink is consuming a lot of resources due to some unknown problem at the moment. No matter it can be perfectly solved or not eventually, users may often want to remove this particular one and bring the cluster back to normal for the moment. Showing how much memory is being used by each MV can help easily locate the problematic MV/sinks. (Don't know if there is an existing to locate the problem? I may miss a few things)
There are cases where too many MVs/sinks exist in the cluster and the workload runs out of memory as a whole, i.e. everything works normally but the cluster is simply under too much stress. However, it is unclear to users how many is too many. Showing how much memory is being used by each MV looks very intuitive. Users will have reasonable expectations of RW and are more likely to be convinced.
The text was updated successfully, but these errors were encountered: