-
Notifications
You must be signed in to change notification settings - Fork 84
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unexpected behaviour when resampling with granularity = end - start #997
Comments
just curious but what is i'd have to look into this it's possible it's because we time is aggregated relative to epoch so maybe 22hrs splits different? i'm completely guessing. |
It returns a search expression to find instances that exist within the given time window:
I think something like that could be a likely explanation. I investigated a bit more and made the following observation: duration 1h returns single value for intervals [0:00,1:00],[1:00,2:00],[2:00,3:00], etc. |
hmmm maybe resample with not sure if this is reason but i don't think get_range_query should be |
also, could you add what the archive policy is for the metrics are? are there multiple granularities? |
Thank you for your help and suggestions. I replaced 'max' by 'count' and here is the output corresponding to the example from my original post:
My interpretation is that we have |
yeah, i don' t know why i was asking about search... i just realised that isn't related to series data at all. apologies. so i think i remember how this all works now :) and it does relate to epoch. when you resample, it's resampling based on that so for example, even though you give it a range of 22 hours, those 22 hours of data does not necessarily all fall in to the same 22 hour groups from epoch. i think pandas behaves differently and groups based on first index. i'd need to think about whether there's a way to actually a way to get the max of each group over a specific 22 hour span. if exactly 22hrs is not requird, then you could change your use or you could add ability to group off first index rather than epoch |
Closing because of no new information, please update and we can open again. |
I am currently working with gnocchi version 4.1.4, since as of now we are unable to upgrade to the latest version due to some OpenStack cross-dependencies.
Our use case is to identify the top most resource consuming instances. Therefore, I started to work out a solution based on the example posted here:
There @sileht wrote:
which is exactly what I want to achieve as well. However, if I run the following minimal working example:
the resulting entries contain two elements in the
max
list, e.g.:However, when I set
start = datetime(2018, 10, 15, 13, 0)
andend = datetime(2018, 10, 15, 14, 0)
I get only one entry as I anticipated. Therefore, I have the feeling that it could be bug. But it could also be a misunderstanding on my side.Thanks
The text was updated successfully, but these errors were encountered: