-
Notifications
You must be signed in to change notification settings - Fork 3.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[0.9.0-rc28] Issue with number values being inserted as strings. Select 'mean(value)' crashes the node. #2346
Comments
Update Sometimes it returns this error (instead of Bad Gateway): {
"results": [
{
"error": "Post http://dbserver02:8086/data/run_mapper: EOF"
}
]
} If you are wondering, what are the values for 'network_in' measurement, same query with removing 'mean' keyword is giving: {
"results": [
{
"series": [
{
"name": "network_in",
"columns": [
"time",
"value"
],
"values": [
[
"2015-04-20T14:27:40Z",
"26517893"
],
[
"2015-04-20T14:27:40Z",
"10174"
],
[
"2015-04-20T14:27:40Z",
"76"
],
[
"2015-04-20T14:27:40Z",
"10312"
],
[
"2015-04-20T14:32:27Z",
"265208164"
],
[
"2015-04-20T14:32:27Z",
"10378"
],
[
"2015-04-20T14:32:27Z",
"228"
],
[
"2015-04-20T14:32:27Z",
"10228"
]
]
}
]
}
]
} Does it have something to do with the number type? (the assumption sounds wrong for that kind of "storage", but who knows) |
Update 2: I tried to store some DISK values, and selecting that is behaving same. Taking the fact of big numbers into account, I think it is related. |
Update 3: 0.9.0-RC26, issue is still there |
i'm guessing that calculating the mean is causing an overflow when calculating the sum of large numbers. calculating the sum can be avoided as long as we have the current average and a count for each of the 2 means we are trying to merge.
since it's probably simple enough to update the mean calculations to use this method and avoid overflow.
and then |
@svscorp i can't reproduce your issue. do you have some curl commands to reproduce this from a clean db? meanwhile, i may still go ahead and change the way mean is calculated but it doesn't seem to be the cause of your problem from what i can tell. |
I'll give it when will reach my laptop. But it happenes when I add a graph in graphana. You can change it editing the query, but that's the default query if you use the US to add filters for a graph. |
I was trying to reproduce it from the scratch and couldn't either. It is only reproducable when the data is being inserted with a scheduled script. I'm now busy logging and making a valid case. Will get back once done. Thanks for reaction, though! |
@neonstalwart I got it again. To not overload this thread I've put it here: CURL sequence with queries |
ok, i should have seen this sooner but it's a very subtle issue. your values are strings - e.g. once i had your sample data, i could see the error for myself
this made me realize that your values were strings. the fix is easy - make them numbers. https://gist.github.com/neonstalwart/00d106d8de7c9a960696/5e0a181eb83a649bbb34965dbf6e49347e2a7e29 is a reduced example that demonstrates the problem and https://gist.github.com/neonstalwart/00d106d8de7c9a960696 is the same example with the values changed to numbers and it works. |
This is a known issue, which we will address before the 0.9.0 release. |
It works when sending values as floats |
I see Influxdb 0.9 crash on a simple select query as well. curl -G 'http://localhost:8086/query' --data-urlencode "db=graphite" --data-urlencode "q=select * from "xxx-com.cloud.gauges.raw_storage_usage"" | python -m json.tool The data in influxdb looks like this: Here are the influxdb logs. [http] 2015/07/29 20:59:10 127.0.0.1 - - [29/Jul/2015:20:59:10 +0000] GET /query?q=select+value+from+%22xxx-com.cloud.gauges.num_sessions%22+where+time+%3E+1435611550s+and+time+%3C+1438203551s&p=root&u=root&db=graphite HTTP/1.1 200 3198 - python-requests/2.2.1 CPython/2.7.6 Linux/3.16.0-38-generic a930fa47-3634-11e5-801c-000000000000 7.753755ms goroutine 3633 [running]: |
Hi,
I am facing a weird issue on 3-server-cluster setup (replication factor = 2), which might be related to #2272
I am having two measurements: memory and network_in (I have more, but let's just pock those two)
Query to the cluster:
answer:
Then, I am making same query
And result is (with crashing a node):
Is there something I am missing?
The text was updated successfully, but these errors were encountered: