-
Notifications
You must be signed in to change notification settings - Fork 3.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update raw queries with a LIMIT to read in only what they need #2030
Milestone
Comments
pauldix
added a commit
that referenced
this issue
Mar 28, 2015
Refactored query engine to have different processing pipeline for raw queries. This enables queries that have a large offset to not keep everything in memory. It also makes it so that queries against raw data that have a limit will only p rocess up to that limit and then bail out. Raw data queries will only read up to a certain point in the map phase before yielding to the engine for further processing. Fixes #2029 and fixes #2030
pauldix
added a commit
that referenced
this issue
Mar 28, 2015
Refactored query engine to have different processing pipeline for raw queries. This enables queries that have a large offset to not keep everything in memory. It also makes it so that queries against raw data that have a limit will only p rocess up to that limit and then bail out. Raw data queries will only read up to a certain point in the map phase before yielding to the engine for further processing. Fixes #2029 and fixes #2030
pauldix
added a commit
that referenced
this issue
Apr 1, 2015
Refactored query engine to have different processing pipeline for raw queries. This enables queries that have a large offset to not keep everything in memory. It also makes it so that queries against raw data that have a limit will only p rocess up to that limit and then bail out. Raw data queries will only read up to a certain point in the map phase before yielding to the engine for further processing. Fixes #2029 and fixes #2030
pauldix
added a commit
that referenced
this issue
Apr 4, 2015
Refactored query engine to have different processing pipeline for raw queries. This enables queries that have a large offset to not keep everything in memory. It also makes it so that queries against raw data that have a limit will only p rocess up to that limit and then bail out. Raw data queries will only read up to a certain point in the map phase before yielding to the engine for further processing. Fixes #2029 and fixes #2030
pauldix
added a commit
that referenced
this issue
Apr 6, 2015
Refactored query engine to have different processing pipeline for raw queries. This enables queries that have a large offset to not keep everything in memory. It also makes it so that queries against raw data that have a limit will only p rocess up to that limit and then bail out. Raw data queries will only read up to a certain point in the map phase before yielding to the engine for further processing. Fixes #2029 and fixes #2030
pauldix
added a commit
that referenced
this issue
Apr 7, 2015
Refactored query engine to have different processing pipeline for raw queries. This enables queries that have a large offset to not keep everything in memory. It also makes it so that queries against raw data that have a limit will only p rocess up to that limit and then bail out. Raw data queries will only read up to a certain point in the map phase before yielding to the engine for further processing. Fixes #2029 and fixes #2030
pauldix
added a commit
that referenced
this issue
Apr 9, 2015
Refactored query engine to have different processing pipeline for raw queries. This enables queries that have a large offset to not keep everything in memory. It also makes it so that queries against raw data that have a limit will only p rocess up to that limit and then bail out. Raw data queries will only read up to a certain point in the map phase before yielding to the engine for further processing. Fixes #2029 and fixes #2030
pauldix
added a commit
that referenced
this issue
Apr 9, 2015
Refactored query engine to have different processing pipeline for raw queries. This enables queries that have a large offset to not keep everything in memory. It also makes it so that queries against raw data that have a limit will only p rocess up to that limit and then bail out. Raw data queries will only read up to a certain point in the map phase before yielding to the engine for further processing. Fixes #2029 and fixes #2030
pauldix
added a commit
that referenced
this issue
Apr 10, 2015
Refactored query engine to have different processing pipeline for raw queries. This enables queries that have a large offset to not keep everything in memory. It also makes it so that queries against raw data that have a limit will only p rocess up to that limit and then bail out. Raw data queries will only read up to a certain point in the map phase before yielding to the engine for further processing. Fixes #2029 and fixes #2030
pauldix
added a commit
that referenced
this issue
Apr 10, 2015
Refactored query engine to have different processing pipeline for raw queries. This enables queries that have a large offset to not keep everything in memory. It also makes it so that queries against raw data that have a limit will only p rocess up to that limit and then bail out. Raw data queries will only read up to a certain point in the map phase before yielding to the engine for further processing. Fixes #2029 and fixes #2030
mark-rushakoff
pushed a commit
that referenced
this issue
Jan 11, 2019
…state fix(ui/dataLoaders): Separate the error state from the data not found
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
The underlying iterators in
tx.go
should read and yield data points only up to the limit if the query is a raw data query (one without an aggregate).The reducer will then have to take the output of the mappers, combine together, sort and then put the limit. So each mapper will return up to what the limit is (assuming there is multiple mappers in a given time interval).
If there is only one mapper per time interval, they should be run sequentially until the limit is hit.
The text was updated successfully, but these errors were encountered: