- 04 5月, 2017 6 次提交
-
-
由 Marbin Tan 提交于
* Address pylint warnings and errors - Fix whitespace and indentation - Remove unused variables - Fix syntax errors Signed-off-by: NMarbin Tan <mtan@pivotal.io>
-
由 Haisheng Yuan 提交于
By setting material->cdb_strict to true if the outer child of material is a partition selector. Before this patch, the following query return incomplete results quite often: create table t(id int, a int); create table pt(id int, b int) DISTRIBUTED BY (id) PARTITION BY RANGE(b) (START (0) END (5) EVERY (1)); insert into t select i, i from generate_series(0,4) i; insert into pt select i, i from generate_series(0,4) i; analyze t; analyze pt; set enable_hashjoin=off; set enable_nestloop=on; select * from t, pt where a = b; In 3 segments cluster, it may return different result shown below: hyuan=# select * from t, pt where a = b; id | a | id | b ----+---+----+--- 0 | 0 | 0 | 0 1 | 1 | 1 | 1 2 | 2 | 2 | 2 (3 rows) hyuan=# select * from t, pt where a = b; id | a | id | b ----+---+----+--- 3 | 3 | 3 | 3 4 | 4 | 4 | 4 (2 rows) hyuan=# select * from t, pt where a = b; id | a | id | b ----+---+----+--- 3 | 3 | 3 | 3 4 | 4 | 4 | 4 0 | 0 | 0 | 0 1 | 1 | 1 | 1 2 | 2 | 2 | 2 (5 rows) But only the last one is correct result. The plan for above query is: ------------------------------------------------------------------- Gather Motion 3:1 (slice2; segments: 3) -> Nested Loop (cost=2.27..9.00 rows=2 width=16) Join Filter: t.a = public.pt.b -> Append (cost=0.00..5.05 rows=2 width=8) -> Result (cost=0.00..1.01 rows=1 width=8) One-Time Filter: PartSelected -> Seq Scan on pt_1_prt_1 pt -> Result (cost=0.00..1.01 rows=1 width=8) One-Time Filter: PartSelected -> Seq Scan on pt_1_prt_2 pt -> Result (cost=0.00..1.01 rows=1 width=8) One-Time Filter: PartSelected -> Seq Scan on pt_1_prt_3 pt -> Result (cost=0.00..1.01 rows=1 width=8) One-Time Filter: PartSelected -> Seq Scan on pt_1_prt_4 pt -> Result (cost=0.00..1.01 rows=1 width=8) One-Time Filter: PartSelected -> Seq Scan on pt_1_prt_5 pt -> Materialize (cost=2.27..2.42 rows=5 width=8) -> Partition Selector for pt (dynamic scan id: 1) Filter: t.a -> Broadcast Motion 3:3 (slice1; segments: 3) -> Seq Scan on t Settings: enable_hashjoin=off; enable_nestloop=on Optimizer status: legacy query optimizer The data distribution for table t and pt in 3 segments environment is: hyuan=# select gp_segment_id, * from t; gp_segment_id | id | a ---------------+----+--- 1 | 3 | 3 1 | 4 | 4 0 | 0 | 0 0 | 1 | 1 0 | 2 | 2 (5 rows) hyuan=# select gp_segment_id, * from pt; gp_segment_id | id | b ---------------+----+--- 0 | 0 | 0 0 | 1 | 1 0 | 2 | 2 1 | 3 | 3 1 | 4 | 4 (5 rows) Tuples {0,1,2} of t and pt are in segment 0, tuples {3,4} of t and pt are in segment 1. Segment 2 has no data for t and pt. In this query, planner decides to prefetch inner child to avoid deadlock hazard and the cdb_strict of Material is set to false. Let's see how the query goes in segment 0. 1. The inner child of nestloop join, material fetch one tuple from partition selector and then output it. Let's assume the output order of partition selector/broadcast motion is {0,1,2,3,4}. So the 1st tuple output by partition selector and material is 0. 2. The partition selector decides that the selected partition for table pt is pt_1_prt_1, because t.a = pt.b = 0 in this partition. The outer child of nestloop join, Append, fetches 1 tuple from that partition, with pt.b=0. 3. Nestloop join continues to execute Material of inner child to fetch other tuples, 1,2,3,4, but all these tuples from t don't match the join condition, because pt=0. No more tuples output by nestloop join for this round of loop. But all the partition of pt are matched and selected. 4. Nestloop join fetch another tuple from pt_1_prt_2, which is 1, which can match with a tuple from inner child, output 1. And then fetch tuple from pt_1_prt_3, which is 2, matched, output 2. But pt_1_prt_4 and pt_1_prt_5 have no data in this segment, so the output ends with {0,1,2} in segment 0. But in segment 1, let's still assume the tuple output order of partition selector/broadcast motion is {0,1,2,3,4}. Since the first output tuple from inner child is 0, only pt_1_prt_1 is selected. But when nestloop join tries to fetch tuple from outer child, which in fact fetch from pt_1_prt_1 in this case, it returns no tuple, because pt_1_prt_1 is empty in this segment. So nestloop join decides that since it can't fetch any tuple from outer child, it must be empty, no need to execute the join, return NULL and finish it directly. Segment 2 has no data for t and pt, no tuple is output. So the final result gathered on master segment is {0,1,2} in this case. But if the broadcast motion output tuple order is {3,4,0,1,2}, the final result may be {3,4}. If the braodcast motion output tuple order on segment 0 is {0,1,2,3,4}, and on segment 1 is {3,4,0,1,2}, then the final result on master is {0,1,2,3,4}, which is correct. The bug is fixed by setting cdb_strict of material to true when planner generates partition selector for the inner child of nestloop join, the material will fetch all the tuples from child and materialize them before emitting any tuple. Thus we can make sure the partitions of pt are selected correctly. RCA by Lirong Jian <jianlirong@gmail.com>
-
由 C.J. Jameson 提交于
[ci skip] Signed-off-by: NMarbin Tan <mtan@pivotal.io>
-
由 Marbin Tan 提交于
Signed-off-by: NLarry Hamel <lhamel@pivotal.io>
-
由 Larry Hamel 提交于
Signed-off-by: NMarbin Tan <mtan@pivotal.io>
-
由 Jingyi Mei 提交于
Signed-off-by: NTushar Dadlani <tdadlani@pivotal.io>
-
- 03 5月, 2017 13 次提交
-
-
由 xiong-gang 提交于
Increase the 'concurrency' limit can take effect immediately, and the queueing transactions can be woken up. Decrease the 'concurrency' is different, if the new limit is smaller than the number of current running transactions, ALTER statement won't cancel the running transactions to the limit. Therefore, we use column 'proposed' in pg_resgroupcapability to represent the effective limit, and use column 'value' to record the historical limit. For example, we have a resource group with concurrency=3, and there are 3 running transactions and 3 queueing transactions. If we alter the concurrency to 2, 'proposed' will be updated to 2 and 'value' will stay as 3. When one running transaction is finished, it won't wake up the transactions in the queue as the current concurrency is 2. If we execute the statement again to alter the concurrency to 2, it will update the 'value' column to 2, and the 'value' is consistent with 'proposed' again. Signed-off-by: NRichard Guo <riguo@pivotal.io>
-
由 Adam Lee 提交于
Support COPY statement that exports the table directly from segment to local file parallelly. This commit adds a keyword "on segment" to save the copied file on "segment" instead of on "master". Two place holders are used, which are "<SEG_DATA_DIR>" and "<SEGID>" and will be replced to segment datadir and segment id. E.g. ``` COPY tbl TO '/tmp/<SEG_DATA_DIR>filename<SEGID>.txt' ON SEGMENT; ``` Signed-off-by: NYuan Zhao <yuzhao@pivotal.io> Signed-off-by: NHaozhou Wang <hawang@pivotal.io> Signed-off-by: NAdam Lee <ali@pivotal.io>
-
由 mkiyama 提交于
-
由 David Yozie 提交于
* removing info for older, incompatible netbackup versions * removing system requirements from backup topic; adding conditionalized reference to release notes for supported netbackup versions
-
由 Jim Doty 提交于
Signed-off-by: NTom Meyer <tmeyer@pivotal.io>
-
由 Daniel Gustafsson 提交于
We don't have pl/java in gpAux/extensions anymore and there are no proprietary modules either. [ci skip]
-
-
由 Larry Hamel 提交于
* add behave test for diskspace_history * remove commented-out test (moved to new tracker story to redo) * refactor test for qamode Signed-off-by: NC.J. Jameson <cjameson@pivotal.io>
-
由 Larry Hamel 提交于
Signed-off-by: NMarbin Tan <mtan@pivotal.io> Signed-off-by: NC.J. Jameson <cjameson@pivotal.io>
-
由 Shreedhar Hardikar 提交于
-
由 Jingyi Mei 提交于
We did the following changes: 1. changed make files to reference sles instead of suse, and use sles11-x86_64 instead of sles11_x86_64 in the zip artifact 2. changed make files to use rhel6/7-x86_64 instead of RHEL6/7-x86_64 in the zip artifact 3. changed set_bld_arch.sh to set BLD_ARCH=sles instead of suse 4. in dependencies ivy file, add sles11_x86_64 as configurations to multiple repos which only had suseXX-x86_64 before 5. set multiple config path/flags for sles11_x86_64 in Makefile which didn't exist before Signed-off-by: NTom Meyer <tmeyer@pivotal.io> Signed-off-by: NKris Macoskey <kmacoskey@pivotal.io> Signed-off-by: NJingyi Mei <jmei@pivotal.io>
-
由 David Yozie 提交于
-
由 Jane Beckman 提交于
* Add new OPERATOR_FAMILY pages * Updates for Postgres 8.3 * Corrections to nav map * Correct xref pointer * xml format fix * Standardize xml for new xrefs * Remove duplicate link * Change ref to Greenplum Database * Updates for new PostreSQL commands ALTER/CREATE/DROP OPERATOR FAMILY, ALTER VIEW, and DISCARD. * Updates from Heikki * Remove 8.3 comparison * Update GET DIAGNOSTICS
-
- 02 5月, 2017 5 次提交
-
-
由 Daniel Gustafsson 提交于
While looking at other things, spotted that there were quite a few functions in catalog.py that were unused. Remove these, and also remove imports of gppylib.db.catalog when unused.
-
由 C.J. Jameson 提交于
- We think that external tables previously broke with newlines. This is no longer the case, so removing space replacement logic. Signed-off-by: NMarbin Tan <mtan@pivotal.io> Signed-off-by: NLarry Hamel <lhamel@pivotal.io>
-
For appendonly_read_check, setting `optimizer_disable_missing_stats_collection=on` to hide the additional NOTICES/HINTS produced by ORCA. For uao/uaocs crash on update tests, we are setting `optimizer=off` as ORCA does not go through the code path intended for these tests.
-
由 Jamie McAtamney 提交于
Authors: Karen Huddleston, Jamie McAtamney
-
由 David Yozie 提交于
-
- 01 5月, 2017 3 次提交
-
-
The sub-transaction tests are injecting faults at a point that only happens in a reader gang on segments. The original query, when planned by ORCA, would lead to a plan where the reader slice executes on master. This commit changes the test (while mostly staying true to its spirits): instead of using `INSERT INTO ... VALUES ...` we put the value in a temporary table, and do an `INSERT INTO ... SELECT ... FROM`. The new test should now pass under both ORCA and planner.
-
由 mkiyama 提交于
-
由 Venkatesh Raghavan 提交于
-
- 29 4月, 2017 6 次提交
-
-
由 Eric Wu 提交于
*polymorphims -- ORCA will not add order by clause in derived table, thus we add 'limit 3' to force it add SORT node to make sure test result is consistent. *alter_table_aocs -- fix the input data to make order by gets stable result
-
This particular delete line fails to do anything because the subquery returns nothing. Remove the line since it does not do anything for the walrep tests.
-
由 C.J. Jameson 提交于
- gpperfmon tracks a variety of queries and statistics and system properties and metrics in a dedicated database - Until recently, it hadn't been tested or built very reliably here. It is now part of the gpdb_master pipeline. - We are doing a lot of work to revitalize and provide tests for this feature; we plan to make many upcoming pull requests for this feature in different ways - We will look into testing this with installcheck, but for now, it is tested with Behave - Building it as part of PRs will help us get faster feedback - This should add negligible time to compilation: the tests themselves run in less than 5 minutes typically, and will be running in parallel with ICW Signed-off-by: NMarbin Tan <mtan@pivotal.io>
-
由 David Yozie 提交于
* moving linux kerberos client instructions to dedicated topic, and placing it parallel with existing windows topic * removing map and dita files for connectivity package, which is no longer provided * removing more references to connectivity packages; conditionalizing postgres odbc/jdbc vs datadirect drivers for oss/pivotal audiences * adding .ditaval for handling pivotal conditions; [ci skip] * relocating pivotal ditaval * adding .ditaval for handling pivotal conditions; [ci skip] * relocating pivotal ditaval * removing/fixing old references to 4.3 version * removing pgcrypto.fips mentions * small fix to version format * removing 'orca' from package version strings * changing rhel5 to rhel6 in example version strings
-
由 Marbin Tan 提交于
Signed-off-by: NMelanie Plageman <mplageman@pivotal.io>
-
由 Larry Hamel 提交于
Signed-off-by: NMarbin Tan <mtan@pivotal.io>
-
- 28 4月, 2017 7 次提交
-
-
由 Daniel Gustafsson 提交于
-
由 Daniel Gustafsson 提交于
src/bin was an incorrect location for gp_session_state, and while contrib might not be ideal it's better than src/bin at least. Move into the gp_internal_tools contrib module which contains various Greenplum specific management functions and views. During the next release cycle we should perhaps make this a proper backend function and move it into gp_toolkit but for now lets collect these utils in one place.
-
由 Daniel Gustafsson 提交于
The gp_aoseg_history function was rolled into gp_ao_co_diagnostics some time ago and has since been disconnected. The cache_stats function in gp_workfile_mgr was simply not used and never exposed such that a user could use it.
-
由 Daniel Gustafsson 提交于
The gp_workfile_mgr functions are installed via gp_toolkit, so being in src/bin is the wrong place for such code. Move to the gp_internal_tools contrib, which also is the wrong place, but at least has precedence for containing gp_toolkit functions. These should at some point be made into backend functions and moved from contrib but not at this point in cycle.
-
由 Daniel Gustafsson 提交于
Minor copyediting, spelling fixes and style tweaks spotted while glancing over the code.
-
由 Daniel Gustafsson 提交于
The workfile_manager unittest code was previously included in all builds, this is however functionality which shouldn't be in prod installations. This moves all unit test code to src/test/regress and links it to the regress object such that it can be tested with the normal pg_regress process. Add a random test from the suite into ICW just to exercise it, none of these test were ever executed in automated suites before so not sure what we actually want to run.
-
由 Jesse Zhang 提交于
Commit 35e0dfdb first bumped this up for an isolation test. Higher and higher concurrency has since crept into installcheck, resulting in a flaky group that contains `mapred`. This hopefully addresses that flake.
-