- 02 6月, 2018 2 次提交
-
-
由 Lisa Owen 提交于
* docs - add resgroup global shmem recommendation * put conditions in a list * explicitly call out vmtracker roles for global shmem * Changing formatting from varname to i
-
由 Mel Kiyama 提交于
* docs - add gprestore options --data-only, --metadata-only. also, fix title link to gpbackup plugins. * docs - gprestore --data-only, --metadata-only. review comment updates
-
- 01 6月, 2018 5 次提交
-
-
由 Taylor Vesely 提交于
Unlike upstream, GPDB needs to keep collations in-sync between multiple databases. Add tests for GPDB specific collation behavior. These tests need to import a system locale, so add a @syslocale@ variable to gpstringstubs.pl in order to test the creation/deletion of collations from system locales. Co-authored-by: NJim Doty <jdoty@pivotal.io>
-
由 Taylor Vesely 提交于
Make CREATE COLLATION and pg_import_system_collations() parallel aware by dispatching collation creation to the QEs. In order for collations to work correctly, we need to be sure that every collation that is created on the QD is also installed on the QEs, and that the OID matches in every database. We will take advantage of two phase commit to prevent collations from being created if there is a problem adding it on any QE. In upstream, collations are created during initdb, but this won't work for GPDB, because while initdb is running there is no way to be sure that every segment has the same locales installed. We disable collation creation during initdb, and make it the responsibility of the system administrator to initialize any needed collations by either running a CREATE COLLATION command, or running the pg_import_system_collations() UDF. Co-authored-by: NJim Doty <jdoty@pivotal.io>
-
由 Tom Lane 提交于
Pull in more recent version of pg_import_system_collations() from upstream. We have not pulled in the ICU collations, so wholesale remove the sections of code that deal with them. This commit is primarily a cherry-pick of 0b13b2a7, but also pulls in prerequisite changes for CollationCreate(). Rethink behavior of pg_import_system_collations(). Marco Atzeri reported that initdb would fail if "locale -a" reported the same locale name more than once. All previous versions of Postgres implicitly de-duplicated the results of "locale -a", but the rewrite to move the collation import logic into C had lost that property. It had also lost the property that locale names matching built-in collation names were silently ignored. The simplest way to fix this is to make initdb run the function in if-not-exists mode, which means that there's no real use-case for non if-not-exists mode; we might as well just drop the boolean argument and simplify the function's definition to be "add any collations not already known". This change also gets rid of some odd corner cases caused by the fact that aliases were added in if-not-exists mode even if the function argument said otherwise. While at it, adjust the behavior so that pg_import_system_collations() doesn't spew "collation foo already exists, skipping" messages during a re-run; that's completely unhelpful, especially since there are often hundreds of them. And make it return a count of the number of collations it did add, which seems like it might be helpful. Also, re-integrate the previous coding's property that it would make a deterministic selection of which alias to use if there were conflicting possibilities. This would only come into play if "locale -a" reports multiple equivalent locale names, say "de_DE.utf8" and "de_DE.UTF-8", but that hardly seems out of the question. In passing, fix incorrect behavior in pg_import_system_collations()'s ICU code path: it neglected CommandCounterIncrement, which would result in failures if ICU returns duplicate names, and it would try to create comments even if a new collation hadn't been created. Also, reorder operations in initdb so that the 'ucs_basic' collation is created before calling pg_import_system_collations() not after. This prevents a failure if "locale -a" were to report a locale named that. There's no reason to think that that ever happens in the wild, but the old coding would have survived it, so let's be equally robust. Discussion: https://postgr.es/m/20c74bc3-d6ca-243d-1bbc-12f17fa4fe9a@gmail.com (cherry picked from commit 0b13b2a7)
-
由 Peter Eisentraut 提交于
Move this logic out of initdb into a user-callable function. This simplifies the code and makes it possible to update the standard collations later on if additional operating system collations appear. Reviewed-by: NAndres Freund <andres@anarazel.de> Reviewed-by: NEuler Taveira <euler@timbira.com.br> (cherry picked from commit aa17c06f)
-
由 Omer Arap 提交于
-
- 31 5月, 2018 1 次提交
-
-
由 Paul Guo 提交于
This seems to be related to AIX system issue or old compiler. Not long ago there is a similar complaint on the pg community also. http://www.postgresql-archive.org/pgsql-Improve-performance-of-SendRowDescriptionMessage-td5987721.html Do not want to waste too much time on this. Instead, I just work around this issue following what we did on auth.c, i.e. +#if defined(_AIX) +int getpeereid(int, uid_t *__restrict__, gid_t *__restrict__); +#endif
-
- 30 5月, 2018 5 次提交
-
-
由 Tang Pengzhou 提交于
* Refine the fault injector framework * Add counting feature so a fault can be triggered N times. * Add a simpler version named gp_inject_fault_infinite. * Refine and make code cleaner include renaming sleepTimes to extraArg so it can be used by other fault types. Now 3 functions provided: 1. gp_inject_fault(faultname, type, ddl, database, tablename, start_occurrence, end_occurrence, extra_arg, db_id) startOccurrence: nth occurrence that a fault starts triggering endOccurrence: nth occurrence that a fault stops triggering, -1 means the fault is always triggered until it is reset. 2. gp_inject_fault(faultname, type, db_id) simpler version for fault triggered only once. 3. gp_inject_fault_infinite(faultname, type, db_id) simpler version for fault always triggered until it's reset. * fix bgwriter_checkpoint case * use gp_inject_fault_infinite here instead of gp_inject_fault so cache of pg_proc that contains gp_inject_fault_infinite is loaded before checkpoint and the following gp_inject_fault_infinite don't dirty the buffer again. * Add a matchsubs to ignore 5 or 6 times hits of fsync_counter. * Fix flaky twophase_tolerance_with_mirror_promotion test * use different session for Scenario 2 and Scenario 3 because the gang of session 2 is no longer valid. * wait for wanted fault to be triggered so no unexpected error occurs. * Add more segment status info to identify error quickly Some cases are right behind FTS test cases. If the segments are not in the desired status, those test cases will fail unexpectedly, this commit adds more debug info at the beginning of test cases to help to identify issues quickly. * Enhance cases to skip fts probe for sure * Do FTS probe request twice to guarantee fts error is triggered
-
由 Paul Guo 提交于
This kind of error could lead to serious problems sometimes.
-
由 Mel Kiyama 提交于
* docs - gpbackup/gprestore plugin for DD Boost - also minor update to S3 plugin- change yaml file parameter backupdir to folder. * docs - review updates for gpbackup plugin for ddboost also an update for S3. change keyword backupdir -> folder in example. * docs - more review updates for gpbackup plugin for ddboost updated files on review comments. also updated HTML on review site. * docs - another set of review updates for gpbackup plugin for ddboost --edits/updates --change toc -move plugin api up a level. --tweaked format of parameter definition. --also update S3 plugin to parallel gpbackup doc. * docs - added pivotal only attribute to topic.
-
由 Lisa Owen 提交于
* docs - misc updates to gpbackup/gprestore plugin api docs * address review comments from david
-
由 Jacob Champion 提交于
The 9.1 merge brought upstream support for `make -k` when running installcheck-world. Use it in the master pipeline.
-
- 29 5月, 2018 3 次提交
-
-
由 Ning Yu 提交于
Some error messages were updated during the 9.1 merge, update the answers for the RETURNING test cases of replicated tables.
-
由 Ning Yu 提交于
* rpt: reorganize data when ALTER from/to replicated. There was a bug that altering from/to a replicated table has no effect, the root cause is that we did not change gp_distribution_policy neither reorganize the data. Now we perform the data reorganization by creating a temp table with the new dist policy and transfering all the data to it. * rpt: support RETURNING for replicated tables. This is to support below syntax (suppose foo is a replicated table): INSERT INTO foo VALUES(1) RETURNING *; UPDATE foo SET c2=c2+1 RETURNING *; DELETE * FROM foo RETURNING *; A new motion type EXPLICIT GATHER MOTION is introduced in EXPLAIN output, data will be received from one explicit sender in this motion type. * rpt: fix motion type under explicit gather motion. Consider below query: INSERT INTO foo SELECT f1+10, f2, f3+99 FROM foo RETURNING *, f1+112 IN (SELECT q1 FROM int8_tbl) AS subplan; We used to generate a plan like this: Explicit Gather Motion 3:1 (slice2; segments: 3) -> Insert -> Seq Scan on foo SubPlan 1 (slice2; segments: 3) -> Gather Motion 3:1 (slice1; segments: 1) -> Seq Scan on int8_tbl A gather motion is used for the subplan, which is wrong and will cause a runtime error. A correct plan is like below: Explicit Gather Motion 3:1 (slice2; segments: 3) -> Insert -> Seq Scan on foo SubPlan 1 (slice2; segments: 3) -> Materialize -> Broadcast Motion 3:3 (slice1; segments: 3) -> Seq Scan on int8_tbl * rpt: add test case for with both PRIMARY and UNIQUE. On a replicated table we could set both PRIMARY KEY and UNIQUE constraints, test cases are added to ensure this feature during future development. (cherry picked from commit 72af4af8)
-
由 Ning Yu 提交于
When altering a table's distribution policy we might need to reorganize the data by creating a __temp__ table, copying the data to it, then swap the underlying relation files. However we always create the __temp__ table as permanent, then when the original table is temp the underlying files can not be found in later queries. CREATE TEMP TABLE t1 (c1 int, c2 int) DISTRIBUTED BY (c1); ALTER TABLE t1 SET DISTRIBUTED BY (c2); SELECT * FROM t1;
-
- 28 5月, 2018 2 次提交
-
-
由 Ning Yu 提交于
* rpt: reorganize data when ALTER from/to replicated. There was a bug that altering from/to a replicated table has no effect, the root cause is that we did not change gp_distribution_policy neither reorganize the data. Now we perform the data reorganization by creating a temp table with the new dist policy and transfering all the data to it. * rpt: support RETURNING for replicated tables. This is to support below syntax (suppose foo is a replicated table): INSERT INTO foo VALUES(1) RETURNING *; UPDATE foo SET c2=c2+1 RETURNING *; DELETE * FROM foo RETURNING *; A new motion type EXPLICIT GATHER MOTION is introduced in EXPLAIN output, data will be received from one explicit sender in this motion type. * rpt: fix motion type under explicit gather motion. Consider below query: INSERT INTO foo SELECT f1+10, f2, f3+99 FROM foo RETURNING *, f1+112 IN (SELECT q1 FROM int8_tbl) AS subplan; We used to generate a plan like this: Explicit Gather Motion 3:1 (slice2; segments: 3) -> Insert -> Seq Scan on foo SubPlan 1 (slice2; segments: 3) -> Gather Motion 3:1 (slice1; segments: 1) -> Seq Scan on int8_tbl A gather motion is used for the subplan, which is wrong and will cause a runtime error. A correct plan is like below: Explicit Gather Motion 3:1 (slice2; segments: 3) -> Insert -> Seq Scan on foo SubPlan 1 (slice2; segments: 3) -> Materialize -> Broadcast Motion 3:3 (slice1; segments: 3) -> Seq Scan on int8_tbl * rpt: add test case for with both PRIMARY and UNIQUE. On a replicated table we could set both PRIMARY KEY and UNIQUE constraints, test cases are added to ensure this feature during future development.
- 26 5月, 2018 1 次提交
-
-
由 Ashwin Agrawal 提交于
-
- 25 5月, 2018 3 次提交
-
-
由 Daniel Gustafsson 提交于
Running the full ICW under Travis was problematic, but the unittests doesn't require a running cluster so let's run them to boost our test coverage a little. This changes the invocation of mocker.py to accomodate how Travis need it to be, without changing the effect of it.
-
由 Ning Yu 提交于
* gdd: alloc MyProcPort on stack. We used to allocate MyProcPort with malloc() and does not check for the result, it's reported to trigger a crash in gprecoverseg test: (gdb) bt 0x00007f56b506794f in __strlen_sse42 () from /lib64/libc.so.6 0x00000000009f7894 in write_message_to_server_log () 0x00000000009fc200 in send_message_to_server_log () 0x00000000009ffb85 in EmitErrorReport () 0x00000000009fccf5 in errfinish () 0x0000000000acd99f in FtsTestSegmentDBIsDown () Now we allocate it directly on stack. We can not find out a way to construct a testcase for this issue, but according to the report it should be covered in existing tests. * gdd: store name of super user on stack. We used to store name of super user in a buffer allocated by strdup() and did not check for the result, it's reported to trigger a crash in gprecoverseg test. Now we store it in a buffer on stack. A fatal error will be triggered if no super user can be found. We can not figure out a way to construct a testcase for this change, but according to the report it should be covered in existing tests.
-
由 Jimmy Yih 提交于
The Postgres 9.1 merge introduced a problem where issuing a COPY FROM to a partition table could result in an unexpected error, "ERROR: extra data after last expected column", even though the input file was correct. This would happen if the partition table had partitions where the relnatts were not all the same (e.g. ALTER TABLE DROP COLUMN, ALTER TABLE ADD COLUMN, and then ALTER TABLE EXCHANGE PARTITION). The internal COPY logic would always use the COPY state's relation, the partition root, instead of the actual partition's relation to obtain the relnatts value. In fact, the only reason this is intermittently seen is because the COPY logic, when working on the leaf partition's relation that has a different relnatts value, was looking beyond a boolean array's allocated memory and got a phony value that would evaluate to TRUE. Co-authored-by: NJimmy Yih <jyih@pivotal.io> Co-authored-by: NAshwin Agrawal <aagrawal@pivotal.io>
-
- 24 5月, 2018 5 次提交
-
-
由 Daniel Gustafsson 提交于
gpfdist was mistankenly misspelled as gpfidst, this fixes all found occurrences in code and documentation.
-
由 Adam Lee 提交于
It's broken due to a gppylib issue which is not going to be fixed.
-
由 Jimmy Yih 提交于
To verify that autovacuum actually freezes template0, we used to just busy wait for about two minutes, expecting to observe the change of pg_database.datfrozenxid. While this "usually works", it's too sensitive to the amount of time it takes to vacuum freeze template0. Specifically, in some of our very I/O-deprived environments, this process sometimes takes slightly longer than two minutes. This patch introduces a fault injector to help us observe the expected vacuuming. The wait-in-a-loop is still there, but the bulk of the uncertain timing is now before the loop, not during the loop. Co-authored-by: NJesse Zhang <sbjesse@gmail.com> Co-authored-by: NJimmy Yih <jyih@pivotal.io>
-
由 Jim Doty 提交于
The gen pipeline script outputs a suggested command when setting a dev pipeline. Currently the git remote and git branch have to be edited before executing the command. Since often times the branch has been created and is tracing remote, it is possible to guess those details. The case statements attempt to prevent suggesting using the production branches, and fall back to the same string as before. Authored-by: NJim Doty <jdoty@pivotal.io>
-
由 Mel Kiyama 提交于
-Move cast examples to install guide -Pointer from install guide to admin guide for cast description.
-
- 23 5月, 2018 5 次提交
-
-
由 Todd Sedano 提交于
Fixed typo [ci skip] Authored-by: NTodd Sedano <tsedano@pivotal.io>
-
由 Ashwin Agrawal 提交于
This was left behind to get converted to seconds by commit 7d59d215ec065983c666b80bc2c982d13e476c48. Most likely reason for gpexpand jobs.
-
由 Ashwin Agrawal 提交于
This is cherry-pcik of upstream commit c61559ec. --------------------- Reduce pg_ctl's reaction time when waiting for postmaster start/stop. pg_ctl has traditionally waited one second between probes for whether the start or stop request has completed. That behavior was embodied in the original shell script written in 1999 (commit 5b912b08) and I doubt anyone's questioned it since. Nowadays, machines are a lot faster, and the shell script is long since replaced by C code, so it's fair to reconsider how long we ought to wait. This patch adjusts the coding so that the wait time can be any even divisor of 1 second, and sets the actual probe rate to 10 per second. That's based on experimentation with the src/test/recovery TAP tests, which include a lot of postmaster starts and stops. This patch alone reduces the (non-parallelized) runtime of those tests from ~4m30s to ~3m5s on my machine. Increasing the probe rate further doesn't help much, so this seems like a good number. In the real world this probably won't have much impact, since people don't start/stop production postmasters often, and the shutdown checkpoint usually takes nontrivial time too. But it makes development work and testing noticeably snappier, and that's good enough reason for me. Also, by reducing the dead time in postmaster restart sequences, this change has made it easier to reproduce some bugs that have been lurking for awhile. Patches for those will follow. Discussion: https://postgr.es/m/18444.1498428798@sss.pgh.pa.us ---------------------
-
由 Jim Doty 提交于
Authored-by: NJim Doty <jdoty@pivotal.io>
-
由 Jim Doty 提交于
Authored-by: NJim Doty <jdoty@pivotal.io>
-
- 22 5月, 2018 7 次提交
-
-
由 Jamie McAtamney 提交于
This test explicitly creates some tables and verifies that they can be read from after gpaddmirrors has been run on the cluster. The existing TINC test that does this will be deleted after all the new gpaddmirrors tests have been added. We've also added some code to allow cluster creation on more than one segment host. Co-authored-by: NNadeem Ghani <nghani@pivotal.io> Co-authored-by: NJamie McAtamney <jmcatamney@pivotal.io>
-
由 Jinbao Chen 提交于
-
由 Richard Guo 提交于
By design, all array types should be hashable by GPDB. Issue 3741 shows that GPDB suffers from text array type being not hashable error. This commit fixes that.
-
由 Jacob Champion 提交于
AIX builds are currently failing with the following error message: ../../config/install-sh: utils/errcodes.h does not exist. There is currently a hardcoded list of headers that must be generated before installing src/include on AIX. errcodes.h came in as part of the 9.1 merge, and it was missed in this list. For good measure, add this to the Windows build as well (it's not currently failing there) so that some future unrelated change doesn't introduce this same build failure. Co-authored-by: NAsim Praveen <apraveen@pivotal.io>
-
由 Jacob Champion 提交于
There are no distributed transactions during binary upgrade, so we can ignore them.
-
由 Jacob Champion 提交于
VACUUM FREEZE must function correctly during binary upgrade so that the new cluster's catalogs don't contain bogus transaction IDs. Do a simple check on the QD in our test script, by querying the age of all the rows in gp_segment_configuration.
-
由 Jacob Champion 提交于
As part of the merge with PostgreSQL 9.1, two gpdb/master commits were incorrectly reverted. This change replays the following commits: commit dd42f3d4 Author: Todd Sedano <tsedano@pivotal.io> Date: Thu May 17 10:47:30 2018 -0700 Syncing python-dependencies with pythonsrc-ext In comparing https://github.com/greenplum-db/pythonsrc-ext to python-dependencies.txt, there are several differences. pythonsrc-ext is the vendored python repo, so it should be in sync with pythonsrc-ext argparse is in pythonsrc-ext, adding it to python-dependencies.txt enum34, parse-type, ptyprocess, six are not in pythonsrc-ext, so moving them to python-developer-dependencies.txt Authored-by: NTodd Sedano <tsedano@pivotal.io> commit 7f9f11d3 Author: Daniel Gustafsson <dgustafsson@pivotal.io> Date: Fri May 18 09:43:48 2018 +0200 Fix typo in comment
-
- 21 5月, 2018 1 次提交
-
-
由 Hubert Zhang 提交于
Add hook framework in signal handler, e.g. StatementCancelHandler or die, to cancel the slow function in a 3rd party library. Add PLy_handle_cancel_interrupt hook to cancel plpython, which use Py_AddPendingCall to cancel embedded python asynchronously.
-