Difference between revisions of "TODO"
(→Pgpool-II TODO list) |
(→TODOs already done) |
||
Line 122: | Line 122: | ||
== TODOs already done == | == TODOs already done == | ||
+ | |||
+ | === Allow to send relation cache query to other than primary node === | ||
+ | : Pgpool-II needs to access PostgreSQL's system catalog to obtain meta info. For now the query is always sent to primary. This is good because it could avoid replication delay for newly created tables. However if primary PostgreSQL is geographically distant, the query could take long time. It would be nice if there's a parameter to allow send such queries to other than primary node. | ||
+ | : This has been implemented in 4.1. | ||
=== Automatically reattach a node in streaming master/slave configuration === | === Automatically reattach a node in streaming master/slave configuration === |
Revision as of 04:19, 15 October 2019
Contents
- 1 Pgpool-II TODO list
- 1.1 Allow to use multiple pgpool-II instances with in-memory query cache enabled
- 1.2 Allow to use pg_rewind in online recovery
- 1.3 Support peer auth
- 1.4 Allow to use client encoding
- 1.5 Recognize multi statemnet queries
- 1.6 Cursor statements are not load balanced, sent to all DB nodes in replication mode
- 1.7 Support IPv6 network
- 1.8 Handle abnormal down of virtual IP interface when watchdog enabled
- 1.9 Do not invalidate query cache created in a transaction in some cases
- 1.10 Fix memory leak in pool_config.c
- 1.11 Put together a definition of error codes into a single header file
- 1.12 Import PostgreSQL's latch module
- 1.13 Allow to use schema qualifications in black_function_list and white_function_list
- 1.14 Support multiple UNIX domain socket directories
- 1.15 Implement "log_timezone"
- 1.16 Allow to reset in memory query cache in the shared memory without restarting Pgpool-II
- 1.17 Do not prevent load balancing in explicit transactions in certain cases
- 1.18 Allow to use comma separated IP address or host names in listen_addresses
- 1.19 Add black/white table list for load balancing control
- 1.20 Support for CRL (Certificate Revocation List)
- 1.21 Support Cert authentication between Pgpool-II and PostgreSQL
- 1.22 Include other file in pgpool.conf file
- 1.23 Detach the standby node with large replication lag
- 1.24 Allow to get primary node info in failback_command script.
- 1.25 Add support for an user/password input file to pg_md5
- 1.26 Change relative path of ssl_key and ssl_cert to DEFAULT_CONFIGDIR
- 1.27 Support for SSL passphrase
- 1.28 Allow to send relation cache query to other than primary node
- 1.29 Allow to specify whether a relcahe entery is for a global table
- 1.30 Grouping config requires duplicate codings
- 1.31 Duplicate functionality: "show pool_status" and "pgpool show all"
- 1.32 Allow to set application name in log_line_prefix
- 1.33 Set ps status in extended query
- 2 TODOs already done
- 2.1 Allow to send relation cache query to other than primary node
- 2.2 Automatically reattach a node in streaming master/slave configuration
- 2.3 Move relation cache to shared memory
- 2.4 Allow to specify which node is dead when starting up
- 2.5 Ability to load balance based on Client IP, database, table etc.
- 2.6 Import PostgreSQL's execption handling
- 2.7 Allow to print user name in the logging
- 2.8 Remove on disk query cache
- 2.9 Restart watchdog process when it abnormaly exits
- 2.10 Synchronize backend nodes information with watchdog when standby pgpool starts up
- 2.11 Avoid multiple pgpools from executing failover.sh simultaneously.
- 2.12 Add new parameter for searching primary node timeout
- 2.13 Allow to load balance even in an explicit transaction in replication mode
- 2.14 Add testing framework
- 2.15 Add switch to control select(2) time out in connecting to PostgreSQL
- 2.16 Allow to specify which node is dead when starting up
- 2.17 Remove parallel query
- 2.18 Enhance pcp commands
- 2.19 Enhance performance of extended protocol case
- 2.20 Import PostgreSQL 9.5's parser
- 2.21 Watchdog feature enhancement
- 2.22 Allow to specify user name, password and database name for health check per backend base
- 2.23 Enhance documents
- 2.24 Add SET commnad
- 2.25 Send read query only to standbys even after fail over
- 2.26 Do not disconnect to clients when a fail over happens
- 2.27 Create separate process for health checking
- 2.28 Health-check timeout for each backend node
- 2.29 Support SCRAM authentication
- 2.30 Allow to choose load balance behavior per SQL statement
- 2.31 Allow to specify load balance weight ratio for load balance parameters
- 2.32 Support for SSL using ECDH
Pgpool-II TODO list
Allow to use multiple pgpool-II instances with in-memory query cache enabled
- For this purpose we not only use memcached but also we need to store the oid map info on it to share the info among pgpool-II instances.
- According to https://www.pgpool.net/pipermail/pgpool-hackers/2018-November/003143.html , attempt to put oid map into memcached was failed due to reliability and performance reason. Maybe we should try with more reliable in memory storage engine, such as Redis.
Allow to use pg_rewind in online recovery
- pg_rewind could speed up online recovery. However it only works when the target node was normally shut down. Can we recognize that?
- Probably yes by looking at pg_controldata.
Support peer auth
- Apparently pool_hba.conf should recognize it if we are going to support it. Also pgpool-II should forward it to PostgreSQL. We need think the case if pg_hba.conf does not use peer auth.
Allow to use client encoding
- It would be nice if pgpool client could use encoding which different from PostgreSQL server encoding.
- To implement this, the parser should be able to handle "unsafe" encodings such as Shift_JIS. psql replaces second byte of each multibyte character to fool the parser. We could hire similar strategy.
Recognize multi statemnet queries
- As stated in the document, pgpool-II does not recognize multi statement queries correctly (BEGIN;SELECT 1;END). Pgpool-II only parses the first element of the query ("BEGIN" in this case) and decides how to behave.
- Of course this will bring various problems. It would be nice if pgpool-II could understand the each part of the multi statement queries.
- Problem is, how PostgreSQL backend handles the multi statement queries. For example, when client sends BEGIN;SELECT 1;END, backend returns "Command Complete" respectively and "Ready for query" is returned only once. Thus, trying to split multi statement queries to non multi statement queries like what psql is doing will not work.
- Simon Riggs suggested that if Pgpool-II cannot process multi-statement query properly, then it should have an option to prohibit the multi stattement queries in the developer unconference held in PGConf.ASIA 2016 on December 1st 2016 in Tokyo. (or maybe we could disregard the 2nd or later queires instead).
Cursor statements are not load balanced, sent to all DB nodes in replication mode
- DECLARE..FETCH are sent to all DB nodes in replication mode. This is because the SELECT might come with FOR UPDATE/FOR SHARE.
- It would be nice if pgpool-II checks if the SELECT uses FOR UPDATE/FOR SHARE and if not, enable load balance (or only sends to the master node if load balance is disabled).
- Note that some applications including psql could use CURSOR for SELECT. For example, from PostgreSQL 8.2, if "\set FETCH_COUNT n" is executed, psql unconditionaly uses a curor named "_psql_cursor".
Support IPv6 network
- As of 3.4, it is allowed to use IPv6 address for PostgreSQL backend server and bind address of pgpool-II itself.
- However, PCP process still only binds to IPv4 and UNIX domain socket. Same thing can be said to watchdog,
Handle abnormal down of virtual IP interface when watchdog enabled
- When virtual IP interface is dropped abnormally by manual ifconfig etc., there are no one holding VIP, and clients aren't able to connect pgpool-II. Watchdog of active pgpool should monitor the interface or VIP, and handle its down.
Do not invalidate query cache created in a transaction in some cases
- Currently new query cache for table t1 created in a transaction is removed at commit if there's DMLs which touch t1 in the same transaction. Apparently this is overkill for same cases:
BEGIN; INSERT INTO t1 VALUES(1); SELECT * FROM t1; COMMIT;
- To enhance this, we need to teach pgpool-II about "order of SELECTs and DMLs.".
Fix memory leak in pool_config.c
- The module in charge of parsing pgpool.conf has memory leak problem. Usually pgpool reads pgpool.conf just once at the start up time, it is not a big problem. However reloading pgpool.conf will leak memory and definitely a problem. Also using memory leak check tools like valgrind emit lots of error messages and very annoying. So it would be nice to fix the problem in the future.
Put together a definition of error codes into a single header file
- Currently most error codes used by pool_send_{error,fatal}_message() etc (e.g. "XX000", "XX001", "57000") are hard-coded in different sources. They should be defined as constants in a single header together.
Import PostgreSQL's latch module
- Pgpool already has similar module but PostgreSQL's one seems more sophiscated and reliable.
Allow to use schema qualifications in black_function_list and white_function_list
- Currently schema qualifications are silently ignored in these parameter.
Support multiple UNIX domain socket directories
- PostgreSQL already does this. See pgpool-hackers: 1433.
Implement "log_timezone"
- (From pgpool-genera: 5215)
I'd like to propose that an addition be made to pgpool to allow for a log timestamp to be written to the log with a timezone other than the locally defined timezone. Where this is helpful is when we use an external tool like logstashforwarder, where we want the logs to be absorbed with a timestamp with a UTC timezone. Postgres offers this feature ('log_timezone'), which we use, and it would be nice to allow pgpool to behave in the same way.
Do not prevent load balancing in explicit transactions in certain cases
- If write queries are issued in an explicit transaction, Following SELECTs are not load balanced, rather sent to primary node. This is intended to allow SELECTs to retrieve the latest data regardless the replication delay. Currently "write query" includes anything other than SELECTs. This is overkill for some class of queries: for example, Since SET command are sent to both primary and standby nodes, sending SELECTs to any of DB nodes could retrieve the latest data.
Allow to use comma separated IP address or host names in listen_addresses
- Currently only single IP or host name or '*' is allowed. PostgreSQL already allows multiple listen addresses.
Add black/white table list for load balancing control
- If a table is in the black list, always send queries to primary server. Probably database.schema.table notion is preferable.
Support for CRL (Certificate Revocation List)
- Our SSL support lacks this (PostgreSQL already has this) and supporting CRL should make Pgpool-II more secure.
Support Cert authentication between Pgpool-II and PostgreSQL
- Pgpool-II 4.0 added support for Cert authentication between frontend and Pgpool-II, but between Pgpool-II and backend is not yet supported.
Include other file in pgpool.conf file
- Add the feature pgpool.conf can include other file, which specify backendname and host specific setting values.
Detach the standby node with large replication lag
- Now no loadbalance to the standby node with large replication lag. But if due to some reason of online-recovery the recoveroed standby node can't connect to primary node, the standby node should be detached.
Allow to get primary node info in failback_command script.
- Now we can get master node info in failback_command script, it will be more useful to get hostname, port and database cluster directory of new primary node.
Add support for an user/password input file to pg_md5
pg_md5 -m -f conf/pgpool.conf --input-file=users.txt
Change relative path of ssl_key and ssl_cert to DEFAULT_CONFIGDIR
- Currently the relative path of ssl_key and ssl_cert are the path to the directory to run pgpool. Change this relative path to DEFAULT_CONFIGDIR. And change default value to use absolute path.
Support for SSL passphrase
- Using passpharase encrypt private key is more sucure. PostgreSQL already has this. Pgpool-II should import it.
Allow to send relation cache query to other than primary node
- Pgpool-II needs to access PostgreSQL's system catalog to obtain meta info. For now the query is always sent to primary. This is good because it could avoid replication delay for newly created tables. However if primary PostgreSQL is geographically distant, the query could take long time. It would be nice if there's a parameter to allow send such queries to other than primary node.
- This has been implemented in 4.1.
Allow to specify whether a relcahe entery is for a global table
- Currently relcache is defined for each database. However some of relcache entry does not depend on databases: for example shared catalogs and misc info including PostgreSQL version. For such info having per-database relcache entry is not only waste of resource, but less efficient. It is desirable to be able to specify if a relcache entry does not depend on databases.
Grouping config requires duplicate codings
- pool_config_variable.c manages each config variables along with its group to belong to. However each group definition also requires which config variable belongs to the group. This is redundant and should be avoided.
Duplicate functionality: "show pool_status" and "pgpool show all"
- Both commands produce almost same output except that "show pool_status" lacks some variables because certain config variables were forgotten to be added. Probably we should keep "pgpool show all" only because it does not require to maintaining pool_process_reporting.c. To keep backward compatibility, if "show pool_status" is requested, "pgpool show all" could be called.
Allow to set application name in log_line_prefix
- Currently, appication name (%a) can only be set if startup packet includes it. It would be nice if Pgpool-II traps "set application_nane..." in the current session and allows log_line_prefix to use it.
Set ps status in extended query
- Currently ps status, sucuh as "SELECT" etc. is only set when simple query is executed. It would be nice if the ps status is set while executing extended queries.
TODOs already done
Allow to send relation cache query to other than primary node
- Pgpool-II needs to access PostgreSQL's system catalog to obtain meta info. For now the query is always sent to primary. This is good because it could avoid replication delay for newly created tables. However if primary PostgreSQL is geographically distant, the query could take long time. It would be nice if there's a parameter to allow send such queries to other than primary node.
- This has been implemented in 4.1.
Automatically reattach a node in streaming master/slave configuration
- In streaming master/slave configuration there could be an option to automatically reattach a node if it's up-to-date with the master (0 bytes behind). It often happens that due to minor network outage a slave node is dropped off from pgpool and stays down even if the the node has resumed replication with master and is up-to-date.pgpool already knows how much slave is behind master so i guess this wouldn't be too difficult to implement? (from bugtrack #17)
- Another concern is whether the standby in question actually connects to the proper primary server or not. It is possible that the standby is up and running but is connected to different primary server. Simon Riggs suggested at the developer unconference on December 1st 2016 held in PGConf.ASiA 2016 in Tokyo that pg_stat_wal_receiver, which is new in PostgreSQL 9.6, can be used to safely judge that the standby in question is actually connected to appropriate primary server.
- pg_stat_replication provides ideal information for this purpose. By using it this will be supported in
- This has been implemented as "auto_failback" 4.1.
- This will bring less inquiry to the system catalogue (thus better performance) and more real-time cache invalidation.
- This has been implemented in 4.1 as "enable_shared_relcache".
Allow to specify which node is dead when starting up
- If we set longer health check timeout and/or many health check retries, starting up pgpool-II will take long time if some of DB nodes have been down because of health checking and retries in creating connection to backend.
- pgpool_status should help here but for the very first starting up, we cannot use it.
- It would be nice if we could tell pgpool-II about down node info.
- As of 3.4, pgpool_status file is changed to a plain ASCII file and users could specify down node by using ordinary text editors.
Ability to load balance based on Client IP, database, table etc.
- From bugid 26: I have recently moved a database from Mysql to postgresql 9.1.5 which is behind a pgpool-II-3.1.4 . Everything went fine until i observed that some "tickets" are not created correctly by the application (OTRS) that populate the database.
- After some debugging i found/guess that the problem is the following:
- when a cron job wants to create a ticket he has to insert info in abut 10 tables, and i guess that the 2-nd, 3-rd ... inserts depends on the first. The problem was that this operation is not performed transactionally so after the first insert, when the app tries to perform the other inserts, first tries to select "the first insert", but this first insert is still not propagated to all nodes, and the error occurs.
- I`m aware of the fact that if this entire operation would be performed transactionally (only on master) the issue is solved, but unfortunately i cannot modify the app.
- So i want to know if there is any way that i can tell to pgpool something like :
- any request from this ip do not load balance.
- PS. temporary i have set the weight factor to 0 to the 2-nd and 3-rd postgresql slaves and it behaves ok, because reads and writes only from master.
- P.P.S. there's also different request regarding load balance.
- http://www.pgpool.net/pipermail/pgpool-general/2014-June/003032.html
- This item has been implemented in 3.4 as "database_redirect_preference_list" and "app_name_redirect_preference_list".
Import PostgreSQL's execption handling
- PostgreSQL's exception handling (elog family) is pretty good tool to make codes to be simple and robust. It would be nice if pgpool could use this. This has been already done in 3.4.
Allow to print user name in the logging
- This will be useful for audit purpose. (done and will appear in pgpool-II-3.4.0).
Remove on disk query cache
- Old on disk query cache has almost 0 user and has sevior limitation, including no automatic cache invalidation. This has been already obsoleted since on memory query cache implemented. We should remove this (this has been already in git master and will appear in 3.4.0).
Restart watchdog process when it abnormaly exits
- It would be nice for pgpool main to restart watchdog process when it dies abormaly.
Synchronize backend nodes information with watchdog when standby pgpool starts up
- For example, when a certain node is detached from active pgpool and then standby pgpool starts up, the standby pgpool can't recognized that the node is detached. Standby pgpool should get information about node information from other pgpool.
Avoid multiple pgpools from executing failover.sh simultaneously.
- In master-slave mode with watchdog, when a backend DB is down, all pgpools execute failover.sh. It might cause something wrong.
Add new parameter for searching primary node timeout
- pgpool-II uses "recovery_timeout" for searching the primary node timeout after failover. Since this is an abuse of the parameter, we should add new parameter for searching the primary node.
Allow to load balance even in an explicit transaction in replication mode
- Currently load balance in an explicit transaction is only allowed in master-slave mode. It should be allowed in the replication mode as well.
Add testing framework
- PostgreSQL has nice regression test suite. It would be nice if pgpool-II has similar test suite. Problem is, such a suite could be very complex system because it should include not only pgpool-II itself, but also multiple PostgreSQL instances. Also don't forget about "watchdog". Even such a test suite should be able to manage multiple pgpool-II instances.
Add switch to control select(2) time out in connecting to PostgreSQL
- In connect_inet_domain_socket_by_port(), select(2) is issued to watch events on the fd created by non blocking connect(2). The time out parameter of select(2) is fixed to 1 second, which is not long enough in flakey network environment like AWS (http://www.pgpool.net/pipermail/pgpool-general/2014-May/002880.html).
- To solve the problem, new switch to control the time out is desired (done for pgpool-II 3.4.0).
Allow to specify which node is dead when starting up
- If we set longer health check timeout and/or many health check retries, starting up pgpool-II will take long time if some of DB nodes have been down because of health checking and retries in creating connection to backend.
- pgpool_status should help here but for the very first starting up, we cannot use it.
- It would be nice if we could tell pgpool-II about down node info (pgpool-II 3.4.0 chages the pgpool_status format to ASCII. Thus users can edit the file if needed).
Remove parallel query
- Parallel query has severe restrictions such as certain queries cannot be used, nor in extended protocol (i.e. JDBC).
- Also it is pain to upgrade to newer version of PostgreSQL's SQL parser (yes, pgpool-II uses PostgreSQL's parser code). In short, parallel query gives us small gain comparing with the work needed to maintain/enhance. So I would like to obsolete parallel query in the future pgpool-II release. (related parameters have been removed from pgpool.conf in 3.4.0. pgpool-II 3.5.0 will remove actual code).
Enhance pcp commands
- There are number of drawbacks in pcp commands including 1)the timeout parameter is not used any more and should be removed 2)error codes returned from the commands are completely useless 3)multiple commands can not be accepted simultaneously.
- This has been already done in 3.5.
Enhance performance of extended protocol case
- When extended protocl (i.e. JDBC etc.) used, pgpool-II's overhead is pretty large compared with simple query. Need to enhance it.
- This has been already done in 3.5.
Import PostgreSQL 9.5's parser
- No need to say for this.
- This has been already done in 3.5.
Watchdog feature enhancement
- Watchdog is a very important feature of pgpool-II as it is used to eliminate the single point of failure and provide HA. But there are few feature requests and bugs in the existing watchdog that require little more than a simple code fix, and requires the complete revisit of its core architecture.
- See the design proposal for watchdog enhancement [here]
- This has been already done in 3.5.
Allow to specify user name, password and database name for health check per backend base
- In some environment it is not allowed to access standard database i.e. postgres and template1. So users need to specify them per backend basis.
- Maybe we need backend_healthcheck_username0 etc? See http://www.pgpool.net/pipermail/pgpool-hackers/2015-June/000942.html for more details.
- This has been already done in 3.5.
Enhance documents
- The current document for is plain HTML, which is a real pain to maintain. Like PostgreSQL, is SGML our direction?
- Pgpool-II 3.6 is going to change the document format to SGML. (This has been already implemented in 3.6. We employ SGML).
Add SET commnad
- Pgpool specific SET command would be usefull. For example, using "SET debug = 1" could produce debug info on the fly for particular session.
- This is being discussed in pgpool-II 3.6 development. (This item has been implemented in 3.6)
Send read query only to standbys even after fail over
- We can configure pgpool-II to not send read queries to the primary. However after a fail over, the role of the node could be changed.
- To solve the problem, we need new flag to specify that read queries always are sent to standbys regardless the fail over ([pgpool-general: 1621] backend weight after failover).
- (This has been already implemented in 3.4 as "database_redirect_preference_list" and "app_name_redirect_preference_list")
Do not disconnect to clients when a fail over happens
- At this moment we don't know how to implement it but this is a desirable feature.
- This has been already implemented in 3.6.
Create separate process for health checking
- To make main process more stable, it would be better to make separate process which is responsible for health checking.
- This has been already implemented in 3.7.
Health-check timeout for each backend node
- In the current, timeout values specified by health_check_timeout means the total time for checking all the backend status. Hence, if it takes a long time to succeed to check a backend, when timeout occurs during checking the next backend, this node is regarded as failed and failovered even though this is healthy.To resolve this issue, we need health-check timeout for each backend.
- This has been implemented in 3.7.
Support SCRAM authentication
- PostgreSQL 10.0 supports SCRM authentication. It seems there's fundamental difficulty with this.
- See http://www.pgpool.net/pipermail/pgpool-hackers/2017-May/002331.html for more details.
- This has been implemented in 4.0.
Allow to choose load balance behavior per SQL statement
- If a query string matches specified regular expression, send the query to either primary or standby.
- This has been implemented in 4.0.
Allow to specify load balance weight ratio for load balance parameters
- Allow to specify load balance weight ratio for database_redirect_preference_list, and app_name_redirect_preference_list like: "postgres:primary(0.3)".
- See https://www.pgpool.net/pipermail/pgpool-hackers/2017-December/002650.html
- This has been implemented in 4.0.
Support for SSL using ECDH
- ECDH is encryption algorithm. Our SSL support lacks this (PostgreSQL already has this) and supporting ECDH should make Pgpool-II more secure.
- This has been implemented in 4.1.
- commit: https://git.postgresql.org/gitweb/?p=pgpool2.git;a=commit;h=51bc494aaa7fd191e14038204d18effe2efb0ec8