The advanced analytics provided by ADB is being used across many verticals, including finance, manufacturing, automotive, government, energy, education, retail, and so on, to address a wide variety of problems.
Some of the Arenadata DB analytics capabilities include the ability to analyze a multitude of data types, leverage existing SQL knowledge, and train more models in less time by using the MPP architecture.
Additionally, ADB provides in-database analytics which allows you to run analytics directly in the database vs exporting and running your data in an external analytics engine.
Arenadata DB is an excellent database for machine learning – the study of computer algorithms that improve automatically through experience. Apache MADlib is an open-source, SQL-based machine learning library that runs in-database on ADB, as well as on PostgreSQL.
This combination helps to improve the parallelism, scalability, and predictive accuracy of a machine learning deployment. Data transformation and feature engineering capabilities are also available through MADlib for machine learning, including descriptive and inferential statistics, pivoting, sessionization, and categorical variables encoding.
With ADB ability to ingest large volumes of data at high speeds, it makes this database a powerful tool for smart applications that need to interact intelligently based on an unlimited number of unique scenarios.
For example, a telecom company may use Arenadata DB AI capabilities in IoT (Internet of Things) systems with smart sensors to analyze and process events for maintenance, security, and operational efficiency purposes.
- configure backup schedules;
- manage backup configurations;
- create backups of different types (full, incremental, differential) on-demand;
- restore cluster databases from existing backups;
- perform audit of actions related to backups.
- high data transmission speed;
- automatic data schema generation;
- flexible partitioning;
- support for push-down operators;
- support for batch operations.
- ability to read and write AVRO data from Kafka topics;
- support for CSV and text formats in data read operations;
- support for transactions in Arenadata DB.
- transactionally load data by automatic creation of staging tables;
- use multiple table engine families in ClickHouse;
- flexibly distribute and parallelize the write load.
Arenadata DB
Compare with
Compare with
Via ADCM.
Only CLI.
ADB Control & Backup Manager
Compare with
Compare with
Via ADCM.
Full training on working with Arenadata products.
Complete release history with service versions and description of the upgraded functionality is available in the open domain.
“Product comparison” section is relevant on the date of 31.08.2023.
- Implemented ADB to ADB Connector 1.0.0
- Added the Uninstall action for Madlib
- Fixed the error that occurred when the session on the master node was ended, while the sessions on the segments continued to work
- The custom timezone setting is now available on the Configuration tab
- New parameters for the metrics offloading job (Export job) were added
- Changed the order of side menu items
- Deleting the last user with the Owner role is now prohibited
- Fixed the error of filtering by Job group on the Jobs → Audit page
- CIFS repository support was added
- Added the option of automated backup copying to a target cluster when a new backup on a source cluster is completed
- Auto detection of an active master host was added to support a backup process after switching to standby
- Fixed: Self-restore (via the Restore action) failed after changing roles on mirrors
- Added LDAP integration and synchronization with ADB from bundle
- Added the pgbouncer-adb6 RPM with LDAP native authentication
- gpbackup: added the --report-dir parameter to gprestore
- ADB is now available as a target database for exporting metrics
- Filters on operations/authorizations audit pages for Object name, Username, Host are now case insensitive and partial
- Fixed: Password for offload database was shown on Audit Operations page
- Fixed: Sorting by columns was not available on Jobs page
- Fixed: After selecting a cluster, agents for other clusters were displayed on Information page
- Added restore with mirrors option for DR copy process
- Repository validation step was added before saving backup configuration
- Backup topology is now sorted by segment_id
- Fixed: Restores list was empty for clusters with spaces in names
- Postgres config files are backed up for the purpose of ADBM Copy (Disaster Recovery) Fixed:
- Cluster was marked Down in ADBM when master was on the same host as segment
- New tables and filters were implemented Fixed:
- Execution time was not shown in plan nodes after the command finished
- Error when changing the user role
- Error when launching the metrics offload job
- Restore for standby master was implemented
- Upgraded gpbackup to 1.28
- Upgraded pxf to 6.6
- ADB ClickHouse connector went out of beta to GA
- Implemented the adb_collect_table_stats function to handle a problem with NULL values in db_files_current
- Implemented the adb_skew_coefficients view to monitor data skews as opposed to gp_toolkit.gp_skew_idle_fractions
- Audit for unsuccessful ADB logon attempts and Audit for ADB Control users management actions are available now
- Added an archiving action for a cluster to stop monitoring of the selected cluster with the ability to activate it again
- Added filtering of queries shorter than a threshold on the adcc-agent side to improve agents stability
- SQL signature (sql_id) was added for every command Fixed:
- Cluster information data didn’t update immediately after switching the cluster on Actions, Backups, Restores tabs
- Implemented Disaster Recovery with Cold Standby — ability to recover a cluster from a backup located on another cluster (Beta-version)
- Added pgbackrest options (-delta, --force, --max-process) for the restore action
- Added an option to save a backup configuration without schedule for running backups on request
- ADBM action tree view was enhanced
- Fixed: the Apply backup config action failed when S3 was selected
- Upgraded diskquota to 2.2.1
- Limited the find command depth when diskquota upgrading
- Fixed diskquota 2.x to allow a huge number of rows in diskquota.table_size
- Fixed: PXF curl error buffer: a transfer closed with external read data remaining
- Fixed: diskquota hung and blocked the cluster stopping
- Upgraded tkh-connector and gptkh to the latest PXF 5 versions
- Fixed FATAL: could not open relation after the segments switchover
- Fixed the gptkh extension creation when extension scripts dropped anything
- Fixed ERROR: gp_relation_node has more entries than pg_aocsseg for relations
- Implemented Tkhemali connector 2.0
- Started to process the IN predicate for filter push-down purposes in PXF JDBC
- Enabled SSL between Client and Master
- Added PXF Monitoring Grafana Dashboards
- Arenadata DB Command Center (ADBCC) is renamed to Arenadata DB Control (ADB Control)
- Optimized storage space with the ability to export metrics to an external database for a long-term storage
- Improved security with CSRF (Cross-Site Request Forgery) protection support
- Improved authentication security with forced password change at first login and user account blocking after several failed authentication attempts
- Added the ability to view total monitoring metrics as well as current-time recalculations for active commands
- Ability to use ADBM on PowerPC
- Ability to restore cluster without mirror segments
- Improved filtering for Restore actions
- Synced with upstream Greenplum Database 6.22.1
- Upgraded pgbouncer to 1.18
- Upgraded gpbackup to 1.27
- Upgraded plcontainer to 2.2
- Implemented a buffer parameter in the gpcheckperf utility
- Ability to restore from backups on a stopped cluster
- Ability to delete the last backup from the stanza
- Ability to restore specific databases from backups
- Added the ability to use ADB Control along with gpperfmon within the same ADB cluster
- LDAP search in several Organizational Units (OU) during authentication
- Flexible sorting on query and transaction monitoring pages
- Synced with upstream Greenplum Database 6.22.0
- Added support for AltLinux 8.4 SP
- Added Data Domain Boost 1.0.0
- PXF: allowed setting of Oracle parallel instructions
- Refactored Planchecker to use an external ADB Control database
- gpbackup: fixed the metadata order so that now gprestore can restore functions after the tables that are used in functions as a returning type
- gptkh: fixed fetching of the actual system.tables columns in ClickHouse (according to the ClickHouse version)
- Added new performance metrics for commands and transactions: Cpu usage total, Read bytes total, Write bytes total
- Added the ability to repeatedly change a resource group for a transaction
- Fixed calculating the number of tuples affected by the request
- Included the Planchecker database objects into migration
- Synced with upstream Greenplum Database 6.21.1
- Enabled core dump files for ADB processes
- Fixed the problem with loss of resource group slots when moving a query
- Added a Planchecker image to the ADCC service (docker-compose) in the ADB bundle
- Added the ability to cancel a transaction
- Added the ability to reassign queries to another resource group
- Added the ability to filter commands by a planner
- Synced with upstream Greenplum Database 6.21.0
- Optimized DML queries against partitioned tables to avoid further planning if a partition was pruned
- Excluded the gpmon background process from the shared memory user list
- Implemented a fallback to PostgreSQL for an empty target list in CTE producer
- ADB bundle: added the ability to specify a cluster network
- A non-blocking socket is now used to communicate with an agent
- Added monitoring of transactions
- Added monitoring of SQL statement groups: DDL, DML, DCL, and TCL
- Added Spill and Spill Skew calculation
- Synced with upstream Greenplum Database 6.20.1
- PXF: added PXF 6.3.0 to the ADB bundle (with ability to upgrade from PXF 5.x)
- PXF: activated a PXF cluster sync command
- PXF: added the ability to override data types mapping in external tables for PXF
- Added ADB ClickHouse connector 1.0.1
- The actual statistics from EXPLAIN ANALYZE for finished queries is processed now
- Integrated the average cluster query metrics
- Implemented compression for huge queries
- Synced with upstream Greenplum Database 6.19.3
- Added ADB Loader tools for RHEL 8
- Added the ability to deploy maintenance scripts for several databases
- Enabled the backlog_lock_waits GUC
- Data audit support
- Added the Background jobs history page
- Support for virtual process memory in system metrics of commands
- Service load ratio support
- Synced with upstream Greenplum Database 6.18.2
- Implemented archive_mode always
- Added Kafka ADB connector 1.0.4
- HTTPS support
- Implemented the backpressure mechanism relative to the memory volume occupied in heap on the agent
- Synced with upstream Greenplum Database 6.18.0
- gpbackup: added an explicit order of tables by using pg_class.relpages
- PXF: added the partitioning query support for Sybase
- Multi-clusters
- New system metrics in the query context: CPU, RAM, IO
- Actualization for hanging queries
- Adding columns dynamically to the History and Monitoring pages
- Synced with upstream Greenplum Database 6.17.5
- ADB bundle: added the Ready to upgrade status for a bundle upgrade action
- Fixed low CPU performance on Power with newly added CGLAGS build options
- Added a build for Power8 LE platform (ppc64le arch)
- Added an agent build for Power8 LE platform (ppc64le arch)
- Synced with upstream Greenplum Database 6.17.1
- ADB bundle: external database connection for ADB Control
- Fixed: PostgreSQL query optimizer built a bad plan for replicated tables with indexes
- adcc-extension: started to retrieve and send an error text
- New user interface
- Ability to cancel and terminate queries
- Time-based Retention Policy
- Extended information on errors
- LDAP authentication
- Synced with upstream Greenplum Database 6.16.2
- Shrinking of relation segment files to zero on TRUNCATE and DELETE
- PXF: removed the tuple count check for JDBC queries INSERT
- Kafka ADB Connector: allowed users to set custom librdkafka options
- Implemented the diskquota extension update
- Added the gp_enable_gpperfmon=on parameter to Master and Segment servers
- Added JVM arguments for logging
- Synced with upstream Greenplum Database 6.15.0
- Implemented switchover from Master to Standby via ADCM
- Added the $PXF_CONF and the $PXF_HOME environment variables to PXF hosts
- Kafka ADB connector: implemented signal handlers to interrupt consuming
- Ported ADB to Alt Linux 8.2
- Support for horizontal scaling of ADB Control backend
- Added the UDS unlink processing, updated the library build for CentOS 7
- Added the innerQueueCapacity parameter that defines the internal message queue size for an agent
- Synced with upstream Greenplum Database 6.14.1
- Removed online loading of static resources from the ADB Control web interface
- Synced with upstream Greenplum Database 6.14.0
- Synced with upstream Greenplum Database 6.13.0
- Supported auxiliary relations for append-optimized tables to be used by the pgstattuple extension to exactly estimate the bloat of those relations
- ADB ClickHouse connector: avoided the intermediate conversion to a byte array
- Removed the obsolete batching options in ADQM connector (since the TEXT is used now)
- Synced with upstream Greenplum Database 6.12.1
- Added the offset function to Kafka connector
- Added the rest committed function to Kafka ADB connector
- Implemented the text format for Kafka ADB connector
- Provided the JVM_OPTS setting for PXF
- Supported AVRO logical types in Kafka ADB connector
- Started to use rd_kafka_query_watermark_offsets to validate partition-offset pairs in Kafka ADB connector