PDaiPostgres provides seamless PostgreSQL integration for the PDai analytics ecosystem. This package enables enterprise-grade database connectivity, optimized data processing, and real-time analytics capabilities for production environments.
- High-Performance Connectivity: Optimized connection pooling and query execution
- Data Pipeline Integration: Seamless ETL/ELT workflows with PDai analytics
- Scalable Architecture: Built for enterprise workloads with millions of records
- Security First: SSL/TLS encryption, role-based access control, and audit logging
- Real-Time Processing: Stream processing capabilities for live data analysis
- Smart Caching: Intelligent query result caching for improved performance
- R (>= 4.0.0)
- PostgreSQL (>= 12.0)
- PDai base package
# Install devtools if not already installed
if (!require(devtools)) {
install.packages("devtools")
}
# Install PDaiPostgres from GitHub
devtools::install_github("embeddedlayers/package-PDaiPostgres")# Load the package
library(PDaiPostgres)
# Configure database connection
conn <- pdai_pg_connect(
host = "your-postgres-host",
port = 5432,
dbname = "your-database",
user = "your-username",
password = "your-password",
ssl = TRUE
)
# Execute analytics pipeline
results <- pdai_pg_pipeline(
connection = conn,
query = "SELECT * FROM sales_data",
analytics = list(
predict = TRUE,
visualize = TRUE,
export = "html"
)
)
# View results
print(results$summary)pdai_pg_connect(): Establish secure database connectionpdai_pg_pool(): Create connection pool for concurrent operationspdai_pg_disconnect(): Safely close connections
pdai_pg_query(): Execute optimized SQL queriespdai_pg_stream(): Stream large datasets efficientlypdai_pg_write(): Bulk insert/update operationspdai_pg_upsert(): Intelligent upsert operations
pdai_pg_pipeline(): Run PDai analytics on PostgreSQL datapdai_pg_aggregate(): Perform in-database aggregationspdai_pg_ml(): Execute machine learning models in-databasepdai_pg_cache(): Manage analytics result caching
pdai_pg_monitor(): Monitor query performancepdai_pg_optimize(): Automatic query optimizationpdai_pg_audit(): Audit trail for compliance
Create a .pdai_pg_config file in your project root:
default:
host: localhost
port: 5432
ssl: true
pool_size: 10
timeout: 30
production:
host: prod-server.example.com
port: 5432
ssl: required
pool_size: 50
timeout: 60- Use connection pooling for applications with multiple concurrent users
- Enable query caching for frequently accessed data
- Leverage in-database analytics to minimize data transfer
- Use streaming for large dataset processing
- Enable compression for network traffic optimization
- Always use SSL/TLS connections in production
- Implement role-based access control (RBAC)
- Use environment variables for credentials
- Enable audit logging for compliance
- Regularly update both PDaiPostgres and PostgreSQL
Comprehensive documentation available at:
- Issues: GitHub Issues
- Email: support@embeddedlayers.com
- Community: Discussions
We welcome contributions! Please see CONTRIBUTING.md for guidelines.
This project is licensed under the MIT License - see the LICENSE file for details.
Copyright (c) 2024 PeopleDrivenAI LLC (DBA EmbeddedLayers)
- PDai: Core analytics package
- MCP Analytics: Statistical analysis tools for AI assistants
Part of the EmbeddedLayers Analytics Ecosystem