Initial commit

This commit is contained in:
Zhongwei Li
2025-11-30 08:55:17 +08:00
commit 5f83ab42ed
12 changed files with 6080 additions and 0 deletions

View File

@@ -0,0 +1,703 @@
# Administration Reference
**Source**: [https://github.com/SAP-docs/sap-datasphere/tree/main/docs/Administering](https://github.com/SAP-docs/sap-datasphere/tree/main/docs/Administering)
---
## Table of Contents
1. [Tenant Configuration](#tenant-configuration)
2. [Spaces and Storage](#spaces-and-storage)
3. [Users and Roles](#users-and-roles)
4. [Identity and Authentication](#identity-and-authentication)
5. [Monitoring](#monitoring)
6. [Elastic Compute Nodes](#elastic-compute-nodes)
7. [Data Provisioning Agent](#data-provisioning-agent)
8. [System Maintenance](#system-maintenance)
---
## Tenant Configuration
### Creating a Tenant
**SAP BTP Service Instance**:
1. Access SAP BTP Cockpit
2. Navigate to Subaccount
3. Create SAP Datasphere service instance
4. Configure initial sizing
**Plan Options**:
| Plan | Description |
|------|-------------|
| Free | Trial with limitations |
| Standard | Production use |
### Configuring Tenant Size
**Capacity Parameters**:
- Storage (GB)
- In-memory (GB)
- Compute units
**Sizing Recommendations**:
| Use Case | Storage | Memory |
|----------|---------|--------|
| Small | 256 GB | 32 GB |
| Medium | 1 TB | 128 GB |
| Large | 4 TB+ | 512 GB+ |
### System Information
**Display System Info**:
- System > About
- View tenant ID
- Check version
- Monitor capacity usage
### SAP HANA Configuration
**Enable Script Server**:
1. System > Configuration
2. Enable SAP HANA Cloud Script Server
3. Required for Python, R, AFL
**Enable SQL Data Warehousing**:
1. System > Configuration
2. Enable SAP HANA SQL Data Warehousing
3. Allows HDI container deployment
### Additional Features
**Enable SAP Business AI**:
- AI-powered features
- Intelligent recommendations
- Natural language queries
**Enable Choropleth Layers**:
- Geographic visualizations
- Map-based analytics
### OAuth 2.0 Configuration
**Client Types**:
| Type | Purpose |
|------|---------|
| Technical User | System-to-system integration |
| API Access | REST API calls |
| Interactive Usage | User authentication |
**Creating OAuth Client**:
1. System > Security > OAuth 2.0 Clients
2. Create new client
3. Configure client type
4. Note client ID and secret
**API Access Configuration**:
```json
{
"clientid": "sb-xxx",
"clientsecret": "xxx",
"url": "[https://xxx.authentication.xxx.hana.ondemand.com",](https://xxx.authentication.xxx.hana.ondemand.com",)
"apiurl": "[https://xxx.hana.ondemand.com"](https://xxx.hana.ondemand.com")
}
```
### Trusted Identity Providers
Add external IdP for authentication:
1. System > Security > Identity Providers
2. Add trusted IdP
3. Configure SAML settings
4. Map user attributes
---
## Spaces and Storage
### Creating Spaces
**Standard Space**:
1. Space Management > Create
2. Enter space name
3. Configure storage
4. Assign users
**File Space**:
1. Space Management > Create File Space
2. Configure object store
3. Set data lake connection
### Space Properties
**Initial Creation Fields**:
| Property | Specifications |
|----------|----------------|
| Space Name | Maximum 30 characters; allows spaces and special characters |
| Space ID | Maximum 20 UPPERCASE letters/numbers; underscores only |
| Storage Type | SAP HANA Database (Disk and In-Memory) |
**General Settings (Read-Only)**:
- Space Status (newly-created spaces are active)
- Space Type (SAP Datasphere only)
- Created By/On timestamps
- Deployment Status and Deployed On
**Optional Configuration**:
| Setting | Description |
|---------|-------------|
| Data Access | Exposure for consumption defaults |
| Database User | Create for external tool connections |
| HDI Container | Associate HDI container |
| Time Data | Generate standardized time tables/dimensions |
| Auditing | Enable read/change action logging |
**Deployment**: Spaces require deployment after creation and re-deployment after modifications.
### Technical Naming Rules (Space ID)
**Valid Space IDs**:
- UPPERCASE letters, numbers, underscores only
- Maximum 20 characters
- No spaces or special characters
**Reserved Prefixes (Avoid)**:
- `_SYS` - System reserved
- `DWC_` - Datasphere reserved
- `SAP_` - SAP reserved
**Example**: `SALES_ANALYTICS_2024`
### Storage Allocation
**Allocate Storage**:
1. Open space settings
2. Set disk storage (GB)
3. Set in-memory storage (GB)
4. Save changes
**Storage Types**:
| Type | Use | Performance |
|------|-----|-------------|
| Disk | Persistent data | Standard |
| In-Memory | Hot data | High |
| Object Store | Large files | Cost-effective |
### Space Priorities
**Priority Levels**:
1. High: Critical workloads
2. Medium: Standard workloads
3. Low: Background tasks
**Statement Limits**:
- Maximum memory per query
- Query timeout
- Concurrent connections
### Space Operations
**Copy Space**:
1. Space Management
2. Select source space
3. Copy with/without data
4. New space name
**Delete Space**:
1. Remove all objects
2. Remove all users
3. Delete space
**Restore from Recycle Bin**:
1. System > Recycle Bin
2. Select deleted space
3. Restore or permanently delete
### Command Line Management
**datasphere CLI**:
```bash
# Login
datasphere login
# List spaces
datasphere spaces list
# Create space
datasphere spaces create --name my_space --storage 100
# Delete space
datasphere spaces delete --name my_space
```
---
## Users and Roles
### User Management
**Creating Users**:
1. Security > Users
2. Create user
3. Enter email
4. Assign roles
**User Properties**:
| Property | Description |
|----------|-------------|
| Email | Login identifier |
| First Name | Display name |
| Last Name | Display name |
| Manager | Reporting structure |
### Role Types
**Global Roles**:
- Apply across all spaces
- System-level permissions
**Scoped Roles**:
- Space-specific permissions
- Object-level access
### Standard Roles
| Role | Description |
|------|-------------|
| DW Administrator | Full system access |
| DW Space Administrator | Space management |
| DW Integrator | Data integration |
| DW Modeler | Data modeling |
| DW Viewer | Read-only access |
### Role Privileges
**System Privileges**:
- Lifecycle: Deploy, monitor, transport
- User Management: Create, assign users
- Security: Manage access controls
**Space Privileges**:
- Create Objects
- Read Objects
- Update Objects
- Delete Objects
- Share Objects
### Creating Custom Roles
1. Security > Roles > Create
2. Enter role name
3. Select privileges
4. Assign to users
### Scoped Roles
**Creating Scoped Role**:
1. Security > Roles > Create Scoped
2. Define base privileges
3. Assign spaces
4. Assign users
**Scope Options**:
- All spaces
- Selected spaces
- Space categories
### Role Assignment
**Direct Assignment**:
- Security > Users > Assign Roles
**SAML Attribute Mapping**:
- Map IdP attributes to roles
- Automatic role assignment
- Dynamic membership
### SCIM 2.0 API
**User Provisioning**:
```http
POST /api/v1/scim/Users
Content-Type: application/json
{
"userName": "user@example.com",
"name": {
"givenName": "John",
"familyName": "Doe"
},
"emails": [{"value": "user@example.com"}]
}
```
### View Authorizations
**By User**:
- All roles assigned
- All spaces accessible
- Effective permissions
**By Role**:
- All users with role
- Permission details
**By Space**:
- All users in space
- Role breakdown
---
## Identity and Authentication
### SAP Cloud Identity Services
**Bundled IdP**:
- Included with SAP Datasphere
- Basic user management
- SAML 2.0 support
**Configuration**:
1. Access Identity Authentication admin
2. Configure application
3. Set user attributes
4. Enable SSO
### Custom SAML Identity Provider
**Requirements**:
- SAML 2.0 compliant IdP
- Metadata exchange
- Attribute mapping
**Setup**:
1. Export Datasphere SAML metadata
2. Import to IdP
3. Export IdP metadata
4. Import to Datasphere
5. Configure attribute mapping
**SAML Attributes**:
| Attribute | Purpose |
|-----------|---------|
| email | User identification |
| firstName | Display name |
| lastName | Display name |
| groups | Role assignment |
### Certificate Management
**SAML Signing Certificates**:
- Update before expiration
- Coordinate with IdP
- Test after update
### Database User Password Policy
**Policy Settings**:
- Minimum length
- Complexity requirements
- Expiration period
- History depth
---
## Monitoring
### Capacity Monitoring
**Monitor**:
- Storage usage
- Memory consumption
- Compute utilization
**Alerts**:
- Configure thresholds
- Email notifications
- Automatic warnings
### Audit Logs
**Database Audit Logs**:
- DDL operations (CREATE, ALTER, DROP)
- DML operations (SELECT, INSERT, UPDATE, DELETE)
- Login/logout events
**Configuration**:
1. System > Audit
2. Enable audit logging
3. Select event types
4. Set retention period
**Delete Audit Logs**:
- Manual deletion
- Scheduled cleanup
- Retention-based removal
### Activity Logs
**Tracked Activities**:
- Object creation
- Object modification
- Object deletion
- Deployments
### Task Logs
**Task Types Logged**:
- Data flows
- Replication flows
- Transformation flows
- Task chains
**Task Log Properties**:
| Property | Description |
|----------|-------------|
| Start date/time | When task started |
| Object name/type | Object being processed |
| Space name | Space containing the object |
| Storage type | SAP HANA Database or Data Lake Files |
| Activity type | persist, replicate, execute |
| Status/substatus | Completion status with failure descriptions |
| SAP HANA Peak Memory (MiB) | Requires expensive statement tracing |
| SAP HANA Used Memory (MiB) | Memory consumption |
| SAP HANA Used CPU Time (ms) | Requires expensive statement tracing |
| SAP HANA Used Disk (MiB) | Disk consumption |
| Apache Spark Peak Memory | Peak memory for Spark tasks |
| Apache Spark Spill to Disk | Data spilled to disk |
| Apache Spark Used Cores | Number of cores used |
| Records count | Only for: views (persist), remote tables (replicate), data flows, intelligent lookups |
**Display Limitations**:
- Only first **1,000 rows** displayed for performance
- Filters applied to all rows, but only first 1,000 filtered rows shown
- Use filters to find specific data
**Decimal Separator Note**: Use '.' (period) as decimal separator regardless of regional settings when filtering on memory/CPU columns.
**CPU Time Measurement**: CPU time measures time used by all threads. If much higher than statement duration, indicates heavy thread usage which can lead to resource bottlenecks.
**Log Management**:
- View execution history
- Download logs
- Delete old logs
### Notifications
**Configure Notifications**:
1. User profile > Notifications
2. Select event types
3. Choose delivery method
**Notification Types**:
- Task completion
- Task failure
- System alerts
- Capacity warnings
### Database Analysis Users
**Create Analysis User**:
1. System > Monitoring
2. Create database analysis user
3. Grant analysis privileges
4. Connect with SQL tools
**Analysis Capabilities**:
- Query monitoring views
- Analyze execution plans
- Debug performance issues
**Stop Running Statements**:
```sql
-- Find running statements
SELECT * FROM M_ACTIVE_STATEMENTS;
-- Cancel statement
ALTER SYSTEM CANCEL SESSION 'connection_id';
```
### SAP HANA Monitoring Views
**System Views**:
| View | Purpose |
|------|---------|
| M_ACTIVE_STATEMENTS | Running queries |
| M_CONNECTIONS | Active connections |
| M_SERVICE_MEMORY | Memory usage |
| M_VOLUME_IO | I/O statistics |
### SAP Cloud ALM Integration
**Health Monitoring**:
- Integration for checking tenant health
- Real-time health status
**Job & Automation Monitoring**:
- Monitor tasks (except child tasks)
- Integration with SAP Cloud ALM dashboard
### SAP HANA Cockpit Integration
Access via "Open SAP HANA Cockpit" links in System Monitor:
- Performance Monitor for real-time CPU/memory utilization
- Database Overview page for HANA analysis
- Admission Control analysis
---
## Elastic Compute Nodes
### Overview
Elastic compute nodes provide additional processing capacity for intensive workloads.
### Creating Elastic Compute Node
1. System > Elastic Compute Nodes
2. Create new node
3. Configure capacity
4. Set warm-up schedule
### Node Configuration
| Parameter | Description |
|-----------|-------------|
| Node Name | Identifier |
| Capacity | Processing units |
| Warm-up Time | Pre-start minutes |
| Auto-shutdown | Idle timeout |
### Running Elastic Compute
**Start Node**:
1. Select node
2. Start manually or schedule
3. Wait for warm-up
4. Execute workloads
**Assign Workloads**:
- Data flows
- Transformation flows
- Specific queries
### Resource Purchase
**Capacity Units**:
- Billed by consumption
- Pre-purchase options
- Monitor usage
---
## Data Provisioning Agent
### Installation
**Requirements**:
- Java 11+
- Network access to sources
- Network access to Datasphere
**Installation Steps**:
1. Download agent from SAP
2. Install on-premise server
3. Configure connection
4. Register with Datasphere
### Configuration
**Agent Properties**:
```properties
# Connection settings
datasphere.tenant.url=[https://xxx.hana.ondemand.com](https://xxx.hana.ondemand.com)
datasphere.agent.name=dp_agent_01
# Performance settings
datasphere.threads.max=10
datasphere.batch.size=10000
```
### Adapter Registration
**Register Adapters**:
1. System > Data Provisioning
2. Select agent
3. Register adapter
4. Configure connection
**Supported Adapters**:
- ABAP ODP
- HANA SDI
- File adapters
- Database adapters
### Agent Monitoring
**Status Monitoring**:
- Connection status
- Replication status
- Error logs
**Log Access**:
1. Enable log access
2. View logs in Datasphere
3. Download for analysis
### Pause Replication
**Pause Agent**:
- Maintenance window
- Network issues
- Source system updates
**Resume Agent**:
- Verify connectivity
- Check queue status
- Resume replication
---
## System Maintenance
### HANA Database Operations
**Restart Database**:
1. System > HANA Cloud
2. Restart database
3. Wait for recovery
4. Verify connections
**Apply Patch Upgrades**:
1. Review available patches
2. Schedule maintenance window
3. Apply patch
4. Validate functionality
### Support Requests
**Request SAP Support**:
1. System > Support
2. Create incident
3. Provide details
4. Attach logs
**Required Information**:
- Tenant ID
- Error messages
- Steps to reproduce
- Screenshots/logs
---
## Documentation Links
- **Tenant Configuration**: [https://help.sap.com/docs/SAP_DATASPHERE/c8a54ee704e94e15926551293243fd1d/2f80b57](https://help.sap.com/docs/SAP_DATASPHERE/c8a54ee704e94e15926551293243fd1d/2f80b57)
- **Space Management**: [https://help.sap.com/docs/SAP_DATASPHERE/c8a54ee704e94e15926551293243fd1d/2ace657](https://help.sap.com/docs/SAP_DATASPHERE/c8a54ee704e94e15926551293243fd1d/2ace657)
- **User Management**: [https://help.sap.com/docs/SAP_DATASPHERE/c8a54ee704e94e15926551293243fd1d/4fb82cb](https://help.sap.com/docs/SAP_DATASPHERE/c8a54ee704e94e15926551293243fd1d/4fb82cb)
- **Monitoring**: [https://help.sap.com/docs/SAP_DATASPHERE/c8a54ee704e94e15926551293243fd1d/28910cd](https://help.sap.com/docs/SAP_DATASPHERE/c8a54ee704e94e15926551293243fd1d/28910cd)
---
**Last Updated**: 2025-11-22