Step-by-step how-to documentation for every MinusNow module — from CMDB discovery to AI-driven auto-healing and forensic analysis.
Configuration Management Database — automatic asset discovery, relationship mapping, and lifecycle tracking.
The CMDB module provides a single source of truth for all IT assets (Configuration Items). It automatically discovers servers, network devices, applications, and cloud resources, then maps dependencies for impact analysis.
Agent-based and agentless discovery via SNMP, WMI, SSH, and cloud APIs. Discovers hardware, OS, installed software, running services, open ports, and network topology.
Automatic dependency mapping between CIs. Visualize impact radius for any asset. Supports parent-child, runs-on, depends-on, and connected-to relationship types.
Every CI change is versioned and audited. Compare configurations over time. Detect configuration drift and unauthorized changes with diff view.
Native sync with AWS (EC2, RDS, Lambda), Azure (VMs, App Service, AKS), and GCP (Compute, Cloud SQL, GKE). Auto-tags and classifies cloud resources.
| Category | Data Points |
|---|---|
| Hardware | CPU model/cores, RAM, disk (type/size/usage), NIC, serial number, BIOS version |
| Operating System | OS name, version, kernel, architecture, uptime, last patch date |
| Software | Installed packages (name, version, source), auto-update status |
| Services | Running services, state, startup type, dependencies, listening ports |
| Network | IP addresses, MAC, gateway, DNS, FQDN, VLAN, open ports |
| Cloud | Instance ID, region, VPC, security groups, tags, cost allocation |
Real-time infrastructure monitoring with correlated alerting, dashboards, and anomaly detection.
| Category | Metrics | Default Interval |
|---|---|---|
| CPU | Total %, per-core %, load average, steal %, iowait | 30s |
| Memory | Used %, available, swap usage, cache, buffers | 30s |
| Disk | Usage %, IOPS, read/write throughput, latency, inodes | 60s |
| Network | TX/RX bytes, packets, errors, drops, connection count | 30s |
| Process | Count, top consumers, zombie count, file descriptors | 60s |
| Service | State (up/down), response time, port availability | 30s |
| Application | HTTP status, response time, error rate, custom metrics | 60s |
The AI engine correlates related alerts (e.g., a disk full alert → service crash → application error) into a single incident, reducing alert fatigue by up to 80%. Correlation rules use topology, timing, and historical pattern matching.
ITIL-aligned incident lifecycle from detection to resolution with AI-assisted triage and escalation.
| High Urgency | Medium Urgency | Low Urgency | |
|---|---|---|---|
| High Impact | P1 — Critical Response: 15 min Resolve: 4 hrs | P2 — High Response: 30 min Resolve: 8 hrs | P3 — Medium Response: 2 hrs Resolve: 24 hrs |
| Medium Impact | P2 — High Response: 30 min Resolve: 8 hrs | P3 — Medium Response: 2 hrs Resolve: 24 hrs | P4 — Low Response: 8 hrs Resolve: 72 hrs |
| Low Impact | P3 — Medium Response: 2 hrs Resolve: 24 hrs | P4 — Low Response: 8 hrs Resolve: 72 hrs | P5 — Planning Response: 24 hrs Resolve: 1 week |
Risk-controlled change lifecycle with approval workflows, CAB reviews, and automated rollback.
| Type | Risk | Approval | Lead Time | Example |
|---|---|---|---|---|
| Standard | Low | Pre-approved | None | Password reset, user onboarding, pre-tested patch |
| Normal | Medium | 1-2 approvers | 3-5 days | Server upgrade, network config change, app deployment |
| Major | High | CAB review | 7-14 days | Infrastructure migration, architecture change |
| Emergency | Varies | Fast-track | Immediate | Security patch for active exploit, hotfix for outage |
Root-cause investigation, known-error tracking, and proactive problem identification.
Self-service catalog for users to request IT services with automated fulfillment.
| Category | Items | Fulfillment |
|---|---|---|
| Access | New user account, password reset, group membership, VPN access | Automated via Directory Sync |
| Hardware | Laptop request, monitor, peripheral, mobile device | Manual (asset assignment) |
| Software | App install, license request, dev environment | Automated via Agent |
| Cloud | VM provisioning, storage, database instance, Kubernetes namespace | Automated via Cloud API |
| Network | Firewall rule, DNS entry, load balancer config | Semi-automated |
Automated incident remediation through runbooks, scripts, and AI-driven recovery actions.
mnow-agent user must have scoped sudo privileges. See User & Permissions for the exact sudoers configuration.Auto-healing on Linux requires the mnow-agent user to have sudo NOPASSWD access for specific commands: systemctl, apt/yum/dnf, reboot, rm (scoped), journalctl. On Windows, the agent service must run as Local Administrator or a domain account with admin rights. Without elevated privileges, auto-healing operates in dry-run / recommendation mode only.
| Runbook | Trigger | Action | Verify |
|---|---|---|---|
| Service Restart | Service down alert | systemctl restart <service> | Check port + health endpoint |
| Disk Cleanup | Disk usage > 90% | Purge old logs, tmp files, journal | Verify usage dropped below 85% |
| Process Kill | Runaway process (CPU > 95%) | Kill process by PID or name | Verify CPU normalized |
| Log Rotation | Log file > 1 GB | Rotate and compress log files | Verify log size reduced |
| Certificate Renewal | Cert expiry < 7 days | Run certbot renew, reload service | Verify cert validity |
| Memory Cleanup | Memory > 95% | Clear caches, restart memory-leaking service | Verify memory dropped |
Forecast resource utilization, plan capacity, and prevent saturation before it impacts services.
| Resource | Metrics Tracked | Forecast Window |
|---|---|---|
| CPU | Avg util %, peak util %, core count, thread count | 30/60/90-day trend |
| Memory | Avg used %, peak used %, total available | 30/60/90-day trend |
| Disk | Used space, growth rate (GB/day), IOPS headroom | Days until full |
| Network | Avg throughput, peak throughput, bandwidth available | 30/60/90-day trend |
Scan for vulnerabilities, prioritize risk, orchestrate patching, and track remediation.
apt, yum, dnf, Windows Update). The agent handles reboot scheduling and verification.In air-gapped or remote environments, a satellite server maintains a local mirror of the vulnerability database. Agents report their package inventory to the satellite, which performs local matching and syncs results to the central server when connectivity is available. See Satellite Requirements.
| Role | Permissions Needed | Actions Available |
|---|---|---|
| Security Admin | Full vulnerability module access | Configure scans, create campaigns, approve patches, manage exceptions |
| IT Operator | Read + Patch execution | View vulnerabilities, deploy approved patches, verify remediation |
| Auditor | Read-only | View reports, export compliance data, review scan history |
| Agent (OS-level) | sudo for package management | Install/update packages, schedule reboots |
How monitoring alerts become actionable incidents with automatic triage, deduplication, and suppression.
Build and deploy automated workflows that respond to events, schedules, or manual triggers.
| Template | Trigger | Actions |
|---|---|---|
| Auto-Assign by Category | Incident created | Check category → assign to correct team → notify via Slack |
| Disk Space Alert Response | Disk > 90% alert | Run cleanup script → verify → create incident if failed |
| New Employee Onboarding | Service request approved | Create AD account → provision email → assign laptop → notify HR |
| Change Approval Reminder | Change pending 48hrs | Send reminder to approver → escalate after 72hrs |
| SSL Certificate Expiry | Cert expiry < 30 days | Create change request → auto-renew → verify → close |
Synchronize users, groups, and roles from Active Directory, LDAP, Azure AD, Okta, and other identity providers.
User.Read.All, Group.Read.All, Directory.Read.All permissionsCreate, manage, and auto-surface knowledge articles linked to live operations.
Automated root-cause analysis, timeline reconstruction, and exportable forensic reports.
| Source | Data | Purpose |
|---|---|---|
| Monitoring | Alert history, metric trends, anomalies | When did symptoms first appear? |
| Change Records | Recent changes on affected CIs | Was the incident caused by a change? |
| CMDB | CI relationships, dependencies | What was the blast radius? |
| Deployment Logs | CI/CD events, release notes | Was new code deployed recently? |
| System Logs | syslog, event log, application logs | What errors occurred at the time? |
| Audit Logs | User actions, access events | Was there unauthorized or unexpected access? |
How user data flows between source directories, the MinusNow application, and client servers.
What syncs: User accounts, group memberships, organizational structure, manager hierarchy, account status (enabled/disabled), authentication credentials (via SSO redirect, not password sync).
Requirements:
What syncs: Agent configuration (which user has access to run commands on which hosts), role-based permissions (who can trigger auto-healing), monitoring exemptions, and maintenance window awareness.
Requirements:
mnow-agent| Data | Source | Destination | Direction | Frequency |
|---|---|---|---|---|
| User accounts | AD / Azure AD | MinusNow App DB | Source → App | Every 15 min |
| Group memberships | AD / Azure AD | MinusNow Role Engine | Source → App | Every 15 min |
| SSO tokens | IdP (SAML/OIDC) | User browser → App | Redirect flow | On login |
| Agent config | MinusNow App | Client agents | App → Client | Real-time push |
| RBAC policies | MinusNow App | Client agents | App → Client | On change + 5 min poll |
| Telemetry data | Client agents | MinusNow App DB | Client → App | Every 30-60s |
| Vulnerability data | Client agents | MinusNow App DB | Client → App | On package change + daily |
All sync communication uses TLS 1.3 encryption. Agent-to-server communication uses mutual TLS (mTLS) with auto-rotated certificates. No passwords are synced — authentication is always handled via the identity provider using SAML 2.0 or OIDC. API tokens are encrypted at rest with AES-256-GCM.