Install an on-premises data gateway to securely connect on-site data with cloud services and keep workloads responsive. For reliable reporting, choose a gateway that supports networking between your local resources and the cloud, and hire an experienced installer. This choice reduces latency, monitoring overhead, and extraneous data movement while you keep the resources needed for peak usage in check, and there is clear cost awareness from the start.
The gateway is a software-based bridge that sits in your architecture of local servers. It runs on a dedicated server or virtual machine and provides a secure channel to cloud services. By design, it keeps data inside your network and transmits only the necessary results to the cloud.
How it works: The gateway runs as a service on a Windows or Linux host and communicates with cloud endpoints using TLS. It authenticates applications and users, transmits only authorized data, and returns results to dashboards. It monitors connectivity and throughput, helps allocate resources for fluctuating workloads, and is based on a vendor-provided image. It requires a supported OS and proper network rules, and it can be deployed with a trusted installer to simplify maintenance.
Benefits: It lowers cost by reducing large data transfers and enabling local pre-processing. It enables reliable networking with cloud dashboards and provides a stable bridge for multiple teams. There is an onyx hardware or virtual appliance option that fits your security posture. Builders and admins can search for connectors and implement basic governance policies. There is clear usage guidance to keep traffic within limits while you scale.
On-Premises Data Gateway: Quick Reference
Register the gateway under your main account to establish a managed, secure connection for organizational data and to enable centralized monitoring of changes. This setup serves users and protects credentials.
Install the gateway on a dedicated server within your data center or a trusted site, then configure network rules so it can reach cloud service endpoints. Keep a clear maintenance window and monitor health metrics to keep operations steady.
Create a cluster for load sharing and fault tolerance, and assign it to the following groups of users. A cluster keeps the gateway resilient and ensures seamless access when one node restarts.
The connection model uses the gateway as the bridge between on-prem data sources and cloud services. It executes queries and returns results with power and minimal data movement, preserving data stored on-site and reducing exposure. The query pathway stays efficient even as data volumes grow.
Follow best practices to achieve optimal throughput: limit concurrent queries per gateway, cache frequently used results, and design queries to minimize traffic. Keep bricks of your architecture organized so operations stay predictable.
Register credentials with a minimal-privilege account, enforce strong authentication, and maintain a controlled access list. This reduces risk as changes occur in the organizational data landscape.
Data management: store sensitive data behind gateway filters and isolate stored data paths. The gateway should provide a single connection point for queries among multiple sources, simplifying governance and audits.
Following practical tips ensures reliability: monitor the main metrics, keep the gateway software updated, and plan for scale by adding additional nodes to the cluster as load grows. This approach maintains seamless performance across users and data sources.
What is an On-Premises Data Gateway? Definition, How It Works, Benefits; Gateway Execution Flow
Deploy a dedicated gateway node on a trusted server to achieve optimal performance; start with a trial in a controlled environment to validate connections and cost implications.
An On-Premises Data Gateway is software that runs on your local servers, creating a secure bridge between on-prem источник данных and cloud services such as Microsoft 365 and the Power Platform. The gateway is registered in your account and managed from the cloud, with clear status shown in the icon in the admin console. For setup details, refer to the microsoft docs.
How it works: installed on a node in your network, the gateway maintains a secure outbound connection to the cloud and listens for data requests from services you use. It stores credentials locally in a protected form and uses them to access the configured sources, then passes results back through the gateway to the cloud service.
Gateway execution flow: the cloud service sends a request through the gateway; the gateway authenticates against your accounts; the node queries the on-prem sources; the data moves back through the gateway to the cloud; the service applies the result and refreshes on schedule. This flow is designed to be resilient, with automatic retries and transparent logging, so changes in one source do not disrupt others.
Benefits include keeping data on your network when possible, reducing unnecessary data movement and bandwidth cost, and enabling centralized management of connections and credentials. It supports multiple sources, guest accounts for collaboration, and a straightforward path to scale as your data footprint grows. Using the gateway saves time by consolidating management under one hosted service, and you can move workloads between on-prem and cloud without relocating data first.
Best practices: choose a registered gateway connected to your primary accounts, run it on a dedicated server, keep the host OS and gateway software current, and document every change in the docs for future audits. Ensure power and network redundancy, plan for failover, and test updates in a trial environment before moving production work. Monitor performance through the admin console and track related changes to ensure ongoing reliability.
Next steps: register the gateway, link your on-prem servers, and validate all connections in a test environment before moving production work.
Definition: What qualifies as an On-Premises Data Gateway
A gateway qualifies when it is designed to run inside the organizational network, securely bridge on-premises data sources to cloud services, and be managed for reliable data flow. It operates as a Windows service, uses dedicated accounts for access, and supports connectors to several sources through a single instance, enabling centralized control over where data travels and how it’s accessed.
Key criteria include location, redundancy, and compatibility. Install on a 64‑bit Windows server or desktop that remains online, and choose between Standard mode for multi-user, enterprise workloads or Personal mode for a single user. During install, a setup window appears; click through the prompts to select the mode and provide a recovery key. There is a dedicated setup window and you can verify status with the icon as you analyze connectivity. It can connect to related data sources such as SQL Server, Oracle, SAP, SharePoint, and other databases through secure tunnels, and publish data into cloud apps and services like Power BI, Power Apps, and Logic Apps through the same bridge, improving compatibility between on-prem and cloud ecosystems. It supports TLS encryption and uses service accounts aligned with organizational accounts, which helps governance and control. A license is required for cloud services; the gateway itself is free, but accessing protected data through cloud services uses license terms. If you cant connect due to network blocks, check firewall rules and ensure ports and proxies allow traffic to cloud endpoints. For recovering configuration, use the recovery key to recover and re-associate the gateway with the cloud tenant.
Organizations should analyze data sources and choose a location near critical sources to achieve reduced latency and exposure. They should selecting related data sources that the gateway supports through the standard connectors, and map accounts with minimal privilege to limit access. By selecting the appropriate gateway and keeping it updated, administrators maintain flow between on-prem sources and cloud consumers while preserving organizational control. Selecting the right gateway profile helps balance security and usability. You can adjust security settings to align with policy requirements.
Supported data sources and connectors
Use the Standard on-premises data gateway in gateway cluster mode to connect core on-prem sources: SQL Server, SQL Server Analysis Services (SSAS), Oracle, IBM DB2, SAP HANA, SAP BW, MySQL, PostgreSQL, Access, Excel, and ODBC data sources. Data remains stored on-prem and is refreshed to cloud services on demand and on schedule. Validate network access from the gateway machine to each database and ensure the gateway service account has the necessary permissions. A subscription refresh model is available in Power BI and other services, enabling predictable usage window. While this covers common needs, plan for specialized ERP or CRM sources if needed with direct connections.
Data sources are defined on the gateway with a data source name and a prefix that helps organize logs and policies. Each connector runs on the side of the gateway – on-prem side communicates with the cloud service, and the cloud side receives the results. The gateway can populate datasets in reports and apps and supports both scheduled refresh and direct query, depending on source capabilities. For larger teams, deploy a cluster to increase throughput and redundancy, enabling scalable solutions across departments.
Performance and scalability: For larger datasets, deploy a gateway cluster to increase throughput and reduce bottlenecks. You can scale capacity, manage concurrency, and set a limit on the number of refresh jobs; this can produce reduced load on backend systems while maintaining high performance. The system uses caching to speed queries and improve connectivity. If you might face maintenance windows for updates, schedule them during low usage.
Supported sources include relational databases (SQL Server, Oracle, MySQL, PostgreSQL, IBM DB2), SSAS, file-based sources (Excel, Access), and ODBC-based data sources. Add SharePoint on-prem lists and Dynamics on-prem where available. For each, ensure the data source is accessible from the gateway host, and that the appropriate drivers are installed. When you need to adjust a connection, you can paste new credentials into the gateway config; the gateway saves them securely and uses them for all refresh tasks. Construction of dashboards is faster when data sources are mapped clearly.
Observability and governance: Sankey diagrams show data movement across sources and targets, highlighting how queries flow from on-prem side to cloud services. A simple chartexpo visualization helps monitor data lineage, usage, and performance across the gateway cluster. If connectivity cant be established, check firewall rules, proxy settings, and TLS configuration. This approach reduces risk and maintains high performance while keeping data secure.
Deployment options and sizing basics
Start with a two-node Standard mode gateway cluster on a supported Windows server to maximize availability and minimize downtime. Place it in a secured network segment with redundant power and a reliable connection. Use clustering so traffic can fail over between nodes, and keep the gateway service separate from heavy data-processing tasks on the same computer to avoid contention. This setup suits production workloads and supports regular dataset refreshes from multiple sources. Each data source connects securely, and the design helps you keep control on the side of the network while reducing risk during peak hours.
Deployment options include Standard mode with clustering (recommended for production) and Personal mode (single-user). For ongoing operations, cant rely on a single gateway; configure at least two gateways in a side-by-side cluster for failover. You can run gateways on physical hardware or as virtual machines; the memory, CPU, and disk needs scale with load. Managed deployments offer centralized updates and policy; you can automate provisioning with powershell, and read gatewayresourceid values for scripts to drive decision-making and maintain a valid, auditable configuration.
Next, sizing basics by workload: Light: 2 vCPU, 4-6 GB memory, 60-100 GB disk; network 100 Mbps. Medium: 4 vCPU, 8-16 GB memory, 120-200 GB disk; network 1 Gbps. Heavy: 8 vCPU, 32 GB memory, 240-500 GB disk; network 1-2 Gbps. Plan for headroom to avoid contention, and allocate memory into the gateway process with room to grow into peaks. The next step is to monitor metrics and adjust allocations. Ensure each data source connects securely and keep the layout aligned with your source data volume and refresh cadence.
Automation and lifecycle: Use powershell to automate installation, configuration, scaling, and health checks across the cluster. Use gatewayresourceid values to target specific gateways in API calls, scripts, or monitoring dashboards. Opt for managed updates where possible to simplify patching and reduces operational overhead. For sizing decisions, maintain a clear decision-making framework that maps concurrency, data volumes, and refresh frequency to node count and memory needs, and validate changes with a controlled failover test.
Operational considerations: Model costs by gateway count, hardware or VM requirements, and licensing for Standard mode versus Personal mode. Regularly review network capacity and storage growth, and keep security postures aligned with secured access to on-prem data sources. Track metrics such as refresh duration, error rate, and idle time to ensure you aren’t overprovisioning. Maintain a valid license state and document gatewayresourceid mappings for audits and ongoing support. This approach keeps deployments scalable, predictable, and ready for next workload spikes into production without surprises.
Security and access controls in gateway operation
Enable MFA for all gateway accounts and apply strict RBAC. Tie permissions to clearly defined roles and keep registered users and service identities accounted. Use conditional access to require compliant devices and trusted locations. This approach reduces risk if a credential is compromised.
- Identity and access design
- Define three roles: viewer, operator, administrator; assign actions to each role and specify what is allowed. Enforce separation of duties and limit privilege to what is necessary.
- Assign roles to groups in your identity provider; when someone changes roles or leaves, revoke access promptly and verify that every active token comes from a registered account. check that the access level aligns with the job function.
- Session management and credentials
- Use connect-azaccount for sign-in with MFA; for automation, prefer service principals and rotate secrets. doesnt embed credentials in scripts; store them in a vault and paste credentials only into secure pipelines.
- Limit session lifetimes, require re-auth for sensitive actions, and set a window for token refresh. Ensure the on-prem location aligns with policy and doesnt expose keys.
- Monitoring, auditing and incident response
- Enable detailed gateway logs and a scheduled check window to review activity; related alerts, device posture and IP anomalies should trigger alerts. check for issues across accounts and devices, and link events to related alerts.
- Maintain a chartexpo data field mapping access events to risk scores and keep the latest policies up-to-date. If issues arise, save incident details and continue with remediation steps.
- Centralize policy files and load the latest configurations from the repo before applying changes. If you need to verify, paste the policy snippet into a secure editor to review.
- Keep other controls in place, such as IP allowlists and device posture checks, to further reduce risk. This helps make sure youre environment stays secure.
- Once a change is approved, push updates to all gateways and document the outcome in the audit log.
Gateway execution flow: data request to cloud service
Enable batching and local validation on the gateway to reduce unnecessary cloud requests and improve response times. This approach grows efficiency as demand grows, under memory-resident caches and secure transport ensuring up-to-date data across networks.
When a client application issues a request, the gateway begins registering the local session and checks that the resource is supported. It attaches metadata, validates the request against security policies, and ensures the gateway has an up-to-date view of access rights. It helps navigate policy layers to ensure authorization.
The gateway then packages the payload, applies compression if needed, and sends it through a secure channel to the cloud service. It uses the on-prem storage and memory to buffer data, supporting retry logic so that a temporary network issue doesnt lose the request. If the cloud is unavailable, the gateway stores the request locally for later processing while avoiding duplicate submissions. This avoids turning the process into a game of retries and improves resilience.
On the cloud, the service validates the incoming request, confirms it is valid and within policy, and routes it to the appropriate resource. The response travels back through the same secure channel, and the gateway updates its local state to reflect the latest status. This overall flow supports fast recovering and retry, reducing cost by avoiding duplicate processing and enabling idempotent handling. Shown results from pilots confirm reliability across existing deployments.
| Pașii | Acțiune | Key considerations | Output |
|---|---|---|---|
| 1. Client request | Gateway registering the local session, authenticates, and validates the request against policy | Resource supported; permissions valid; policy up-to-date | Request accepted and queued for transport |
| 2. Packaging & transport | Payload packaged, compressed if needed, sent through TLS to cloud | Memory/storage buffers; retry/backoff; secure networks | Cloud-ready request delivered |
| 3. Cloud processing | Cloud validates and routes to the correct resource | Policy check passes; valid data; idempotent handling | Processing result generated |
| 4. Response & sync | Response travels back; gateway updates local state and cache | Up-to-date status; existing sessions preserved; memory keeps recent state | Application receives result; system ready for next request |
What is an On-Premises Data Gateway? Definition, How It Works, and Benefits">
