Skip to content

Commit

Permalink
Merge pull request #12991 from MicrosoftDocs/repo_sync_working_branch
Browse files Browse the repository at this point in the history
Confirm merge from repo_sync_working_branch to main to sync with https://github.com/MicrosoftDocs/powerbi-docs (branch main)
  • Loading branch information
Albertyang0 authored Feb 17, 2025
2 parents 40e4164 + 9ae5b56 commit 346a5dd
Showing 1 changed file with 1 addition and 1 deletion.
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,7 @@ Keep the following considerations and limitations in mind when working with Azur
* Semantic models created on the web by uploading a CSV file don't generate logs.
* If you have Multi-Factor Auth (MFA) in place for Azure but not Power BI, the configuration screens will give general Azure errors. A workaround is to first sign in to the [Azure portal](https://portal.azure.com), complete the MFA challenge and then log into Power BI in the same browser session.
* If you're using private links/VNets to isolate your Log Analytics workspaces, data ingestion into Log Analytics is unaffected. However, the [Log Analytics Template app(https://appsource.microsoft.com/product/power-bi/pbi_pcmm.powerbiloganalyticsforasengine?tab=Overview)] won't work because it relies on a public endpoint that is no longer accessible by the Power Service as a private link. A workaround is to use the [.pbit report template(https://github.com/microsoft/PowerBI-LogAnalytics-Template-Reports)] and refresh the data from inside the private VNet. You must set up a custom DNS mapping to ensure the public endpoint uses a private internal IP.
* For the Log Analytics feature, Power BI only sends data to the *PowerBIDatasetsWorkspace* table and doesn't send data to the to *PowerBIDatasetsTenant* table. This avoids storing duplicate data about log analytics in both locations.
* For the Log Analytics feature, Power BI only sends data to the *PowerBIDatasetsWorkspace* table and doesn't send data to the *PowerBIDatasetsTenant* table. This avoids storing duplicate data about log analytics in both locations.


## Related content
Expand Down

0 comments on commit 346a5dd

Please sign in to comment.