cancel
Showing results for 
Search instead for 
Did you mean: 
Warehousing & Analytics
Engage in discussions on data warehousing, analytics, and BI solutions within the Databricks Community. Share insights, tips, and best practices for leveraging data for informed decision-making.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

MadelynM
by Databricks Employee
  • 1878 Views
  • 0 replies
  • 0 kudos

[Recap] Data + AI Summit 2024 - Warehousing & Analytics | Improve performance and increase insights

Here's your Data + AI Summit 2024 - Warehousing & Analytics recap as you use intelligent data warehousing to improve performance and increase your organization’s productivity with analytics, dashboards and insights.  Keynote: Data Warehouse presente...

Screenshot 2024-07-03 at 10.15.26 AM.png
Warehousing & Analytics
AI BI Dashboards
AI BI Genie
Databricks SQL
  • 1878 Views
  • 0 replies
  • 0 kudos
Kaz
by New Contributor II
  • 7491 Views
  • 3 replies
  • 1 kudos

Show full logs on job log

Is it possible to show the full logs of a databricks job? Currently, the logs are skipped with:*** WARNING: max output size exceeded, skipping output. ***However, I don't believe our log files are more than 20 MB. I know you can press the logs button...

  • 7491 Views
  • 3 replies
  • 1 kudos
Latest Reply
Isi
Contributor III
  • 1 kudos

Hey @Kaz ,Unfortunately, the output truncation limit in the Databricks job UI cannot be changed. Once that limit is exceeded, the rest of the logs are skipped, and the full logs become accessible only through the “Logs” button, which, as you mentione...

  • 1 kudos
2 More Replies
teixeire
by New Contributor II
  • 285 Views
  • 3 replies
  • 0 kudos

How to restrict external access to SQL Warehouse but allow workspace queries?

Hi everyone,I'm currently setting up access controls in our Databricks development workspace. The goal is to enable business users to explore data and build their SQL skills within the workspace itself (e.g., via SQL editor or notebooks), but prevent...

Warehousing & Analytics
Access Rights
Endpoint
sql
  • 285 Views
  • 3 replies
  • 0 kudos
Latest Reply
Isi
Contributor III
  • 0 kudos

Hi @teixeire ,To prevent external tools like Power BI or DBeaver from connecting to your SQL Warehouse, one effective approach is to restrict personal access token (PAT) creation for users who should only query data inside the Databricks workspace.Th...

  • 0 kudos
2 More Replies
barchiel33
by New Contributor
  • 118 Views
  • 1 replies
  • 0 kudos

Databricks SQL Python - Result fetching takes extremely long time

Hello All!I have a python script which utilizes the databricks SQL for python package in order to pull a databricks table into a pandas dataframe which is used to create a table in a Spotfire report. The table contains ~1.28 million rows, with 155 co...

  • 118 Views
  • 1 replies
  • 0 kudos
Latest Reply
Isi
Contributor III
  • 0 kudos

Hey @barchiel33 ,Hi! Let me share my thoughts, I’m not entirely sure if all of this applies directly to your use case, but hopefully some of it helps, and we can iterate further if needed.From what I understand, you’re using a Python data function in...

  • 0 kudos
meghana_tulla
by New Contributor III
  • 324 Views
  • 5 replies
  • 0 kudos

UCX V0.58.0 Dashboard in Databricks Showing 0 Tables Despite Existing Tables.

Hi,I am currently installing UCX version 0.58.0 in Non-Unity enabled Databricks Workspace. It installed successfully and the job completes without errors. However, The Dashboard that goanna generate showing "0 Total Tables" and other metrics as 0.Has...

  • 324 Views
  • 5 replies
  • 0 kudos
Latest Reply
meghana_tulla
New Contributor III
  • 0 kudos

Hi,Still facing the same issue actually, I am trying to automate the UCX installation using bashscript. while running the bashscript the job starts and run successfully with all run jobs. Even all jobs Ran successfully, I am getting 0 tables and 0 va...

  • 0 kudos
4 More Replies
lucy-ji
by New Contributor
  • 93 Views
  • 4 replies
  • 0 kudos

How to download widget result into CSV

In the databricks Tutorial (https://docs.databricks.com/gcp/en/sql/get-started/sample-dashboards) , it says we can right click the widget and choose "Download as CSV".  However, in my databricks dashboard, when I right click the widgets, there is no ...

  • 93 Views
  • 4 replies
  • 0 kudos
Latest Reply
Isi
Contributor III
  • 0 kudos

Hi @lucy-ji,Could you share a bit more detail about your setup? Specifically:Are you creating a new dashboard from the SQL > Dashboard interface?When you right-click and choose “View dataset”, are you taken to the Data tab (on the right side of the “...

  • 0 kudos
3 More Replies
alex456897874
by New Contributor II
  • 1073 Views
  • 6 replies
  • 0 kudos

how to reuse filter between pages in dashboards?

Hi!I have a multi-page dashboard based on a handful of tables sharing a common run_id field.I created a dropdown filter widget on the first page to filter datasets based on the table's run_id field.It worked fine for this page, but how to reuse it fo...

  • 1073 Views
  • 6 replies
  • 0 kudos
Latest Reply
Benjamin_DavidS
New Contributor II
  • 0 kudos

I have the same problem here. filters are not applied while switching pages

  • 0 kudos
5 More Replies
ogs
by New Contributor II
  • 131 Views
  • 2 replies
  • 2 kudos

Data selection from adls2 in Serverless Warehouse

Hi everyone,I'm trying to query data from our adls2 delta lake using a serverless sql warehouse. We've already set up private connectivity via NCC, but hitting a snag when running queries like:SELECT * FROM delta.`abfss://[email protected]...

  • 131 Views
  • 2 replies
  • 2 kudos
Latest Reply
ogs
New Contributor II
  • 2 kudos

Hi, thanks for the detailed explanation.Unfortunately, configuring fs.azure.account.key in the Serverless advanced options didn’t help (I’m sure I wrote it correctly) - still receiving the same error.I saw in some sources around the net that I should...

  • 2 kudos
1 More Replies
Sweta
by New Contributor II
  • 209 Views
  • 1 replies
  • 0 kudos

Optimized option to write updates to Aurora PostgresDB from Databricks/spark

Hello All,    We want to update our postgres tables from our spark structured streaming workflow on Databricks. We are using foreachbatch utility to write to this sink. I want to understand an optimized way to do this at near real time latency avoidi...

  • 209 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

  If near real-time latency is critical, focus on optimizing parallel writes with batch updates (Option 1).If you prioritize transactional stability and can tolerate slightly higher latency due to staging, continue refining your temporary table with ...

  • 0 kudos
AP01
by New Contributor
  • 325 Views
  • 1 replies
  • 0 kudos

Databricks JDBC Error: Job Aborted Due to Stage Failure (Executor OOM - Error Code 52)

java.sql.SQLException: [Databricks][JDBCDriver](500051) ERROR processing query/statement. Error Code: 0, SQL state: null, Query: SELECT `ma***, Error message from Server: org.apache.hive.service.cli.HiveSQLException: Error running query: org.apache.s...

Warehousing & Analytics
Databricks JDBC SparkSQL OOM HiveThriftServer Error500051
Databricks SQL
JDBC Driver
SparkSQL
sql
  • 325 Views
  • 1 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

The executor does not seem to have enough memory to process the assigned tasks, OOM error.

  • 0 kudos
jakobhaggstrom
by New Contributor
  • 489 Views
  • 1 replies
  • 0 kudos

Issue when using M2M authentication with azure databricks jdbc driver 2.7.1

Hi!I try to connect to an azure databricks sql warehouse in dbeaver, which uses the azure databricks jdbc driver version 2.7.1 ## and I cannot get M2M authentication to work. I get a 'Not Authorized' (401) response when I try to connect, and it seems...

  • 489 Views
  • 1 replies
  • 0 kudos
Latest Reply
Renu_
Contributor
  • 0 kudos

Hi @jakobhaggstrom, this error likely occurs due to the type of secret you're using. For M2M authentication, the Databricks JDBC driver requires a Databricks generated OAuth secret, not a Microsoft Entra ID client secret. While your service principal...

  • 0 kudos
Kyle2
by New Contributor II
  • 1504 Views
  • 3 replies
  • 1 kudos

Databricks JDBC driver fails with socket read timeout

We work with a application that connects to our Databricks serverless SQL warehouse via Databricks JDBC driver. It runs a few thousand SQL select statements everyday, and a small percentage of them will fail with the following error details: java.sql...

  • 1504 Views
  • 3 replies
  • 1 kudos
Latest Reply
Rafael-Sousa
Contributor II
  • 1 kudos

Did you tried to increase the socket timeout?

  • 1 kudos
2 More Replies
pavel_cerny
by New Contributor II
  • 2020 Views
  • 1 replies
  • 1 kudos

Databricks JDBC driver fails with socket read timeout

Application connects to Databricks serverless SQL warehouse via Databricks JDBC driver. It executes SQL select statements only. We see small number of statements failed each day with the following error detail: java.sql.SQLException: [Databricks][JDB...

  • 2020 Views
  • 1 replies
  • 1 kudos
Latest Reply
banghartt
New Contributor II
  • 1 kudos

Did you ever find a solution to this? Hitting the same

  • 1 kudos
meghana_tulla
by New Contributor III
  • 211 Views
  • 1 replies
  • 0 kudos

Issue with UCX Dashboard Display on Databricks v0.5.7 Installation

Hi,I am trying to install UCX for Databricks workspace using the Databricks CLI. When I attempt to install a specific older version instead of the latest, I get a dashboard creation error. This seems to be due to the April 7 update, which removed the...

  • 211 Views
  • 1 replies
  • 0 kudos
Latest Reply
Advika
Databricks Employee
  • 0 kudos

Hello @meghana_tulla! Older versions of UCX may fail to install because they try to create legacy dashboards, which are no longer supported following the April 7 update. To work around this, make sure your Databricks environment is aligned with the l...

  • 0 kudos
Akshay_Petkar
by Contributor III
  • 1802 Views
  • 6 replies
  • 2 kudos

How to Display Top Categories in Databricks AI/BI Dashboard?

In a Databricks AI/BI dashboard, I have a field with multiple categories (e.g., district-wise sales with 50 districts). How can I display only the top few categories (like the top 10) based on a specific metric such as sales?

  • 1802 Views
  • 6 replies
  • 2 kudos
Latest Reply
erbg
New Contributor II
  • 2 kudos

Did you find a solution? 

  • 2 kudos
5 More Replies
ahid
by New Contributor
  • 358 Views
  • 2 replies
  • 0 kudos

Databricks Connector for Looker Studio – No Aggregation Pushdown + 1M Row Limit

Hi Databricks Community,I'm trying to understand which team is responsible for maintaining the Databricks Connector for Looker Studio . We’re currently facing a major performance bottleneck with how this connector operates.Specifically:The connector ...

  • 358 Views
  • 2 replies
  • 0 kudos
Latest Reply
AndreyMirskiy
Databricks Employee
  • 0 kudos

Thank you for the feedback! Unfortunately there is a limitation in Looker Studio Community Connector API. getData method does not specify aggregation expectations for the data source. Therefore, a connector is expected to retrieve non-aggregated resu...

  • 0 kudos
1 More Replies
OSZAR »