GitHub user oliverbell-klv created a discussion: Failure on creation of 
Snowflake dataset due to staging S3 download errors (works in SQL Lab, 
regression with new connector)

### Bug description

### Description

We have a Snowflake connection that works fine in SQL Lab (queries succeed).
When creating a new dataset or running “Test Connection,” Superset fails with:

```
An Error Occurred
Unable to load columns for the selected table. Please select a different table.

```

Logs show underlying errors from the Snowflake connector when trying to fetch 
staged results:

```
HTTPSConnectionPool(host='<Snowflake staging S3 bucket in us-west-2>', 
port=443):
Max retries exceeded … Remote end closed connection without response
```

### Screenshots

<img width="1719" height="863" alt="Image" 
src="https://github.com/user-attachments/assets/34edbfe1-77ee-4c76-8c18-c082ec672498";
 />

### Repro steps

1. Go to Datasets
3. Click + Dataset
5. Pick Snowflake DB connection
7. Choose schema + table
9. Observe error

### Expected

Columns load, Test Connection succeeds.

### Actual

- Dataset creation fails.
- Test Connection fails.
- SQL Lab queries continue to work (likely because small results don’t hit S3 
staging).

### Environment

- Superset 4.1.1 (dockerized, AWS ECS Fargate)
- Python 3.9 (default in base image)
- Snowflake connector: 3.16.0
- Snowflake SQLAlchemy: 1.7.6
- Snowflake region: AWS us-west-2

### Troubleshooting performed

- Confirmed Snowflake network policy allows our NAT egress IPs.
- Verified no failed logins in Snowflake login history (issue occurs before 
auth).
- Increased Superset/Gunicorn/ALB timeouts.
- Disabled proxy variables, set NO_PROXY for Snowflake/AWS domains.
- Tried connector options: ocsp_fail_open, insecure_mode, session params 
(CLIENT_PREFETCH_THREADS, CLIENT_RESULT_CHUNK_SIZE, USE_S3_REGIONAL_URL, etc.).
- Added curl sidecar: Snowflake account host + generic S3 reachable, but 
staging S3 bucket sometimes fails.
- Tested with minimal 40-row tables - still fails
- Confirmed issue is specific to metadata/reflection queries, not query size

Why this might be Superset-related

- SQL Lab queries work, but inspector/metadata queries consistently fail.
- Looks like a difference in how Superset uses the connector for 
reflection/metadata (larger results, staging) vs. SQL Lab.
- Want to confirm if this is a known issue with staging downloads in Superset’s 
Snowflake integration, and if there are recommended config flags / retries / 
version pins.

### Screenshots/recordings

_No response_

### Superset version

master / latest-dev

### Python version

3.9

### Node version

16

### Browser

Chrome

### Additional context

_No response_

### Checklist

- [x] I have searched Superset docs and Slack and didn't find a solution to my 
problem.
- [x] I have searched the GitHub issue tracker and didn't find a similar bug 
report.
- [x] I have checked Superset's logs for errors and if I found a relevant 
Python stacktrace, I included it here as text in the "additional context" 
section.

GitHub link: https://github.com/apache/superset/discussions/34778

----
This is an automatically sent email for [email protected].
To unsubscribe, please send an email to: 
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to