ayshaysha / aws-csv-to-parquet-converterLinks
This Script gets CSV file from Amazon S3 using Python Library Boto3 and converts it to Parquet Format before uploading the new Parquet Version again to S3.
☆9Updated last month
Alternatives and similar repositories for aws-csv-to-parquet-converter
Users that are interested in aws-csv-to-parquet-converter are comparing it to the libraries listed below
Sorting:
- DuckDB with Dashboarding tools demo evidence, streamlit and rill☆16Updated last year
- ☆15Updated 4 years ago
- ☆10Updated 10 months ago
- The code to follow along our tutorials for the dlt rest_api source☆10Updated last year
- AWS Glue Configurable Test Data Generator for S3 Data Lakes and DynamoDB☆17Updated last year
- Serverless Datalake architecture☆13Updated 2 years ago
- An experimental Athena extension for DuckDB 🐤☆54Updated 6 months ago
- A python package to create a database on the platform using our moj data warehousing framework☆22Updated this week
- Use cases examples using Versions☆11Updated 2 months ago
- This repository contains example patterns for storing large objects with DynamoDB.☆12Updated last year
- 股市情報系統☆11Updated 10 months ago
- a Retrieval-Augmented Generation (RAG) multi-document chatbot application using Llama 3, Langchain, Streamlit, and Groq API☆10Updated 10 months ago
- Building Product Descriptions with AWS Bedrock and Rekognition☆10Updated 8 months ago
- A serverless duckDB deployment at GCP☆39Updated 2 years ago
- ☆16Updated 2 years ago
- An infrastructure as code approach to deploying Snowflake using Terraform☆25Updated 2 years ago
- Fully unit tested utility functions for data engineering. Python 3 only.☆17Updated 10 months ago
- ☆52Updated this week
- Retrieval Augmented Generation, but no servers involved. Backed by S3☆10Updated last year
- Boilerplates for running DLT on AWS Lambda to create well-structured datasets from unstructured JSON without breaking a sweat☆18Updated 3 months ago
- Run dbt serverless in the Cloud (AWS)☆42Updated 5 years ago
- demo examples how to load data from different sources to different destinations☆21Updated 5 months ago
- Code accompanying AWS blog post "Build a Semantic Search Engine for Tabular Columns with Transformers and Amazon OpenSearch Service"☆17Updated last year
- Using DuckDB with AWS Lambda to process Delta Lake data☆28Updated 5 months ago
- dbt / Amazon Redshift Demonstration Project☆34Updated 2 years ago
- This repo contains examples of high throughput ingestion using Apache Spark and Apache Iceberg. These examples cover IoT and CDC scenario…☆25Updated last week
- API for distributing Data Lake Data☆11Updated 3 months ago
- https://duyet.github.io/related-skills-visualization/index.html☆11Updated 5 years ago
- ☆13Updated 2 years ago
- Building 3D Trusted Data Pipelines With Dagster, Dbt, and Duckdb☆21Updated last year