Jump to content

Juq470 Today

def sum_sales(acc, row): return acc + row["sale_amount"]

(pipeline() .source(read_csv("visits.csv")) .pipe(enrich) .filter(lambda r: r["country"] == "US") .sink(write_jsonl("us_visits.jsonl")) ).run() juq470 provides a catch operator to isolate faulty rows without stopping the whole pipeline: juq470

from juq470 import pipeline, read_csv

def safe_int(val): return int(val)

juq470 is a lightweight, open‑source utility library designed for high‑performance data transformation in Python. It focuses on providing a concise API for common operations such as filtering, mapping, aggregation, and streaming large datasets with minimal memory overhead. Key Features | Feature | Description | Practical Benefit | |---------|-------------|--------------------| | Zero‑copy streaming | Processes data in chunks using generators. | Handles files > 10 GB without exhausting RAM. | | Typed pipelines | Optional type hints for each stage. | Improves readability and catches errors early. | | Composable operators | Functions like filter , map , reduce can be chained. | Builds complex workflows with clear, linear code. | | Built‑in adapters | CSV, JSONL, Parquet readers/writers. | Reduces boilerplate when working with common formats. | | Parallel execution | Simple parallel() wrapper uses concurrent.futures . | Gains speedups on multi‑core machines with minimal code changes. | Installation pip install juq470 The package requires Python 3.9+ and has no external dependencies beyond the standard library. Basic Usage 1. Simple pipeline from juq470 import pipeline, read_csv, write_jsonl | Handles files > 10 GB without exhausting RAM

def capitalize_name(row): row["name"] = row["name"].title() return row | | Composable operators | Functions like filter

def enrich_with_geo(row): # Assume get_geo is a fast lookup function row["country"] = get_geo(row["ip"]) return row

×
×
  • Create New...

Important Information

By using this site, you agree to our Terms of Use. This site uses cookies to enhance the users' browsing experience and collect infomation about the site utilization. We use both technical cookies and third party cookies. If you continue browsing the site you accept the use of cookies; otherwise you can just leave the site.