Coordinates data pipeline tasks (ETL, analytics, feature engineering). Use when implementing data ingestion, transformations, quality checks, or analytics.
Score 70/100
Coordinates data pipeline tasks (ETL, analytics, feature engineering). Use when implementing data ingestion, transformations, quality checks, or analytics.
Score 70/100
Data Partitioner - Auto-activating skill for Data Pipelines. Triggers on: data partitioner, data partitioner Part of the Data Pipelines skill category.
Score 70/100
Implements data persistence systems including DataStore patterns, session locking, data migration, error handling, and backup systems.
Score 70/100
Data pipeline and ETL automation - extract, transform, load workflows for data integration and analytics
Score 70/100
Designs ETL/ELT data pipelines with proper extraction, transformation, and loading patterns, including orchestration, error handling, and data quality validation.
Score 70/100
Expert data engineer for ETL/ELT pipelines, streaming, data warehousing. Activate on: data pipeline, ETL, ELT, data warehouse, Spark, Kafka, Airflow, dbt, data modeling, star…
Score 70/100
Develop and manage data ingestion, processing, and transformation pipelines for pilot projects. Use when automating ETL workflows, integrating new data sources, or building…
Score 70/100
Follow these patterns when implementing data pipelines, ETL, data ingestion, or data validation in OptAIC.
Score 70/100
Process data files through transformation pipelines with validation, cleaning, and export. Use for CSV/Excel/JSON data processing, encoding handling, batch operations, and data…
Score 70/100
Load and preprocess imaging mass cytometry (IMC) and MIBI data. Covers MCD/TIFF handling, hot pixel removal, and image normalization.
Score 70/100
Draft and fill data privacy agreement templates — DPA, data processing agreement, GDPR, HIPAA BAA, business associate agreement, AI addendum.
Score 70/100
Data privacy and regulatory compliance specialist for GDPR, CCPA, HIPAA, and international data protection laws.
Score 70/100
GPU-accelerated data curation for LLM training. Supports text/image/video/audio. Features fuzzy deduplication (16× faster), quality filtering (30+ heuristics), semantic…
Score 70/100
Scalable data processing for ML workloads. Streaming execution across CPU/GPU, supports Parquet/CSV/JSON/images. Integrates with Ray Train, PyTorch, TensorFlow.
Score 70/100
Process and transform arrays of data with common operations like filtering, mapping, and aggregation
Score 70/100
Data product design patterns with contracts, SLAs, and governance for building self-serve data platforms using Data Mesh principles.
Score 70/100
Profile datasets to understand schema, quality, and characteristics. Use when analyzing data files (CSV, JSON, Parquet), discovering dataset properties, assessing data quality, or…
Score 70/100
Audit datasets for completeness, consistency, accuracy, and validity. Profile data distributions, detect anomalies and outliers, surface structural issues, and produce an…
Score 70/100
Data Quality Checker - Auto-activating skill for Data Pipelines. Triggers on: data quality checker, data quality checker Part of the Data Pipelines skill category.
Score 70/100
Enforce data quality rules and validations on pilot data streams and repositories. Use when checking for missing values, schema compliance, consistency issues, or anomalies before…
Score 70/100
Implement data quality validation with Great Expectations, dbt tests, and data contracts. Use when building data quality pipelines, implementing validation rules, or establishing…
Score 70/100
Implement data quality validation with Great Expectations, dbt tests, and data contracts. Use when building data quality pipelines, implementing validation rules, or establishing…
Score 70/100
Techniques and tools for ensuring the accuracy, completeness, and reliability of data across the pipeline.
Score 70/100
Profiles data assets to assess quality dimensions, detect anomalies, and generate comprehensive data quality reports with actionable recommendations.
Score 70/100
See the main Data Validation Rules skill for comprehensive coverage of data quality rule implementation.
Score 70/100
Write and verify SQL queries with BigQuery. Use when executing bq commands, writing SQL queries, or including query results in documents.
Score 70/100
Patterns for reconciling Salesforce data with external systems: count-level, field-level, and record-level reconciliation, external ID upsert patterns, Change Data Capture for…
Score 70/100
Build and refresh eval datasets from Front, run routing evals, and analyze agent response quality.
Score 70/100
Build and maintain the investor data room — organize financials, metrics, legal docs, contracts, and customer references for fundraising due diligence.
Score 70/100
Use when designing database schemas, need to model domain entities and relationships clearly, building knowledge graphs or ontologies, creating API data models, defining system…
Score 70/100
TradingView 데이터 스크래핑 자동화 파이프라인 가이드. 사용 시점: (1) 스크래퍼 실행 방법 문의 시, (2) 데이터 파이프라인 아키텍처 이해 필요 시, (3) DB 업로드 설정/디버깅 시, (4) 스크래핑 자동화 확장 작업 시.
Score 70/100
数据语义服务 API - 提供表单视图的语义理解功能。 用于: (1) 查询字段语义和业务对象识别结果 (2) 触发/批量理解表单视图 (3) 批量业务对象匹配
Score 70/100
Efficient data serialization for game networking including Protobuf, FlatBuffers, and custom binary
Score 70/100
Diagnose and mitigate Salesforce data skew — ownership skew (single user owns >10,000 records) and parent-child skew (>10,000 children under one parent) — that cause sharing…
Score 70/100
본 스킬은 대한민국 공공데이터포털(https://www.data.go.kr/)에서 제공하는 각종 공공데이터 API 를 사용하기 위한 설명입니다. 본 스킬은 공공데이터 개발 또는 공공 API 개발을 할 때 사용하면 됩니다.
Score 70/100
Connect your own data source to replace the demo unicorns data. Use when the user wants to use their own database URL or CSV file instead of the sample data.
Score 70/100
Optimize provider selection, routing, and credit usage across 150+ enrichment sources for company/contact intelligence.
Score 70/100
Optimize provider selection, routing, and credit usage across 150+ enrichment sources for company/contact intelligence.
Score 70/100
Requêtes SQL avancées pour l'analytique — window functions, CTEs récursives, pivots et optimisation de requêtes complexes.
Score 70/100
Recommend basic data structures for a task. Use when a junior developer needs help choosing lists, maps, or sets.
Score 70/100
Give agents persistent structural memory of a codebase — navigate dependencies, track public APIs, and understand why connections exist without re-reading the whole repo.
Score 70/100
Use when disjoint sets, union-find, dynamic connectivity, connected components, weighted union, union by rank, path compression, inverse Ackermann bounds, linked-list set…
Score 70/100
Analyze fundamental data primitives, type systems, and state management patterns in a codebase. Use when (1) evaluating typing strategies (Pydantic vs TypedDict vs loose dicts),…
Score 70/100
Conception de dashboards Tableau incluant calculated fields, LOD expressions, actions et storytelling.
Score 70/100
Manage AI training data, monitor content freshness, detect repetition, and update training samples for continuous learning.
Score 70/100
Transform, clean, reshape, and preprocess data using pandas and numpy. Works with ANY LLM provider (GPT, Gemini, Claude, etc.).
Score 70/100
Transform, clean, reshape, and preprocess data using pandas and numpy. Works with ANY LLM provider (GPT, Gemini, Claude, etc.).
Score 70/100
Centralized transformation logic for consistent data shaping across API routes. Includes aggregators, rankers, trend calculators, and data sanitizers.
Score 70/100
Convert between data formats (JSON, CSV, XML, YAML, TOML). Handles nested structures, arrays, and preserves data types where possible.
Score 70/100
QA an analysis before sharing -- methodology, accuracy, and bias checks
Score 70/100
Use when implementing data validation for API payloads, form inputs, or database writes. Triggers for: Pydantic models, Zod schemas, input sanitization, type validation, field…
Score 70/100
Generate interactive validation reports with quality scoring, missing data analysis, and type checking.
Score 70/100
Implementing comprehensive validation rules across database, application, and pipeline layers to ensure data integrity.
Score 70/100
Creating effective data visualizations using charts, graphs, and visual representations to communicate insights clearly and accurately following Tufte and Few principles.
Score 70/100
Data Visualization Helper - Auto-activating skill for Visual Content. Triggers on: data visualization helper, data visualization helper Part of the Visual Content skill category.
Score 70/100
Chart and visualization generation for DBX Studio. Use when a user wants to visualize data — bar charts, line graphs, pie charts, scatter plots, etc.
Score 70/100
Generate data visualizations, plots, and charts. Analyzes data structure to select optimal visualization types. supports bar charts, line graphs, and scatter plots for clarity.
Score 70/100
Create effective data visualizations with the right chart types, color palettes, and interactive features. Based on Anthropic's Claude Cookbooks (vision capabilities).
Score 70/100
Create publication-quality plots and visualizations using matplotlib and seaborn. Works with ANY LLM provider (GPT, Gemini, Claude, etc.).
Score 70/100