--- dataset_info: features: - name: text dtype: string splits: - name: train num_bytes: 4483331 num_examples: 342 - name: validation num_bytes: 622617 num_examples: 39 download_size: 2534957 dataset_size: 5105948 configs: - config_name: default data_files: - split: train path: data/train-* - split: validation path: data/validation-* license: apache-2.0 --- # Q Code Pretraining Corpus This dataset provides a corpus of Q programming language code and documentation, curated for pretraining large language models and code models. ## 📊 Dataset Overview - **Total Data**: Over 1.6 million Q tokens, 5+ million characters - **Documents**: 342 training chunks, 39 validation chunks - **Source Types**: - Open-source Q repositories (MIT/Apache 2.0 licenses) - Official KDB+/Q documentation and tutorials - Hand-curated code snippets and scripts - **Format**: Cleaned, deduplicated, chunked for efficient pretraining ## 🎯 Key Features - **Q-Only**: All data is pure Q language (no mixed Python or non-code noise) - **Permissive Licensing**: All source code is MIT or Apache 2.0, suitable for both research and commercial use - **Coverage**: Includes code from analytics, time-series, database queries, and utilities - **Filtered & Scored**: LLM-assisted quality scoring plus manual review for top-tier data fidelity - **Chunked & Ready**: Delivered as 4k-token chunks for immediate use with Hugging Face, TRL, or custom pipelines ## 🏗️ Dataset Structure Each record is a text chunk, containing code or documentation in Q. Splits: - `train`: Main corpus for pretraining (342 chunks) - `validation`: Holdout set for evaluation (39 chunks) Sample record: ```python { "text": str # Raw Q code or documentation chunk } ``` ## 🧑‍💻 Usage ### Loading the Dataset ```python from datasets import load_dataset # Load the full Q pretraining dataset dataset = load_dataset("morganstanley/q_pretrained_dataset") # Access splits train_data = dataset["train"] val_data = dataset["validation"] ``` ### Example: Previewing Data ```python sample = dataset["train"][0] print(sample["text"]) ``` ### Training Usage This dataset is designed for language model pretraining using next-token prediction or masked language modeling objectives. Supports efficient training with Hugging Face Transformers, TRL, or custom frameworks. ## 🔤 About Q Programming Language Q is a vector and array programming language developed by Kx Systems for high-performance analytics, finance, and time-series applications. It features: - Concise, functional, array-oriented syntax - Powerful built-in operators for large-scale data manipulation - Industry adoption in trading, banking, and real-time analytics ## 📁 Source Repositories Major open-source Q repos included: - DataIntellectTech/TorQ - psaris/qtips - psaris/funq - KxSystems/ml - finos/kdb - LeslieGoldsmith/qprof - jonathonmcmurray/reQ - ...and more All with permissive licenses (MIT or Apache 2.0). ## 📈 Data Preparation & Filtering - **Automated Scoring**: Qwen-2.5-32B was used to score each file (0–10) for quality and relevance; only files scoring ≥4 were included. - **Manual Review**: Additional cleaning to remove non-Q files or low-value content. - **Deduplication**: Duplicate and boilerplate code removed. ## 📝 Citation If you use this dataset in your research, please cite: ```bibtex @dataset{q_pretraining_corpus_2024, title={Q Code Pretraining Corpus}, author={Brendan Rappazzo Hogan}, year={2024}, url={https://huggingface.co/datasets/bhogan/q-pretraining-corpus}, note={Dataset for domain-adaptive pretraining of language models on the Q programming language} } ``` **Associated Paper:** [Link to paper will be added here]