Coding
PromptBeginner5 minmarkdown
Markdown Converter
Agent skill for markdown-converter
7
This project explores using human-readable explanations as machine learning models themselves. By leveraging Large Language Models (LLMs) to interpret and apply natural language rules, we can create inherently interpretable models that can be refined interactively.
Sign in to like and favorite skills
This project explores using human-readable explanations as machine learning models themselves. By leveraging Large Language Models (LLMs) to interpret and apply natural language rules, we can create inherently interpretable models that can be refined interactively.
make: Build everything (all datasets and analysis)make wisconsin: Build Wisconsin dataset resultsmake titanic: Build Titanic dataset resultsmake southgermancredit: Build South German Credit dataset resultsmake ensembles: Generate ensemble model resultsuv run <script.py>: Run Python scripts with UV package managersqlite3 <file.sqlite> < file.sql: Create/update SQLite databaseThe project evaluates various LLMs including:
results_chart_by_size.py: Compare model performance vs. model sizeresults_chart_by_elo.py: Compare model performance vs. ELO ratingresults_error_rate_by_wordcount.py: Analyze error rates relative to prompt complexityresults_error_rate_by_herdan.py: Analyze error rates relative to lexical complexityresults_ensembling.py: Create ensemble models from multiple base models
language_models table to track model release datesensemble_results table for later analysispython results_ensembling.py titanic --summary outputs/titanic_ensemble_summary.txtresultssampleimpact.py: Measure the impact of sample count on model performanceconfigs/: Configuration JSON files for datasetsdatasets/: CSV data filesdbtemplates/: SQL templates for database initializationenvs/: Environment files for different models (structure: envs/{dataset}/{model}.env)modules/: Core functionality modulesresults/: Output files and SQLite databasesoutputs/: Generated charts, tables, and CSV resultsobfuscations/: Dataset obfuscation plansconversions/: Dataset conversion/encoding guidelinespostgres-schemas/model_release_dates.sql: Table definition and data for language models used in chronological ensemblingpostgres-schemas/ensemble_results.sql: Schema for storing ensemble evaluation resultsinitialise_database.py: Set up task databasestrain.py: Run training iterationspredict.py: Make predictions using trained modelsreport-script.py: Generate performance reportsmake_result_charts.py: Create visualization chartscreate_task_csv_file.py: Generate CSV results from environment fileslexicostatistics.py: Calculate Herdan and Zipf coefficients for a
language model after all investigations have completedenv_settings.py: Parse and validate model environment settingsresultdistribution.py: Generate distribution charts for model outputsbaseline.py: Create baseline model performance metricsobfuscation_plan_generator.py: Generate dataset obfuscation plans