<!-- markdownlint-disable MD033 MD041 -->
<h1 align="center">
Explore
9,659 skills indexed with the new KISS metadata standard.
<h1 align="center">
Microsoft takes the security of our software products and services seriously, which includes all source code repositories managed through our GitHub organizations, which include [Microsoft](https://github.com/Microsoft), [Azure](https://github.com/Azure), [DotNet](https://github.com/dotnet), [AspNet
*.bin
**REPO OWNER**: Do you want Customer Service & Support (CSS) support for this product/project?
repos:
This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/).
This project welcomes contributions and suggestions. Most contributions require you to
<img src="./.asset/logo.color.svg" width="45" /> TaskWeaver
Microsoft takes the security of our software products and services seriously, which includes all source code repositories managed through our GitHub organizations, which include [Microsoft](https://github.com/Microsoft), [Azure](https://github.com/Azure), [DotNet](https://github.com/dotnet), [AspNet
generic skill
!sample/*.csv
[](https://arxiv.org/abs/2303.17580)
Microsoft takes the security of our software products and services seriously, which includes all source code repositories managed through our GitHub organizations, which include [Microsoft](https://github.com/microsoft), [Azure](https://github.com/Azure), [DotNet](https://github.com/dotnet), [AspNet
**REPO OWNER**: Do you want Customer Service & Support (CSS) support for this product/project?
*.dev.yaml
This project has adopted the [Microsoft Open Source Code of Conduct](https://opensource.microsoft.com/codeofconduct/).
1. Fork the repository you want to contribute to by clicking the "Fork" button on the project page.
.chainlit
path = reasoning-from-scratch
This repository contains the code for developing, pretraining, and finetuning a GPT-like LLM and is the official code repository for the book [Build a Large Language Model (From Scratch)](https://amzn.to/4fqvn0D).
.idea

*.ipynb linguist-generated
LLMs in simple, pure C/CUDA with no need for 245MB of PyTorch or 107MB of cPython. Current focus is on pretraining, in particular reproducing the [GPT-2](https://github.com/openai/gpt-2) and [GPT-3](https://arxiv.org/abs/2005.14165) miniseries, along with a parallel PyTorch reference implementation