{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"lighteval-multilingual","owner":"swiss-ai","isFork":true,"description":"LightEval is a lightweight LLM evaluation suite that Hugging Face has been using internally with the recently released LLM data processing library datatrove and LLM training library nanotron.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":78,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-29T14:51:26.531Z"}},{"type":"Public","name":"nanotron","owner":"swiss-ai","isFork":true,"description":"Minimalistic large language model 3D-parallelism training","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":1,"starsCount":7,"forksCount":107,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-19T11:12:15.545Z"}},{"type":"Public","name":"nanotron-multilingual","owner":"swiss-ai","isFork":true,"description":"A copy of nanotron for multilingual training","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":0,"starsCount":0,"forksCount":107,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-08T09:47:19.417Z"}},{"type":"Public","name":"llm-proxy","owner":"swiss-ai","isFork":true,"description":"LLM Serving and User Control","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-30T15:48:34.318Z"}},{"type":"Public","name":"containers","owner":"swiss-ai","isFork":false,"description":"Containers for multimodal initiative (and maybe more across Swiss AI?)","allTopics":[],"primaryLanguage":{"name":"Dockerfile","color":"#384d54"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-07T16:22:05.189Z"}},{"type":"Public","name":"ml-4m-v2","owner":"swiss-ai","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-05T21:07:04.988Z"}},{"type":"Public","name":"ml-4m","owner":"swiss-ai","isFork":true,"description":"4M: Massively Multimodal Masked Modeling (NeurIPS 2023 Spotlight)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":4,"issueCount":13,"starsCount":0,"forksCount":90,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-05T14:51:30.148Z"}},{"type":"Public","name":"video2dataset","owner":"swiss-ai","isFork":true,"description":"Easily create large video dataset from video urls","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":1,"forksCount":65,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-02T12:00:19.837Z"}},{"type":"Public","name":"vllm","owner":"swiss-ai","isFork":true,"description":"A high-throughput and memory-efficient inference and serving engine for LLMs","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":4063,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-31T09:13:57.522Z"}},{"type":"Public","name":"data-tooling","owner":"swiss-ai","isFork":true,"description":"Tool set for data preparation and selection in the context of Swiss-AI (forked from DataTrove)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":139,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-15T09:34:57.698Z"}},{"type":"Public","name":"data-PDF-pipeline","owner":"swiss-ai","isFork":false,"description":"PDF pipeline for creating training corpora (mainly for llm, multimodal and alignment horizontals)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-08T09:07:13.012Z"}},{"type":"Public","name":"MoE","owner":"swiss-ai","isFork":false,"description":"some mixture of experts architecture implementations","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":5,"forksCount":1,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-22T14:20:16.102Z"}},{"type":"Public","name":"Megatron-LLM","owner":"swiss-ai","isFork":true,"description":"distributed trainer for LLMs","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":76,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-08T21:00:36.231Z"}}],"repositoryCount":13,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"swiss-ai repositories"}