{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"llumnix","owner":"AlibabaPAI","isFork":false,"description":"Efficient and easy multi-instance LLM serving","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":0,"starsCount":109,"forksCount":9,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-10T03:50:00.461Z"}},{"type":"Public","name":"xla","owner":"AlibabaPAI","isFork":true,"description":"Enabling PyTorch on XLA Devices (e.g. Google TPU)","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":3,"issueCount":0,"starsCount":1,"forksCount":453,"license":"Other","participation":[26,20,9,14,14,8,17,10,14,12,11,33,48,71,21,3,12,26,26,31,45,33,20,34,22,37,23,8,9,6,2,3,9,0,0,1,3,3,3,0,2,2,0,3,2,0,1,1,0,0,1,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-09T09:56:05.555Z"}},{"type":"Public","name":"FLASHNN","owner":"AlibabaPAI","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":1,"starsCount":67,"forksCount":7,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,3,2,1,2,3,2,0,0,1,1,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-09T08:22:40.123Z"}},{"type":"Public","name":"torchacc","owner":"AlibabaPAI","isFork":false,"description":"PyTorch distributed training acceleration framework","allTopics":["pytorch","xla","llm-training"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":0,"starsCount":14,"forksCount":2,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,5,1,1,4,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-09T06:35:22.328Z"}},{"type":"Public","name":"FlashModels","owner":"AlibabaPAI","isFork":false,"description":"Fast and easy distributed model training examples.","allTopics":["deep-learning","pytorch","zero","data-parallelism","model-parallelism","distributed-training","xla","tensor-parallelism","llm","fsdp","sequence-parallelism"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":0,"starsCount":6,"forksCount":3,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,2,0,2,1,0,1,0,0,1,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-05T11:57:25.517Z"}},{"type":"Public","name":"transformers","owner":"AlibabaPAI","isFork":true,"description":"🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":0,"forksCount":26238,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-27T12:13:00.268Z"}},{"type":"Public","name":"accelerate","owner":"AlibabaPAI","isFork":true,"description":"🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":926,"license":"Apache License 2.0","participation":[16,2,5,5,11,4,10,10,15,19,3,8,14,5,6,5,7,15,10,6,9,14,13,9,20,11,8,10,9,13,15,8,12,14,10,6,7,6,10,12,3,4,12,8,4,0,0,0,0,0,2,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-27T09:45:13.851Z"}},{"type":"Public","name":"openxla","owner":"AlibabaPAI","isFork":true,"description":"A machine learning compiler for GPUs, CPUs, and ML accelerators","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":1,"issueCount":0,"starsCount":0,"forksCount":400,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-27T03:37:00.511Z"}},{"type":"Public","name":"Megatron","owner":"AlibabaPAI","isFork":true,"description":"Ongoing research training transformer models at scale","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":2237,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-26T09:00:29.387Z"}},{"type":"Public","name":"pytorch","owner":"AlibabaPAI","isFork":true,"description":"Tensors and Dynamic neural networks in Python with strong GPU acceleration","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":22005,"license":"Other","participation":[281,251,214,282,257,206,302,283,309,328,168,254,278,249,191,102,233,222,267,287,303,281,241,218,284,263,9,5,6,23,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-05T03:31:44.708Z"}},{"type":"Public","name":"OLMo","owner":"AlibabaPAI","isFork":true,"description":"Modeling, training, eval, and inference code for OLMo","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":433,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-04T15:22:36.727Z"}},{"type":"Public archive","name":"Modelzoo-Data","owner":"AlibabaPAI","isFork":false,"description":"DeepRec modelzoo's data set","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":2,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-02-16T12:14:35.158Z"}},{"type":"Public","name":"EasyParallelLibrary","owner":"AlibabaPAI","isFork":true,"description":"Easy Parallel Library (EPL) is a general and efficient deep learning framework for distributed giant model training.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":49,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-02-28T07:14:47.002Z"}},{"type":"Public","name":"nni","owner":"AlibabaPAI","isFork":true,"description":"An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1806,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-11-30T06:59:53.945Z"}},{"type":"Public","name":"DAPPLE","owner":"AlibabaPAI","isFork":false,"description":"An Efficient Pipelined Data Parallel Approach for Training Large Model","allTopics":["pipeline-parallelism","hybrid-parallelism","distribution-strategy-planner"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":69,"forksCount":15,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-12-11T03:13:47.726Z"}},{"type":"Public","name":"one_shot_text_labeling","owner":"AlibabaPAI","isFork":false,"description":"code and data for paper \"One-shot Text Field Labeling using Attention and BeliefPropagation for Structure Information Extraction\"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":3,"starsCount":61,"forksCount":6,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-08-09T02:40:47.373Z"}},{"type":"Public","name":"sparsehash-c11","owner":"AlibabaPAI","isFork":true,"description":"Experimental C++11 version of sparsehash","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":59,"license":"BSD 3-Clause \"New\" or \"Revised\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-05-26T13:03:27.701Z"}},{"type":"Public","name":"gradient-checkpointing","owner":"AlibabaPAI","isFork":true,"description":"Make huge neural nets fit in memory","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":270,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2018-01-19T20:53:11.544Z"}},{"type":"Public","name":"tpu-demos","owner":"AlibabaPAI","isFork":true,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1772,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2018-01-09T17:12:55.901Z"}}],"repositoryCount":19,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"AlibabaPAI repositories"}