{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"VLMEvalKit","owner":"open-compass","isFork":false,"description":"Open-source evaluation toolkit of large vision-language models (LVLMs), support ~100 VLMs, 40+ benchmarks","allTopics":["computer-vision","evaluation","pytorch","gemini","openai","vqa","vit","gpt","multi-modal","clip","claude","openai-api","gpt4","large-language-models","llm","chatgpt","llava","qwen","gpt-4v"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":8,"issueCount":22,"starsCount":977,"forksCount":137,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,10,88,18,14,17,25,31,28,24,20,25,2,3,4,6,3,9,12,12,2,12,33,23,6,23,16,13,9,2,4,27,12,14,47,46,36,33,34,4,39,25,19],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-09T14:29:03.530Z"}},{"type":"Public","name":"opencompass","owner":"open-compass","isFork":false,"description":"OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.","allTopics":["benchmark","evaluation","openai","llm","chatgpt","large-language-model","llama2","llama3"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":25,"issueCount":173,"starsCount":3698,"forksCount":395,"license":"Apache License 2.0","participation":[10,20,9,4,6,3,12,9,15,12,21,12,7,16,13,12,11,13,12,14,7,11,0,7,7,11,12,9,2,5,3,5,20,6,11,13,8,11,3,0,0,6,8,4,8,18,10,7,5,15,6,13],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-06T11:17:40.049Z"}},{"type":"Public","name":"MMBench","owner":"open-compass","isFork":false,"description":"Official Repo of \"MMBench: Is Your Multi-modal Model an All-around Player?\"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":1,"starsCount":141,"forksCount":10,"license":"Apache License 2.0","participation":[0,0,0,5,0,0,2,0,0,1,0,0,1,0,0,2,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-01T11:44:12.290Z"}},{"type":"Public","name":"hinode","owner":"open-compass","isFork":true,"description":"A clean documentation and blog theme for your Hugo site based on Bootstrap 5","allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":50,"license":"MIT License","participation":[9,24,29,6,56,94,27,67,59,12,22,0,18,17,24,164,73,12,14,35,35,46,17,27,30,15,40,28,25,13,3,0,41,23,31,22,46,41,34,16,27,18,19,26,17,30,22,35,86,29,22,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-01T03:11:24.280Z"}},{"type":"Public","name":"storage","owner":"open-compass","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-18T18:25:33.773Z"}},{"type":"Public","name":"CompassBench","owner":"open-compass","isFork":false,"description":"Demo data of CompassBench","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":2,"starsCount":2,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-08-07T12:39:22.420Z"}},{"type":"Public","name":"CIBench","owner":"open-compass","isFork":false,"description":"Official Repo of \"CIBench: Evaluation of LLMs as Code Interpreter \"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":3,"forksCount":1,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,2,2,1,3,0,1,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-19T04:51:48.990Z"}},{"type":"Public","name":"GAOKAO-Eval","owner":"open-compass","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":2,"starsCount":86,"forksCount":6,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,8,0,0,0,10,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-17T10:58:18.281Z"}},{"type":"Public","name":"ANAH","owner":"open-compass","isFork":false,"description":"[ACL 2024] ANAH: Analytical Annotation of Hallucinations in Large Language Models","allTopics":["acl","gpt","llms","hallucination-detection"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":17,"forksCount":1,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-12T18:57:53.120Z"}},{"type":"Public","name":"MathBench","owner":"open-compass","isFork":false,"description":"[ACL 2024 Findings] MathBench: A Comprehensive Multi-Level Difficulty Mathematics Evaluation Dataset","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":3,"starsCount":76,"forksCount":1,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,4,0,0,0,0,0,0,6,0,0,0,0,0,0,0,0,0,9,2,0,0,0,0,4,2,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-12T05:17:51.572Z"}},{"type":"Public","name":"GTA","owner":"open-compass","isFork":false,"description":"Official repository for paper \"GTA: A Benchmark for General Tool Agents\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":27,"forksCount":2,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-12T03:23:24.351Z"}},{"type":"Public","name":".github","owner":"open-compass","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T02:27:09.928Z"}},{"type":"Public","name":"DevBench","owner":"open-compass","isFork":false,"description":"A Comprehensive Benchmark for Software Development.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":84,"forksCount":4,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-30T13:10:52.870Z"}},{"type":"Public","name":"CodeBench","owner":"open-compass","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-21T11:38:31.710Z"}},{"type":"Public","name":"Ada-LEval","owner":"open-compass","isFork":false,"description":"The official implementation of \"Ada-LEval: Evaluating long-context LLMs with length-adaptable benchmarks\"","allTopics":["gpt4","llm","long-context"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":48,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-22T09:36:27.993Z"}},{"type":"Public","name":"T-Eval","owner":"open-compass","isFork":false,"description":"[ACL2024] T-Eval: Evaluating Tool Utilization Capability of Large Language Models Step by Step","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":32,"starsCount":207,"forksCount":12,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-03T21:05:37.907Z"}},{"type":"Public","name":"human-eval","owner":"open-compass","isFork":true,"description":"Code for the paper \"Evaluating Large Language Models Trained on Code\"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":326,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-14T11:55:53.553Z"}},{"type":"Public","name":"OpenFinData","owner":"open-compass","isFork":false,"description":"","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":3,"starsCount":32,"forksCount":2,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-08T06:32:20.542Z"}},{"type":"Public","name":"CriticBench","owner":"open-compass","isFork":false,"description":"A comprehensive benchmark for evaluating critique ability of LLMs","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":25,"forksCount":1,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,23,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-24T01:31:08.418Z"}},{"type":"Public","name":"code-evaluator","owner":"open-compass","isFork":false,"description":"A multi-language code evaluation tool.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":17,"forksCount":6,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,8,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-26T04:12:27.117Z"}},{"type":"Public","name":"evalplus","owner":"open-compass","isFork":true,"description":"EvalPlus for rigourous evaluation of LLM-synthesized code","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":102,"license":"Apache License 2.0","participation":[0,0,2,1,7,4,26,2,7,2,13,5,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-20T08:13:21.996Z"}},{"type":"Public","name":"MixtralKit","owner":"open-compass","isFork":false,"description":"A toolkit for inference and evaluation of 'mixtral-8x7b-32kseqlen' from Mistral AI","allTopics":["moe","mistral","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":12,"starsCount":763,"forksCount":81,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-15T19:10:55.603Z"}},{"type":"Public","name":"LawBench","owner":"open-compass","isFork":false,"description":"Benchmarking Legal Knowledge of Large Language Models","allTopics":["law","benchmark","llm","chatgpt"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":232,"forksCount":31,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-13T06:42:45.191Z"}},{"type":"Public","name":"BotChat","owner":"open-compass","isFork":false,"description":"Evaluating LLMs' multi-round chatting capability via assessing conversations generated by two LLM instances.","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":1,"starsCount":128,"forksCount":6,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-02T12:38:19.203Z"}},{"type":"Public","name":"pytorch_sphinx_theme","owner":"open-compass","isFork":true,"description":"Sphinx Theme for OpenCompass - Modified from PyTorch","allTopics":[],"primaryLanguage":{"name":"CSS","color":"#563d7c"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":138,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-08-30T04:20:35.062Z"}}],"repositoryCount":25,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"open-compass repositories"}