{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"DiffSynth-Studio","owner":"modelscope","isFork":false,"description":"Enjoy the magic of Diffusion models!","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":41,"starsCount":5594,"forksCount":510,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,1,2,3,4,7,0,1,3,6,0,0,0,0,1,0,0,2,0,0,1,0,3,2,0,0,6,3,1,20,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T05:37:09.244Z"}},{"type":"Public","name":"FunASR","owner":"modelscope","isFork":false,"description":"A Fundamental End-to-End Speech Recognition Toolkit and Open Source SOTA Pretrained Models, Supporting Speech Recognition, Voice Activity Detection, Text Post-processing etc.","allTopics":["pytorch","speech-recognition","vad","punctuation","whisper","audio-visual-speech-recognition","speaker-diarization","voice-activity-detection","conformer","pretrained-model","rnnt","dfsmn","paraformer","speechgpt","speechllm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":187,"starsCount":4534,"forksCount":504,"license":"Other","participation":[26,54,71,41,105,45,44,18,31,48,25,23,0,42,40,19,6,61,39,41,15,24,16,36,16,18,46,63,46,20,13,0,98,84,55,34,51,63,21,9,17,21,38,29,32,33,23,47,89,24,18,17],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T02:21:57.215Z"}},{"type":"Public","name":"data-juicer","owner":"modelscope","isFork":false,"description":"A one-stop data processing system to make data higher-quality, juicier, and more digestible for (multimodal) LLMs! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷为大模型提供更高质量、更丰富、更易”消化“的数据!","allTopics":["nlp","data-science","opendata","data-visualization","pytorch","dataset","chinese","data-analysis","llama","gpt","multi-modal","pre-training","sora","streamlit","gpt-4","large-language-models","llm","llms","instruction-tuning","llava"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":7,"issueCount":7,"starsCount":1707,"forksCount":112,"license":"Apache License 2.0","participation":[0,0,0,10,5,0,2,0,2,1,5,1,0,2,4,5,2,1,9,6,4,2,2,5,6,4,4,4,4,3,1,0,5,5,4,16,12,6,0,4,7,5,1,4,3,0,2,2,0,0,3,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-05T02:21:28.615Z"}},{"type":"Public","name":"swift","owner":"modelscope","isFork":false,"description":"ms-swift: Use PEFT or Full-parameter to finetune 300+ LLMs or 40+ MLLMs. (Qwen2, GLM4, Internlm2.5, Yi, Llama3, Llava, MiniCPM-V, Deepseek, Baichuan2, Gemma2, Phi3-Vision, ...)","allTopics":["agent","deploy","llama","lora","gemma","peft","multimodal","sft","dpo","pre-training","awq","llm","modelscope","llava","qwen","qwen2","unsloth","llama3","glm4","internvl"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":158,"starsCount":2230,"forksCount":214,"license":"Apache License 2.0","participation":[0,0,0,9,18,3,3,3,8,8,16,11,1,6,5,1,11,5,8,11,5,14,11,19,7,16,25,9,14,20,28,0,13,16,18,33,15,17,11,13,28,22,19,20,22,24,15,28,15,25,20,29],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-04T23:38:14.518Z"}},{"type":"Public","name":"eval-scope","owner":"modelscope","isFork":false,"description":"A streamlined and customizable framework for efficient large model evaluation and performance benchmarking","allTopics":["performance","evaluation","llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":9,"starsCount":98,"forksCount":12,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,3,28,2,0,4,3,4,19,0,0,0,5,18,2,1,7,2,1,1,1,0,1,0,0,0,3,8,0,0,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-04T21:02:38.534Z"}},{"type":"Public","name":"modelscope-agent","owner":"modelscope","isFork":false,"description":"ModelScope-Agent: An agent framework connecting models in ModelScope with the world","allTopics":["agent","chatbot","android-application","multi-agents","rag","mobile-agents","gpts","llm","multimodal-large-language-models","qwen","assistantapi","chatglm-4","open-gpts","mobile-agent"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":3,"issueCount":52,"starsCount":2200,"forksCount":244,"license":"Apache License 2.0","participation":[1,0,0,1,1,3,4,4,59,2,7,3,0,0,0,2,2,7,16,23,38,19,31,15,1,3,1,1,0,2,2,1,0,5,4,9,10,5,1,6,12,4,6,3,11,9,1,11,9,7,5,11],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-04T14:59:52.536Z"}},{"type":"Public","name":"agentscope","owner":"modelscope","isFork":false,"description":"Start building LLM-empowered multi-agent applications in an easier way.","allTopics":["agent","chatbot","multi-agent","multi-modal","distributed-agents","gpt-4","large-language-models","llm","llm-agent","llama3","gpt-4o"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":9,"issueCount":14,"starsCount":2859,"forksCount":194,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,8,9,2,3,0,3,10,5,8,16,15,12,7,8,7,9,5,15,6,10,8,10,9,3,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-04T09:06:44.886Z"}},{"type":"Public","name":"modelscope","owner":"modelscope","isFork":false,"description":"ModelScope: bring the notion of Model-as-a-Service to life.","allTopics":["nlp","science","cv","speech","multi-modal","python","machine-learning","deep-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":55,"starsCount":6452,"forksCount":671,"license":"Apache License 2.0","participation":[42,19,55,36,18,17,25,44,30,15,23,49,1,20,23,18,7,11,10,1,7,9,4,3,12,7,6,9,4,1,1,0,11,9,12,2,0,3,0,4,4,3,0,0,4,1,10,6,0,0,1,7],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-04T06:03:16.071Z"}},{"type":"Public","name":"FunClip","owner":"modelscope","isFork":false,"description":"Open-source, accurate and easy-to-use video speech recognition & clipping tool, LLM based AI clipping intergrated.","allTopics":["speech-recognition","speech-to-text","gradio","video-clip","subtitles-generator","video-subtitles","llm","gradio-python-llm"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":8,"starsCount":2697,"forksCount":267,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,11,7,0,0,3,0,0,0,23,6,30,29,12,4,2,6,4,4,5],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-04T05:39:49.916Z"}},{"type":"Public","name":"modelscope-classroom","owner":"modelscope","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":164,"forksCount":20,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,9,6,1,1,0,2,4,0,0,0,2,0,0,7,1,29,0,3],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-03T12:25:30.696Z"}},{"type":"Public","name":"facechain","owner":"modelscope","isFork":false,"description":"FaceChain is a deep-learning toolchain for generating your Digital-Twin.","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":5,"starsCount":8680,"forksCount":810,"license":"Apache License 2.0","participation":[0,0,0,0,7,81,22,73,34,31,84,5,0,4,4,5,4,8,0,1,8,2,1,5,0,0,3,1,0,0,0,0,0,1,1,3,0,7,0,1,0,0,0,0,0,0,4,2,1,1,2,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-03T03:50:49.572Z"}},{"type":"Public","name":"dash-infer","owner":"modelscope","isFork":false,"description":"DashInfer is a native LLM inference engine aiming to deliver industry-leading performance atop various hardware architectures, including x86 and ARMv9.","allTopics":["cpu","llm","llm-inference","native-engine"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":4,"starsCount":102,"forksCount":11,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,5,0,2,1,0,2,0,5,3,0,0,3,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-01T05:06:27.519Z"}},{"type":"Public","name":"richdreamer","owner":"modelscope","isFork":false,"description":"Live Demo:https://modelscope.cn/studios/Damo_XR_Lab/3D_AIGC","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":6,"starsCount":359,"forksCount":13,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,8,5,2,0,3,14,0,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-30T07:31:17.760Z"}},{"type":"Public","name":"modelscope-studio","owner":"modelscope","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"TypeScript","color":"#3178c6"},"pullRequestCount":0,"issueCount":1,"starsCount":26,"forksCount":4,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,9,1,1,4,4,4,5,0,0,2,1,4,0,0,0,3,0,0,2,3,1,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-27T09:11:35.378Z"}},{"type":"Public","name":"3D-Speaker","owner":"modelscope","isFork":false,"description":"A Repository for Single- and Multi-modal Speaker Verification, Speaker Recognition and Speaker Diarization","allTopics":["speaker-verification","speaker-diarization","language-identification","voxceleb","modelscope","campplus","eres2net","3d-speaker","rdino","cnceleb"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":907,"forksCount":77,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-25T06:41:03.445Z"}},{"type":"Public","name":"scepter","owner":"modelscope","isFork":false,"description":"SCEPTER is an open-source framework used for training, fine-tuning, and inference with generative models.","allTopics":["generative-model","scedit","aigc","lar-gen","stylebooth"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":285,"forksCount":16,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,2,3,0,3,9,0,10,0,0,1,0,0,0,0,8,2,1,9,3,0,0,0,6,4,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-05T08:14:32.279Z"}},{"type":"Public","name":"lite-sora","owner":"modelscope","isFork":false,"description":"An initiative to replicate Sora","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":92,"forksCount":6,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,7,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-04-10T08:35:58.543Z"}},{"type":"Public","name":"normal-depth-diffusion","owner":"modelscope","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":2,"starsCount":110,"forksCount":8,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,5,0,6,0,1,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-07T14:39:51.966Z"}},{"type":"Public","name":"FunCodec","owner":"modelscope","isFork":false,"description":"FunCodec is a research-oriented toolkit for audio quantization and downstream applications, such as text-to-speech synthesis, music generation et.al. ","allTopics":["tts","speech-synthesis","codec","speech-to-text","audio-generation","encodec","voicecloning","audio-quantization"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":14,"starsCount":305,"forksCount":26,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,5,0,0,13,0,0,0,0,0,0,33,0,7,0,11,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-25T11:56:17.823Z"}},{"type":"Public","name":"KAN-TTS","owner":"modelscope","isFork":false,"description":"KAN-TTS is a speech-synthesis training framework, please try the demos we have posted at https://modelscope.cn/models?page=1&tasks=text-to-speech","allTopics":["modelscope","speech","tts","speech-synthesis"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":40,"starsCount":464,"forksCount":75,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-28T06:21:05.296Z"}},{"type":"Public","name":"AdaSeq","owner":"modelscope","isFork":false,"description":"AdaSeq: An All-in-One Library for Developing State-of-the-Art Sequence Understanding Models","allTopics":["natural-language-processing","information-extraction","chinese-nlp","word-segmentation","bert","sequence-labeling","relation-extraction","natural-language-understanding","entity-typing","token-classification","multi-modal-ner","nlp","crf","pytorch","named-entity-recognition","ner"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":25,"starsCount":381,"forksCount":34,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,5,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-15T12:08:20.988Z"}},{"type":"Public","name":"motionagent","owner":"modelscope","isFork":false,"description":"MotionAgent is your AI assistent to convert ideas into motion pictures.","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":265,"forksCount":31,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,6,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-09-07T14:44:46.195Z"}},{"type":"Public","name":"AdaDet","owner":"modelscope","isFork":false,"description":"AdaDet: A Development Toolkit for Object Detection based on ModelScope","allTopics":["python","tracking","machine-learning","ocr","computer-vision","deep-learning","domain-specific-models","keypoint-detection","detection-toolkit","detection","face-detection","object-detection"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":44,"forksCount":7,"license":"Apache License 2.0","participation":[0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-07-17T07:31:31.697Z"}},{"type":"Public","name":"kws-training-suite","owner":"modelscope","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":73,"forksCount":15,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-05-26T06:25:17.442Z"}}],"repositoryCount":24,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"modelscope repositories"}