{"payload":{"pageCount":5,"repositories":[{"type":"Public","name":"hifi-gan-template","owner":"inferless","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":"MIT License","participation":[13,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,12],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-17T21:55:45.435Z"}},{"type":"Public","name":"DINet","owner":"inferless","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,16],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-17T21:55:08.531Z"}},{"type":"Public","name":"template-method","owner":"inferless","isFork":false,"description":"GPT-Neo 125M is a transformer model designed using EleutherAI's replication of the GPT-3 architecture. GPT-Neo refers to the class of models, while 125M represents the number of parameters of this particular pre-trained model.","allTopics":["text-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":4,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-16T22:12:45.862Z"}},{"type":"Public","name":"template_input_batch","owner":"inferless","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-16T21:58:52.391Z"}},{"type":"Public","name":"speaker-pipeline","owner":"inferless","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-16T21:44:15.220Z"}},{"type":"Public","name":"Demuc-Pipeline","owner":"inferless","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-16T21:38:16.998Z"}},{"type":"Public","name":"translation-pipeline","owner":"inferless","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-16T21:27:12.251Z"}},{"type":"Public","name":"Stablediffusion-controlnet","owner":"inferless","isFork":false,"description":"ControlNet is a neural network structure to control diffusion models by adding extra conditions. This checkpoint corresponds to the ControlNet conditioned on Canny edges. It can be used in combination with Stable Diffusion.","allTopics":["text-to-image"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,9],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-16T20:52:24.968Z"}},{"type":"Public","name":"Facebook-bart-cnn","owner":"inferless","isFork":false,"description":"BART model pre-trained on English language, and fine-tuned on CNN Daily Mail. It was introduced in the paper BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension by Lewis et al. and first released in [this repository (https://github.com/pytorch/fairseq/tree/master/examples/bart).","allTopics":["text-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T19:20:10.092Z"}},{"type":"Public","name":"idefics-9b-instruct-8bit","owner":"inferless","isFork":false,"description":"IDEFICS (Image-aware Decoder Enhanced à la Flamingo with Interleaved Cross-attentionS) is an open-access reproduction of Flamingo, a closed-source visual language model developed by Deepmind. Like GPT-4, the multimodal model accepts arbitrary sequences of image and text inputs and produces text outputs.","allTopics":["multimodal"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":3,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T19:15:53.789Z"}},{"type":"Public","name":"Vicuna-13b-8k","owner":"inferless","isFork":false,"description":"Vicuna is a chat assistant trained by fine-tuning Llama 2 on user-shared conversations collected from ShareGPT.","allTopics":["text-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T18:36:51.522Z"}},{"type":"Public","name":"universal-sentence-encoder-multilingual-tensorflow","owner":"inferless","isFork":false,"description":"Universal-Sentence-Encoder-Multilingual-QA is a model developed by researchers at Google mainly for the purpose of question answering. You can use this template to import the model in Inferless.","allTopics":["text-embedding"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":3,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T17:58:43.087Z"}},{"type":"Public","name":"DreamShaper","owner":"inferless","isFork":false,"description":"Controlnet v1.1 was released in lllyasviel/ControlNet-v1-1 by Lvmin Zhang. This checkpoint is a conversion of the original checkpoint into diffusers format. It can be used in combination with Stable Diffusion, such as runwayml/stable-diffusion-v1-5.","allTopics":["image-to-image"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T16:52:24.919Z"}},{"type":"Public","name":"Falcon-7b-instruct","owner":"inferless","isFork":false,"description":"Falcon-7B-Instruct is a 7B parameters causal decoder-only model built by TII based on Falcon-7B and finetuned on a mixture of chat/instruct datasets. It is made available under the Apache 2.0 license.","allTopics":["text-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T16:04:49.492Z"}},{"type":"Public","name":"Llama-2-7b-hf","owner":"inferless","isFork":false,"description":"Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 13B fine-tuned model, optimized for dialogue use cases and converted for the Hugging Face Transformers format. Links to other models can be found in the index at the bottom.","allTopics":["text-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T15:42:49.475Z"}},{"type":"Public","name":"Vicuna-7b-1.1","owner":"inferless","isFork":false,"description":"Vicuna GPTQ is a text-generation model. You can use this template to import the model in Inferless.","allTopics":["text-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T15:18:12.414Z"}},{"type":"Public","name":"Vicuna-7b-8k","owner":"inferless","isFork":false,"description":"Vicuna is an open-source chatbot model that was trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT. It is an auto-regressive language model based on the transformer architecture. Vicuna is primarily used for research on large language models and chatbots.","allTopics":["text-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T14:18:29.134Z"}},{"type":"Public","name":"s3-model-import","owner":"inferless","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T13:36:21.226Z"}},{"type":"Public","name":"stable-diffusion-s3-image-save","owner":"inferless","isFork":false,"description":"This stable-diffusion-2-1 model is fine-tuned from stable-diffusion-2 (768-v-ema.ckpt) with an additional 55k steps on the same dataset (with punsafe=0.1), and then fine-tuned for another 155k extra steps with punsafe=0.98.","allTopics":["text-to-image"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T13:26:22.404Z"}},{"type":"Public","name":"Llama-2-13b-chat-hf","owner":"inferless","isFork":false,"description":"Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 13B fine-tuned model, optimized for dialogue use cases and converted for the Hugging Face Transformers format. Links to other models can be found in the index at the bottom.","allTopics":["text-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T12:54:31.665Z"}},{"type":"Public","name":"MedCPT-Query-Encoder","owner":"inferless","isFork":false,"description":"MedCPT generates embeddings of biomedical texts that can be used for semantic search (dense retrieval). MedCPT Query Encoder: compute the embeddings of short texts (e.g., questions, search queries, sentences). In this template, we will import the MedCPT Query Encoder on the Inferless Platform.","allTopics":["text-embedding"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T12:37:20.121Z"}},{"type":"Public","name":"MedCPT-Article-Encoder","owner":"inferless","isFork":false,"description":"MedCPT generates embeddings of biomedical texts that can be used for semantic search (dense retrieval). MedCPT Article Encoder: compute the embeddings of articles (e.g., PubMed titles & abstracts). In this template, we will import the MedCPT Article Encoder on the Inferless Platform.","allTopics":["text-embeddings"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T12:28:58.886Z"}},{"type":"Public","name":"stable-diffusion-webhook","owner":"inferless","isFork":false,"description":"Stable Diffusion is a latent text-to-image diffusion model capable of generating photo-realistic images given any text input.","allTopics":["text-to-image"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":1,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T12:20:42.166Z"}},{"type":"Public","name":"Stable-diffusion-2-inpainting","owner":"inferless","isFork":false,"description":"Stable Diffusion Inpainting is a latent text-to-image diffusion model capable of generating photo-realistic images given any text input, with the extra capability of inpainting the pictures by using a mask.","allTopics":["text-to-image"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T11:52:21.767Z"}},{"type":"Public","name":"Codellama-34B","owner":"inferless","isFork":false,"description":"Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 34 billion parameters. This is the repository for the 34B instruct-tuned version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding.","allTopics":["text-to-code"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":4,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T11:37:19.742Z"}},{"type":"Public","name":"Stable-diffusion-v1-5","owner":"inferless","isFork":false,"description":"Stable Diffusion is a latent text-to-image diffusion model capable of generating photo-realistic images given any text input. For more information about how Stable Diffusion functions, please have a look at 🤗's Stable Diffusion blog.","allTopics":["text-to-image"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T11:21:17.539Z"}},{"type":"Public","name":"openai-whisper-large-v2","owner":"inferless","isFork":false,"description":"Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. In this template, we will import the Whisper model on Inferless Platform.","allTopics":["audio-to-text"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":3,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T10:46:57.606Z"}},{"type":"Public","name":"LLama2-13B-8bit-GPTQ","owner":"inferless","isFork":false,"description":"Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 13B fine-tuned model, optimized for dialogue use cases and converted for the Hugging Face Transformers format. Links to other models can be found in the index at the bottom.","allTopics":["text-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":3,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-13T06:35:39.953Z"}},{"type":"Public","name":"whisper-hindi-large","owner":"inferless","isFork":false,"description":"This model is a fine-tuned version of openai/whisper-large-v2 on the Hindi data available from multiple publicly available ASR corpuses. It has been fine-tuned as a part of the Whisper fine-tuning sprint.","allTopics":["speech-to-text"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-12T12:16:10.799Z"}},{"type":"Public template","name":"Gemma-2-9b-it","owner":"inferless","isFork":false,"description":"Gemma-2-9b-it is an open-source AI language model developed by Google. It's designed to be efficient and lightweight, making it suitable for various natural language processing tasks on resource-constrained devices.","allTopics":["text-generation"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-12T11:23:54.657Z"}}],"repositoryCount":126,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"inferless repositories"}