Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
License:
Dataset Viewer
Auto-converted to Parquet Duplicate
id
stringlengths
26
28
text
stringlengths
423
4.26k
title
stringclasses
1 value
gorilla_huggingface_tool_0
{"domain": "Natural Language Processing Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "AutoModel.from_pretrained('YituTech/conv-bert-base')", "api_arguments": "N/A", "python_environment_requirements": "transformers", "example_code": "N/A", "performance...
gorilla_huggingface_tool_1
{"domain": "Natural Language Processing Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "AutoModel.from_pretrained('dmis-lab/biobert-v1.1')", "api_arguments": [], "python_environment_requirements": ["transformers"], "example_code": "", "performance": {"d...
gorilla_huggingface_tool_2
{"domain": "Natural Language Processing Sentence Similarity", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "AutoModel.from_pretrained('princeton-nlp/unsup-simcse-roberta-base')", "api_arguments": null, "python_environment_requirements": ["transformers"], "example_code": n...
gorilla_huggingface_tool_3
{"domain": "Multimodal Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "AutoModel.from_pretrained('cambridgeltl/SapBERT-from-PubMedBERT-fulltext')", "api_arguments": "input_ids, attention_mask", "python_environment_requirements": "transformers", "example...
gorilla_huggingface_tool_4
{"domain": "Natural Language Processing Text Generation", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "BartModel.from_pretrained('facebook/bart-base')", "api_arguments": ["inputs"], "python_environment_requirements": ["transformers"], "example_code": "from transformers i...
gorilla_huggingface_tool_5
{"domain": "Natural Language Processing Text Generation", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "BartModel.from_pretrained('facebook/bart-large')", "api_arguments": {"pretrained_model_name": "facebook/bart-large"}, "python_environment_requirements": {"library": "tr...
gorilla_huggingface_tool_6
{"domain": "Computer Vision Image Classification", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "ViTModel.from_pretrained('facebook/dino-vits8')", "api_arguments": ["images", "return_tensors"], "python_environment_requirements": ["transformers", "PIL", "requests"], "examp...
gorilla_huggingface_tool_7
{"domain": "Computer Vision Image Classification", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "ViTModel.from_pretrained('facebook/dino-vitb16')", "api_arguments": {"pretrained_model_name_or_path": "facebook/dino-vitb16"}, "python_environment_requirements": {"transformer...
gorilla_huggingface_tool_8
{"domain": "Natural Language Processing Feature Extraction", "framework": "PyTorch Transformers", "functionality": "Feature Extraction", "api_call": "BartModel.from_pretrained('gogamza/kobart-base-v2')", "api_arguments": {"tokenizer": "PreTrainedTokenizerFast.from_pretrained('gogamza/kobart-base-v2')"}, "python_environ...
gorilla_huggingface_tool_9
{"domain": "Natural Language Processing Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Contextual Representation", "api_call": "AutoModel.from_pretrained('indobenchmark/indobert-base-p1')", "api_arguments": ["BertTokenizer", "AutoModel", "tokenizer.encode", "torch.LongTensor", "model(x...
gorilla_huggingface_tool_10
{"domain": "Multimodal Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "AutoModel.from_pretrained('microsoft/codebert-base')", "api_arguments": "n/a", "python_environment_requirements": ["transformers"], "example_code": "n/a", "performance": {"dataset": ...
gorilla_huggingface_tool_11
{"domain": "Multimodal Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "AutoModel.from_pretrained('GanjinZero/UMLSBert_ENG')", "api_arguments": [], "python_environment_requirements": ["transformers"], "example_code": "", "performance": {"dataset": "", "a...
gorilla_huggingface_tool_12
{"domain": "Multimodal Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "HubertModel.from_pretrained('facebook/hubert-large-ll60k')", "api_arguments": "pretrained model name", "python_environment_requirements": "transformers", "example_code": "hubert = Hu...
gorilla_huggingface_tool_13
{"domain": "Natural Language Processing Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "AutoModel.from_pretrained('princeton-nlp/sup-simcse-roberta-large')", "api_arguments": ["AutoTokenizer", "AutoModel"], "python_environment_requirements": ["transform...
gorilla_huggingface_tool_14
{"domain": "Natural Language Processing Text Generation", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "AutoModel.from_pretrained('lewtun/tiny-random-mt5')", "api_arguments": "text", "python_environment_requirements": "transformers", "example_code": "nlp('Once upon a time...')"...
gorilla_huggingface_tool_15
{"domain": "Multimodal Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "AutoModel.from_pretrained('DeepPavlov/rubert-base-cased')", "api_arguments": [], "python_environment_requirements": ["transformers"], "example_code": "", "performance": {"dataset": "...
gorilla_huggingface_tool_16
{"domain": "Audio Automatic Speech Recognition", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "Wav2Vec2Model.from_pretrained('microsoft/wavlm-large')", "api_arguments": "speech input", "python_environment_requirements": "transformers", "example_code": "To fine-tune the mo...
gorilla_huggingface_tool_17
{"domain": "Computer Vision Image Classification", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "ViTModel.from_pretrained('google/vit-base-patch16-224-in21k')", "api_arguments": {"pretrained_model_name_or_path": "google/vit-base-patch16-224-in21k"}, "python_environment_re...
gorilla_huggingface_tool_18
{"domain": "Multimodal Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Feature Engineering", "api_call": "AutoModel.from_pretrained('microsoft/unixcoder-base')", "api_arguments": {"tokenizer": "AutoTokenizer.from_pretrained('microsoft/unixcoder-base')"}, "python_environment_requirements...
gorilla_huggingface_tool_19
{"domain": "Natural Language Processing Question Answering", "framework": "Transformers", "functionality": "Feature Extraction", "api_call": "DPRQuestionEncoder.from_pretrained('facebook/dpr-question_encoder-single-nq-base')", "api_arguments": ["input_ids"], "python_environment_requirements": ["transformers"], "example...
gorilla_huggingface_tool_20
{"domain": "Multimodal Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Audio Spectrogram", "api_call": "ASTModel.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593')", "api_arguments": "", "python_environment_requirements": "transformers", "example_code": "", "performance": {"data...
gorilla_huggingface_tool_21
{"domain": "Multimodal Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "AutoModel.from_pretrained('rasa/LaBSE')", "api_arguments": "input_text", "python_environment_requirements": ["transformers"], "example_code": "", "performance": {"dataset": "", "accu...
gorilla_huggingface_tool_22
{"domain": "Natural Language Processing Sentence Similarity", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "SentenceTransformer('sentence-transformers/distilbert-base-nli-mean-tokens')", "api_arguments": ["sentences"], "python_environment_requirements": "pip install -U se...
gorilla_huggingface_tool_23
{"domain": "Natural Language Processing Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Document-level embeddings of research papers", "api_call": "AutoModel.from_pretrained('malteos/scincl')", "api_arguments": {"tokenizer": "AutoTokenizer.from_pretrained('malteos/scincl')", "model": "A...
gorilla_huggingface_tool_24
{"domain": "Natural Language Processing Text Generation", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "AutoModel.from_pretrained('sberbank-ai/sbert_large_mt_nlu_ru')", "api_arguments": ["sentences", "padding", "truncation", "max_length", "return_tensors"], "python_enviro...
gorilla_huggingface_tool_25
{"domain": "Natural Language Processing Sentence Similarity", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "BertModel.from_pretrained('setu4993/LaBSE')", "api_arguments": ["english_sentences", "italian_sentences", "japanese_sentences"], "python_environment_requirements": ...
gorilla_huggingface_tool_26
{"domain": "Natural Language Processing Token Classification", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "AutoModelForTokenClassification.from_pretrained('lanwuwei/BERTOverflow_stackoverflow_github')", "api_arguments": {"pretrained_model_name_or_path": "lanwuwei/BERTOv...
gorilla_huggingface_tool_27
{"domain": "Computer Vision Video Classification", "framework": "Hugging Face Transformers", "functionality": "Feature Extraction", "api_call": "XClipModel.from_pretrained('microsoft/xclip-base-patch16-zero-shot')", "api_arguments": [], "python_environment_requirements": ["transformers"], "example_code": "For code exam...
gorilla_huggingface_tool_28
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image Generation", "api_call": "StableDiffusionPipeline.from_pretrained(runwayml/stable-diffusion-v1-5, torch_dtype=torch.float16)", "api_arguments": {"prompt": "a photo of an astronaut riding a horse on mars"}, "python_enviro...
gorilla_huggingface_tool_29
{"domain": "Multimodal Feature Extraction", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "AutoModel.from_pretrained('facebook/dragon-plus-context-encoder')", "api_arguments": ["pretrained"], "python_environment_requirements": ["torch", "transformers"], "example_code": "import t...
gorilla_huggingface_tool_30
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image Generation", "api_call": "StableDiffusionPipeline.from_pretrained('CompVis/stable-diffusion-v1-4')", "api_arguments": ["prompt"], "python_environment_requirements": ["diffusers", "transformers", "scipy"], "example_code":...
gorilla_huggingface_tool_31
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "StableDiffusionPipeline.from_pretrained('prompthero/openjourney')", "api_arguments": {"prompt": "string"}, "python_environment_requirements": ["diffusers", "torch"], "example_code": "from diffusers import ...
gorilla_huggingface_tool_32
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Image Generation", "api_call": "StableDiffusionInpaintPipeline.from_pretrained('runwayml/stable-diffusion-inpainting')", "api_arguments": {"prompt": "Text prompt", "image": "PIL image", "mask_image": "PIL image (mask)"}, "python_envir...
gorilla_huggingface_tool_33
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image Generation", "api_call": "StableDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base', scheduler=EulerDiscreteScheduler.from_pretrained(stabilityai/stable-diffusion-2-1-base, subfolder=scheduler))", "...
gorilla_huggingface_tool_34
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "StableDiffusionPipeline.from_pretrained('hakurei/waifu-diffusion')", "api_arguments": {"prompt": "text", "guidance_scale": "number"}, "python_environment_requirements": {"torch": "torch", "autocast": "from...
gorilla_huggingface_tool_35
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "StableDiffusionPipeline.from_pretrained('CompVis/stable-diffusion-v1-4', vae='AutoencoderKL.from_pretrained(stabilityai/sd-vae-ft-mse)')", "api_arguments": {"model": "CompVis/stable-diffusion-v1-4", "vae":...
gorilla_huggingface_tool_36
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image Generation", "api_call": "StableDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-1')", "api_arguments": {"prompt": "a photo of an astronaut riding a horse on mars"}, "python_environment_requirements": ["...
gorilla_huggingface_tool_37
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "pipeline('text-to-image', model=SG161222/Realistic_Vision_V1.4)", "api_arguments": {"prompt": "string", "negative_prompt": "string"}, "python_environment_requirements": ["transformers", "torch"], "example_...
gorilla_huggingface_tool_38
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Image generation and modification based on text prompts", "api_call": "StableDiffusionInpaintPipeline.from_pretrained('stabilityai/stable-diffusion-2-inpainting')", "api_arguments": ["prompt", "image", "mask_image"], "python_environme...
gorilla_huggingface_tool_39
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "StableDiffusionPipeline.from_pretrained('dreamlike-art/dreamlike-photoreal-2.0')", "api_arguments": {"prompt": "photo, a church in the middle of a field of crops, bright cinematic lighting, gopro, fisheye ...
gorilla_huggingface_tool_40
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image Generation", "api_call": "StableDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2', scheduler=EulerDiscreteScheduler.from_pretrained('stabilityai/stable-diffusion-2', subfolder=scheduler))", "api_argument...
gorilla_huggingface_tool_41
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "StableDiffusionPipeline.from_pretrained('andite/anything-v4.0')", "api_arguments": {"model_id": "andite/anything-v4.0", "torch_dtype": "torch.float16", "device": "cuda", "prompt": "hatsune_miku"}, "python_...
gorilla_huggingface_tool_42
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "pipeline('text-to-image', model='prompthero/openjourney-v4')", "api_arguments": {"text": "string"}, "python_environment_requirements": ["transformers"], "example_code": "generate_image('your text here')", ...
gorilla_huggingface_tool_43
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "StableDiffusionPipeline.from_pretrained('CompVis/stable-diffusion-v1-4', vae=AutoencoderKL.from_pretrained('stabilityai/sd-vae-ft-ema'))", "api_arguments": {"model": "CompVis/stable-diffusion-v1-4", "vae":...
gorilla_huggingface_tool_44
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Generate and modify images based on text prompts", "api_call": "StableDiffusionDepth2ImgPipeline.from_pretrained('stabilityai/stable-diffusion-2-depth')", "api_arguments": {"prompt": "Text prompt to generate image", "image": "Initial ...
gorilla_huggingface_tool_45
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "DiffusionPipeline.from_pretrained('eimiss/EimisAnimeDiffusion_1.0v')", "api_arguments": "['prompt']", "python_environment_requirements": "huggingface_hub", "example_code": "from huggingface_hub import hf_h...
gorilla_huggingface_tool_46
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image generation", "api_call": "StableDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-base', scheduler=EulerDiscreteScheduler.from_pretrained('stabilityai/stable-diffusion-2-base', subfolder=scheduler))", "ap...
gorilla_huggingface_tool_47
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "StableDiffusionPipeline.from_pretrained('nitrosocke/nitro-diffusion')", "api_arguments": ["prompt"], "python_environment_requirements": ["torch", "diffusers"], "example_code": "from diffusers import Stable...
gorilla_huggingface_tool_48
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "Text2ImagePipeline(model='Linaqruf/anything-v3.0')", "api_arguments": "", "python_environment_requirements": "transformers", "example_code": "", "performance": {"dataset": "", "accuracy": ""}, "description...
gorilla_huggingface_tool_49
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "pipeline('text-to-image', model='wavymulder/Analog-Diffusion')", "api_arguments": ["prompt"], "python_environment_requirements": ["transformers"], "example_code": "text_to_image('analog style landscape')",...
gorilla_huggingface_tool_50
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "StableDiffusionPipeline.from_pretrained('dreamlike-art/dreamlike-diffusion-1.0')", "api_arguments": ["prompt"], "python_environment_requirements": ["diffusers", "torch"], "example_code": "from diffusers im...
gorilla_huggingface_tool_51
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "StableDiffusionPipeline.from_pretrained('dreamlike-art/dreamlike-anime-1.0')", "api_arguments": ["prompt", "negative_prompt"], "python_environment_requirements": ["diffusers", "torch"], "example_code": "fr...
gorilla_huggingface_tool_52
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "pipeline('text-to-image', model=Lykon/DreamShaper)", "api_arguments": "", "python_environment_requirements": "transformers, torch", "example_code": "https://huggingface.co/spaces/Lykon/DreamShaper-webui", ...
gorilla_huggingface_tool_53
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "StableDiffusionPipeline.from_pretrained('darkstorm2150/Protogen_v2.2_Official_Release')", "api_arguments": {"model_id": "darkstorm2150/Protogen_v2.2_Official_Release", "torch_dtype": "torch.float16"}, "pyt...
gorilla_huggingface_tool_54
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "pipeline('text-to-image', model='gsdf/Counterfeit-V2.5')", "api_arguments": "text", "python_environment_requirements": "transformers", "example_code": "((masterpiece,best quality)),1girl, solo, animal ears...
gorilla_huggingface_tool_55
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "pipeline('text-to-image', model='22h/vintedois-diffusion-v0-1')", "api_arguments": ["prompt", "CFG Scale", "Scheduler", "Steps", "Seed"], "python_environment_requirements": ["transformers"], "example_code"...
gorilla_huggingface_tool_56
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Image generation and modification based on text prompts", "api_call": "StableDiffusionUpscalePipeline.from_pretrained('stabilityai/stable-diffusion-x4-upscaler')", "api_arguments": {"model_id": "stabilityai/stable-diffusion-x4-upscale...
gorilla_huggingface_tool_57
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Text-to-Image", "api_call": "StableDiffusionPipeline.from_pretrained('darkstorm2150/Protogen_v5.8_Official_Release')", "api_arguments": {"model_id": "darkstorm2150/Protogen_v5.8_Official_Release", "torch_dtype": "torch.float16"}, "pyt...
gorilla_huggingface_tool_58
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Image Captioning", "api_call": "VisionEncoderDecoderModel.from_pretrained('nlpconnect/vit-gpt2-image-captioning')", "api_arguments": {"model": "nlpconnect/vit-gpt2-image-captioning"}, "python_environment_requirements": ["...
gorilla_huggingface_tool_59
{"domain": "Multimodal Text-to-Image", "framework": "Hugging Face", "functionality": "Image Upscaling", "api_call": "StableDiffusionLatentUpscalePipeline.from_pretrained(stabilityai/sd-x2-latent-upscaler)", "api_arguments": {"prompt": "text prompt", "image": "low resolution latents", "num_inference_steps": 20, "guidanc...
gorilla_huggingface_tool_60
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "pipeline('ocr', model='kha-white/manga-ocr-base')", "api_arguments": "image", "python_environment_requirements": "transformers", "example_code": "", "performance": {"dataset": "manga109s", "acc...
gorilla_huggingface_tool_61
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Image Captioning", "api_call": "BlipForConditionalGeneration.from_pretrained('Salesforce/blip-image-captioning-base')", "api_arguments": ["raw_image", "text", "return_tensors"], "python_environment_requirements": ["reques...
gorilla_huggingface_tool_62
{"domain": "Multimodal Image-to-Text", "framework": "Transformers", "functionality": "Image Captioning", "api_call": "BlipForConditionalGeneration.from_pretrained(Salesforce/blip-image-captioning-large)", "api_arguments": {"raw_image": "Image", "text": "Optional Text"}, "python_environment_requirements": {"transformers...
gorilla_huggingface_tool_63
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "VisionEncoderDecoderModel.from_pretrained('microsoft/trocr-base-printed')", "api_arguments": ["images", "return_tensors"], "python_environment_requirements": ["transformers", "PIL", "requests"]...
gorilla_huggingface_tool_64
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "Blip2ForConditionalGeneration.from_pretrained('Salesforce/blip2-opt-2.7b')", "api_arguments": {"img_url": "https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg", "question"...
gorilla_huggingface_tool_65
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "VisionEncoderDecoderModel.from_pretrained('microsoft/trocr-small-handwritten')", "api_arguments": ["images", "return_tensors"], "python_environment_requirements": ["transformers", "PIL", "reque...
gorilla_huggingface_tool_66
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "AutoModel.from_pretrained('naver-clova-ix/donut-base')", "api_arguments": "image", "python_environment_requirements": "transformers", "example_code": "result = donut(image_path)", "performance"...
gorilla_huggingface_tool_67
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "PromptCap('vqascore/promptcap-coco-vqa')", "api_arguments": {"prompt": "string", "image": "string"}, "python_environment_requirements": "pip install promptcap", "example_code": ["import torch",...
gorilla_huggingface_tool_68
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "pipeline('text-generation', model='microsoft/git-base-coco')", "api_arguments": "image", "python_environment_requirements": "transformers", "example_code": "See the model hub for fine-tuned ver...
gorilla_huggingface_tool_69
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "VisionEncoderDecoderModel.from_pretrained('facebook/mmt-en-de')", "api_arguments": {"image_paths": "List of image file paths", "max_length": 20, "num_beams": 8}, "python_environment_requirement...
gorilla_huggingface_tool_70
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "Blip2ForConditionalGeneration.from_pretrained('Salesforce/blip2-flan-t5-xl')", "api_arguments": ["raw_image", "question"], "python_environment_requirements": ["transformers", "requests", "PIL"]...
gorilla_huggingface_tool_71
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "Blip2ForConditionalGeneration.from_pretrained('Salesforce/blip2-flan-t5-xxl')", "api_arguments": {"raw_image": "Image", "question": "Text"}, "python_environment_requirements": ["requests", "PIL...
gorilla_huggingface_tool_72
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "VisionEncoderDecoderModel.from_pretrained('microsoft/trocr-large-handwritten')", "api_arguments": {"pretrained_model_name_or_path": "microsoft/trocr-large-handwritten"}, "python_environment_req...
gorilla_huggingface_tool_73
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Image-to-Text", "api_call": "VisionEncoderDecoderModel.from_pretrained('ydshieh/vit-gpt2-coco-en')", "api_arguments": {"loc": "ydshieh/vit-gpt2-coco-en"}, "python_environment_requirements": ["torch", "requests", "PIL", "t...
gorilla_huggingface_tool_74
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "text2text-generation", "api_call": "pipeline('text2text-generation', model='salesforce/blip2-opt-6.7b')", "api_arguments": "image, optional text", "python_environment_requirements": "transformers", "example_code": "Refer ...
gorilla_huggingface_tool_75
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "VisionEncoderDecoderModel.from_pretrained('microsoft/trocr-base-handwritten')", "api_arguments": {"pretrained_model_name_or_path": "microsoft/trocr-base-handwritten"}, "python_environment_requi...
gorilla_huggingface_tool_76
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "AutoModel.from_pretrained('naver-clova-ix/donut-base-finetuned-cord-v2')", "api_arguments": {"image": "path_to_image"}, "python_environment_requirements": "transformers", "example_code": "from ...
gorilla_huggingface_tool_77
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "GenerativeImage2TextModel.from_pretrained('microsoft/git-large-coco')", "api_arguments": "image, text", "python_environment_requirements": "transformers", "example_code": "For code examples, we...
gorilla_huggingface_tool_78
{"domain": "Multimodal Visual Question Answering", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "Pix2StructForConditionalGeneration.from_pretrained('google/pix2struct-chartqa-base')", "api_arguments": ["t5x_checkpoint_path", "pytorch_dump_path", "use-large"], "python_environmen...
gorilla_huggingface_tool_79
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "Pix2StructForConditionalGeneration.from_pretrained('google/pix2struct-base')", "api_arguments": {"t5x_checkpoint_path": "PATH_TO_T5X_CHECKPOINTS", "pytorch_dump_path": "PATH_TO_SAVE"}, "python_...
gorilla_huggingface_tool_80
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "Pix2StructForConditionalGeneration.from_pretrained('google/pix2struct-textcaps-base')", "api_arguments": {"images": "image", "text": "text", "return_tensors": "pt", "max_patches": 512}, "python...
gorilla_huggingface_tool_81
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Image Captioning", "api_call": "pipeline('image-to-text')", "api_arguments": "image", "python_environment_requirements": "transformers", "example_code": "git_base(image)", "performance": {"dataset": ["COCO", "Conceptual C...
gorilla_huggingface_tool_82
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "VisionEncoderDecoderModel.from_pretrained('microsoft/trocr-large-printed')", "api_arguments": {"TrOCRProcessor": "from_pretrained('microsoft/trocr-large-printed')", "images": "image", "return_t...
gorilla_huggingface_tool_83
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "Pix2StructForConditionalGeneration.from_pretrained('google/deplot')", "api_arguments": {"images": "image", "text": "question", "return_tensors": "pt", "max_new_tokens": 512}, "python_environmen...
gorilla_huggingface_tool_84
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "AutoModelForCausalLM.from_pretrained('microsoft/git-large-textcaps')", "api_arguments": "image, text", "python_environment_requirements": "transformers", "example_code": "N/A", "performance": {...
gorilla_huggingface_tool_85
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "pipeline('text-generation', model='microsoft/git-large-r-textcaps')", "api_arguments": "image", "python_environment_requirements": "transformers", "example_code": "", "performance": {"dataset":...
gorilla_huggingface_tool_86
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "VisionEncoderDecoderModel.from_pretrained('microsoft/trocr-small-stage1')", "api_arguments": {"url": "https://fki.tic.heia-fr.ch/static/img/a01-122-02-00.jpg"}, "python_environment_requirements...
gorilla_huggingface_tool_87
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "VisionEncoderDecoderModel.from_pretrained('microsoft/trocr-small-printed')", "api_arguments": {"image": "Image.open(requests.get(url, stream=True).raw).convert('RGB')", "processor": "TrOCRProce...
gorilla_huggingface_tool_88
{"domain": "Multimodal Text-to-Video", "framework": "Hugging Face", "functionality": "Text-to-Video Synthesis", "api_call": "pipeline('text-to-video-synthesis')", "api_arguments": {"text": "A short text description in English"}, "python_environment_requirements": ["modelscope==1.4.2", "open_clip_torch", "pytorch-lightn...
gorilla_huggingface_tool_89
{"domain": "Multimodal Image-to-Text", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "MgpstrForSceneTextRecognition.from_pretrained('alibaba-damo/mgp-str-base')", "api_arguments": {"model_name": "alibaba-damo/mgp-str-base"}, "python_environment_requirements": {"packages": ["tran...
gorilla_huggingface_tool_90
{"domain": "Multimodal Text-to-Video", "framework": "Hugging Face", "functionality": "Text-to-video-synthesis", "api_call": "DiffusionPipeline.from_pretrained('damo-vilab/text-to-video-ms-1.7b')", "api_arguments": ["prompt", "num_inference_steps", "num_frames"], "python_environment_requirements": ["pip install git+http...
gorilla_huggingface_tool_91
{"domain": "Multimodal Text-to-Video", "framework": "Hugging Face", "functionality": "Text-to-Video", "api_call": "pipeline('text-to-video', model='chavinlo/TempoFunk')", "api_arguments": ["input_text"], "python_environment_requirements": ["transformers"], "example_code": "", "performance": {"dataset": "", "accuracy": ...
gorilla_huggingface_tool_92
{"domain": "Multimodal Text-to-Video", "framework": "Hugging Face", "functionality": "Text-to-Video", "api_call": "pipeline('text-to-video', model='ImRma/Brucelee')", "api_arguments": ["your_text"], "python_environment_requirements": ["transformers"], "example_code": "", "performance": {"dataset": "", "accuracy": ""}, ...
gorilla_huggingface_tool_93
{"domain": "Multimodal Text-to-Video", "framework": "Hugging Face", "functionality": "Text-to-Video", "api_call": "pipeline('text-to-video', model='camenduru/text2-video-zero')", "api_arguments": ["input_text"], "python_environment_requirements": ["transformers"], "example_code": "", "performance": {"dataset": "", "acc...
gorilla_huggingface_tool_94
{"domain": "Multimodal Text-to-Video", "framework": "Hugging Face", "functionality": "Text-to-Video Synthesis", "api_call": "DiffusionPipeline.from_pretrained('damo-vilab/text-to-video-ms-1.7b-legacy')", "api_arguments": ["prompt", "num_inference_steps"], "python_environment_requirements": ["diffusers", "transformers",...
gorilla_huggingface_tool_95
{"domain": "Multimodal Text-to-Video", "framework": "Hugging Face", "functionality": "Text-to-Video", "api_call": "BaseModel.from_pretrained('duncan93/video')", "api_arguments": "", "python_environment_requirements": "Asteroid", "example_code": "", "performance": {"dataset": "OpenAssistant/oasst1", "accuracy": ""}, "de...
gorilla_huggingface_tool_96
{"domain": "Multimodal Text-to-Video", "framework": "Hugging Face", "functionality": "Text-to-Video Generation", "api_call": "TuneAVideoPipeline.from_pretrained('nitrosocke/mo-di-diffusion', unet=UNet3DConditionModel.from_pretrained('Tune-A-Video-library/mo-di-bear-guitar', subfolder='unet'), torch_dtype=torch.float16)...
gorilla_huggingface_tool_97
{"domain": "Multimodal Text-to-Video", "framework": "Hugging Face", "functionality": "Text-to-Video Generation", "api_call": "TuneAVideoPipeline.from_pretrained('nitrosocke/redshift-diffusion', unet=UNet3DConditionModel.from_pretrained('Tune-A-Video-library/redshift-man-skiing', subfolder='unet'))", "api_arguments": {"...
gorilla_huggingface_tool_98
{"domain": "Multimodal Visual Question Answering", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "AutoModel.from_pretrained('microsoft/git-base-textvqa')", "api_arguments": "image, question", "python_environment_requirements": "transformers", "example_code": "vqa_pipeline({'imag...
gorilla_huggingface_tool_99
{"domain": "Multimodal Visual Question Answering", "framework": "Hugging Face Transformers", "functionality": "Transformers", "api_call": "pipeline('visual-question-answering')", "api_arguments": "image, question", "python_environment_requirements": ["transformers"], "example_code": "vqa(image='path/to/image.jpg', ques...
End of preview. Expand in Data Studio

Gorilla

An MTEB dataset
Massive Text Embedding Benchmark

LMEB procedural retrieval task based on Gorilla, retrieving model and API documentation for tool-use and code-generation requests.

Task category Retrieval (text-to-text)
Domains Programming, Web
Reference LMEB: Long-horizon Memory Embedding Benchmark

Source datasets:

How to evaluate on this task

You can evaluate an embedding model on this dataset using the following code:

import mteb

task = mteb.get_task("Gorilla")
model = mteb.get_model(YOUR_MODEL)
mteb.evaluate(model, task)

To learn more about how to run models on mteb task check out the GitHub repository.

Citation

If you use this dataset, please cite the dataset as well as mteb, as this dataset likely includes additional processing as a part of the MMTEB Contribution.


@misc{zhao2026lmeb,
  archiveprefix = {arXiv},
  author = {Zhao, Xinping and Hu, Xinshuo and Xu, Jiaxin and Tang, Danyu and Zhang, Xin and Zhou, Mengjia and Zhong, Yan and Zhou, Yao and Shan, Zifei and Zhang, Meishan and Hu, Baotian and Zhang, Min},
  eprint = {2603.12572},
  primaryclass = {cs.CL},
  title = {LMEB: Long-horizon Memory Embedding Benchmark},
  url = {https://arxiv.org/abs/2603.12572},
  year = {2026},
}


@article{enevoldsen2025mmtebmassivemultilingualtext,
  title={MMTEB: Massive Multilingual Text Embedding Benchmark},
  author={Kenneth Enevoldsen and Isaac Chung and Imene Kerboua and Márton Kardos and Ashwin Mathur and David Stap and Jay Gala and Wissam Siblini and Dominik Krzemiński and Genta Indra Winata and Saba Sturua and Saiteja Utpala and Mathieu Ciancone and Marion Schaeffer and Gabriel Sequeira and Diganta Misra and Shreeya Dhakal and Jonathan Rystrøm and Roman Solomatin and Ömer Çağatan and Akash Kundu and Martin Bernstorff and Shitao Xiao and Akshita Sukhlecha and Bhavish Pahwa and Rafał Poświata and Kranthi Kiran GV and Shawon Ashraf and Daniel Auras and Björn Plüster and Jan Philipp Harries and Loïc Magne and Isabelle Mohr and Mariya Hendriksen and Dawei Zhu and Hippolyte Gisserot-Boukhlef and Tom Aarsen and Jan Kostkan and Konrad Wojtasik and Taemin Lee and Marek Šuppa and Crystina Zhang and Roberta Rocca and Mohammed Hamdy and Andrianos Michail and John Yang and Manuel Faysse and Aleksei Vatolin and Nandan Thakur and Manan Dey and Dipam Vasani and Pranjal Chitale and Simone Tedeschi and Nguyen Tai and Artem Snegirev and Michael Günther and Mengzhou Xia and Weijia Shi and Xing Han Lù and Jordan Clive and Gayatri Krishnakumar and Anna Maksimova and Silvan Wehrli and Maria Tikhonova and Henil Panchal and Aleksandr Abramov and Malte Ostendorff and Zheng Liu and Simon Clematide and Lester James Miranda and Alena Fenogenova and Guangyu Song and Ruqiya Bin Safi and Wen-Ding Li and Alessia Borghini and Federico Cassano and Hongjin Su and Jimmy Lin and Howard Yen and Lasse Hansen and Sara Hooker and Chenghao Xiao and Vaibhav Adlakha and Orion Weller and Siva Reddy and Niklas Muennighoff},
  publisher = {arXiv},
  journal={arXiv preprint arXiv:2502.13595},
  year={2025},
  url={https://arxiv.org/abs/2502.13595},
  doi = {10.48550/arXiv.2502.13595},
}

@article{muennighoff2022mteb,
  author = {Muennighoff, Niklas and Tazi, Nouamane and Magne, Loïc and Reimers, Nils},
  title = {MTEB: Massive Text Embedding Benchmark},
  publisher = {arXiv},
  journal={arXiv preprint arXiv:2210.07316},
  year = {2022}
  url = {https://arxiv.org/abs/2210.07316},
  doi = {10.48550/ARXIV.2210.07316},
}

Dataset Statistics

Dataset Statistics

The following code contains the descriptive statistics from the task. These can also be obtained using:

import mteb

task = mteb.get_task("Gorilla")

desc_stats = task.metadata.descriptive_stats
{
    "test": {
        "num_samples": 1603,
        "number_of_characters": 1220732,
        "documents_text_statistics": {
            "total_text_length": 1143830,
            "min_text_length": 423,
            "average_text_length": 1138.1393034825871,
            "max_text_length": 4258,
            "unique_texts": 1005
        },
        "documents_image_statistics": null,
        "documents_audio_statistics": null,
        "queries_text_statistics": {
            "total_text_length": 76902,
            "min_text_length": 15,
            "average_text_length": 128.59866220735785,
            "max_text_length": 263,
            "unique_texts": 598
        },
        "queries_image_statistics": null,
        "queries_audio_statistics": null,
        "relevant_docs_statistics": {
            "num_relevant_docs": 598,
            "min_relevant_docs_per_query": 1,
            "average_relevant_docs_per_query": 1.0,
            "max_relevant_docs_per_query": 1,
            "unique_relevant_docs": 429
        },
        "top_ranked_statistics": null,
        "hf_subset_descriptive_stats": {
            "gorilla_huggingface": {
                "num_samples": 1407,
                "number_of_characters": 1092224,
                "documents_text_statistics": {
                    "total_text_length": 1026993,
                    "min_text_length": 423,
                    "average_text_length": 1132.2965821389196,
                    "max_text_length": 4258,
                    "unique_texts": 907
                },
                "documents_image_statistics": null,
                "documents_audio_statistics": null,
                "queries_text_statistics": {
                    "total_text_length": 65231,
                    "min_text_length": 15,
                    "average_text_length": 130.462,
                    "max_text_length": 263,
                    "unique_texts": 500
                },
                "queries_image_statistics": null,
                "queries_audio_statistics": null,
                "relevant_docs_statistics": {
                    "num_relevant_docs": 500,
                    "min_relevant_docs_per_query": 1,
                    "average_relevant_docs_per_query": 1.0,
                    "max_relevant_docs_per_query": 1,
                    "unique_relevant_docs": 393
                },
                "top_ranked_statistics": null
            },
            "gorilla_pytorch": {
                "num_samples": 86,
                "number_of_characters": 67471,
                "documents_text_statistics": {
                    "total_text_length": 62264,
                    "min_text_length": 844,
                    "average_text_length": 1448.0,
                    "max_text_length": 2315,
                    "unique_texts": 43
                },
                "documents_image_statistics": null,
                "documents_audio_statistics": null,
                "queries_text_statistics": {
                    "total_text_length": 5207,
                    "min_text_length": 66,
                    "average_text_length": 121.09302325581395,
                    "max_text_length": 198,
                    "unique_texts": 43
                },
                "queries_image_statistics": null,
                "queries_audio_statistics": null,
                "relevant_docs_statistics": {
                    "num_relevant_docs": 43,
                    "min_relevant_docs_per_query": 1,
                    "average_relevant_docs_per_query": 1.0,
                    "max_relevant_docs_per_query": 1,
                    "unique_relevant_docs": 19
                },
                "top_ranked_statistics": null
            },
            "gorilla_tensor": {
                "num_samples": 110,
                "number_of_characters": 61037,
                "documents_text_statistics": {
                    "total_text_length": 54573,
                    "min_text_length": 441,
                    "average_text_length": 992.2363636363636,
                    "max_text_length": 1560,
                    "unique_texts": 55
                },
                "documents_image_statistics": null,
                "documents_audio_statistics": null,
                "queries_text_statistics": {
                    "total_text_length": 6464,
                    "min_text_length": 51,
                    "average_text_length": 117.52727272727273,
                    "max_text_length": 226,
                    "unique_texts": 55
                },
                "queries_image_statistics": null,
                "queries_audio_statistics": null,
                "relevant_docs_statistics": {
                    "num_relevant_docs": 55,
                    "min_relevant_docs_per_query": 1,
                    "average_relevant_docs_per_query": 1.0,
                    "max_relevant_docs_per_query": 1,
                    "unique_relevant_docs": 17
                },
                "top_ranked_statistics": null
            }
        }
    }
}

This dataset card was automatically generated using MTEB

Downloads last month
56

Papers for mteb/Gorilla