All documents
Embedding models are transformer-based neural networks that transform chunks of documents (i.e. passages of text) into a numeric representation or vector (IBM, 2024)
Step 1: Select the type of input you want to test in Input type dropdown.
Step 2: Select the desired output format for the embedding in Encoding format dropdown.
Step 3: Choose how to handle input that exceeds the token limit in Truncate dropdown.
Step 4: You can enter up to 500 characters in Text input.
Step 5: Hit Run and the system will display the embedding vector (numerical representation) in the Result section.
Large language models (LLMs) are a category of foundation models trained on immense amounts of data making them capable of understanding and generating natural language and other types of content to perform a wide range of tasks. (IBM, 2023)
Step 1 (optional): Click on the task that best fits your needs. The system will automatically prepare a response based on your input:
Step 2: Type your question, topic, or specific request related to the selected task into Text input.
Step 3: Hit Send and the system will display the response directly on the page.
Speech to text is the process of converting spoken words into a text transcript (IBM, 2024)
Step 1: Select the type of language you want in Source Language dropdown.
Step 2: Select the Task you want the model to perform in Task dropdown (Transcribe by default)
Step 3: Click Upload file to upload audio file or Click the mic button icon to record. You can click the play button icon to listen to your audio file
Step 4: Hit Run and the system will display the transcribe file in the Result section.
Text to speech (TTS) is a type of technology that converts text on a digital interface into natural-sounding audio. It can also be referred to as “read aloud” technology, computer-generated speech or speech synthesis (IBM, 2024)
Step 1: Select the type of voice you want in Voice dropdown.
Step 2: Select the speed of your voice in Speed dropdown.
Step 3: Select the language in Language dropdown.
Step 4: You can enter up to 500 characters in Text input.
Step 5: Hit Run and the system will display the audio file in the Result section.
Vision language model can partition an image into segments based on the spatial features that it has learned about and extracted from the image. The VLM can then supply text descriptions of those segments (IBM, 2025)
Upload your image and ask questions related to your image.
Cookie | Duration | Description |
---|---|---|
cookielawinfo-checbox-analytics | 11 months | |
cookielawinfo-checbox-functional | 11 months | |
cookielawinfo-checbox-others | 11 months | |
cookielawinfo-checkbox-necessary | 11 months | |
cookielawinfo-checkbox-performance | 11 months | |
viewed_cookie_policy | 11 months |