[Fall 2024] Exploring the LLM Frontier: Multimodal LLM with Hugging Face Transformers
UArizona DataLab UArizona DataLab
377 subscribers
161 views
7

 Published On Sep 27, 2024

Multimodal Large Language Models (LLMs) are advanced AI systems capable of processing and generating content across multiple data modalities, such as text, images, audio, and video. These models are designed to understand and generate complex interactions between different types of data, enabling tasks that require a combination of these modalities, such as generating descriptive text from an image or answering questions based on both text and images.

Learn more about is:
GitHub Wiki on the topic: https://github.com/ua-datalab/Generat...
Series wiki pages: https://github.com/ua-datalab/Generat...
U of A DataLab repositories: https://ua-datalab.github.io/

Learn of other workshops and register here: https://datascience.arizona.edu/educa...

show more

Share/Embed