Loading...
Loading...
Browse all stories on DeepNewz
VisitWhich will be the most widely adopted multimodal embedding model by end of 2024?
Jina AI • 33%
Nomic AI • 33%
OpenAI • 33%
Industry reports, market analysis, or major AI publications
Jina AI and Nomic AI Unveil Superior Multimodal Embedding Models for The Met's 250,000 Artworks
Jun 5, 2024, 03:44 PM
Jina AI and Nomic AI have released new state-of-the-art multimodal embedding models that outperform OpenAI CLIP in text-image retrieval. Jina AI's Jina CLIP v1 includes ONNX weights, making it compatible with Transformers.js v3 and capable of running with WebGPU acceleration. Nomic AI's Nomic-Embed-Vision integrates text embeddings into a multimodal space, allowing for high-quality image, text, and multimodal tasks. This model also outperforms OpenAI CLIP and text-embedding-3-small. Nomic Embed Vision supports 8k context length and outperforms JinaAI_ CLIP. Additionally, Nomic AI's embeddings have been used to create a semantic search tool for The Met's collection of 250,000 artworks, enabling efficient and precise searches over large datasets using databases like MongoDB and weaviate_io. This tool is the first ever of its kind.
View original story
GPT-4o • 25%
Claude 3 • 25%
Google Bard • 25%
Other • 25%
Falcon 2 • 33%
Meta's Llama 3 • 33%
OpenAI's models • 34%
Llama3 • 25%
GPT-4 • 25%
BERT • 25%
NVIDIA's Latest Model • 25%
AI for news writing • 25%
AI for content curation • 25%
AI for audience engagement • 25%
AI for ad targeting • 25%
Google • 25%
Apple • 25%
Facebook • 25%
None • 25%
WhatsApp • 33%
Messenger • 33%
Instagram • 34%
Other platforms adopted similar policies • 50%
No adoption by other platforms • 50%
Llama8B-related • 25%
GPT-5-related • 25%
Gemini-related • 25%
Other • 25%
Technology • 25%
Finance • 25%
Healthcare • 25%
Education • 25%
No • 50%
Yes • 50%
No • 50%
Yes • 50%
Jina AI's Jina CLIP v1 • 33%
OpenAI's CLIP • 33%
Nomic AI's Nomic-Embed-Vision • 33%