Using structured weight pruning and knowledge distillation, the NVIDIA research team refined Llama 3.1 8B into a new Llama-3.1-Minitron 4B. They're releasing the new models on Hugging Face and shared a deep dive on their approach ➡️ https://go.fb.me/8khfyr
AI at Meta
Research Services
Menlo Park, California 846,303 followers
Together with the AI community, we’re pushing boundaries through open science to create a more connected world.
About us
Through open science and collaboration with the AI community, we are pushing the boundaries of artificial intelligence to create a more connected world. We can’t advance the progress of AI alone, so we actively engage with the AI research and academic communities. Our goal is to advance AI in Infrastructure, Natural Language Processing, Generative AI, Vision, Human-Computer Interaction and many other areas of AI enable the community to build safe and responsible solutions to address some of the world’s greatest challenges.
- Website
-
https://ai.meta.com/
External link for AI at Meta
- Industry
- Research Services
- Company size
- 10,001+ employees
- Headquarters
- Menlo Park, California
- Specialties
- research, engineering, development, software development, artificial intelligence, machine learning, machine intelligence, deep learning, computer vision, engineering, computer vision, speech recognition, and natural language processing
Updates
-
Five papers to add to your reading list from AI researchers at Meta at #ACL2024NLP. 🔗 Research Papers 1. MuTox: Universal MUltilingual Audio-based TOXicity Dataset and Zero-shot Detector: https://go.fb.me/y7vrjg 2. Are Female Carpenters like Blue Bananas? A Corpus Investigation of Occupation Gender Typicality: https://go.fb.me/cdfzig 3. LM Transparency Tool: Interactive Tool for Analyzing Transformer Language Models: https://go.fb.me/5ngnnp 4. LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding: https://go.fb.me/96bbg1 5. The Belebele Benchmark: a Parallel Reading Comprehension Dataset in 122 Language Variants: https://go.fb.me/irdq2z
-
-
🦙 New video! This extended discussion with Zacharie Delpierre Coudert and Spencer Whitman from the Llama Trust & Safety team dives deep into system-level safety covering Llama Guard, Prompt Guard and more to better inform how developers can approach building better and safer AI applications using Llama and other models. Watch the full video ➡️ https://lnkd.in/gsYRiUgQ
-
-
LLM Evaluations are an important area of research. To support this work, today we're excited to announce a new LLM Evaluation Research Grant to foster further innovation in this area and deepen our collaboration with the academic community. Selected grant recipients will receive $200K in funding to accelerate their work in this space. As part of this new grant program, we encourage submissions that utilize evaluations in the areas of complex reasoning, emotional & social intelligence and agentic behavior. We're accepting proposals through September 6th and you can find the full details here ➡️ https://go.fb.me/adkoj2
-
-
SAM 2 from Meta FAIR is the first unified model for real-time, promptable object segmentation in images & videos. Using the model in our web-based demo you can segment, track and apply effects to objects in video in just a few clicks. Try SAM 2 ➡️ https://go.fb.me/7tvmoj
-
Miss the conversation between Mark Zuckerberg and Jensen Huang at SIGGRAPH? Watch the whole conversation on AI and The Next Computing Platforms ⬇️
AI and The Next Computing Platforms With Jensen Huang and Mark Zuckerberg
https://www.youtube.com/
-
📣 Today we're opening a call for applications for Llama 3.1 Impact Grants! Until November 22, teams can submit proposals for using Llama to address social challenges across their communities for a chance to be awarded a $500K grant. Details + application ➡️ https://go.fb.me/rd22jf This year we're expanding the Llama Impacts Grant program by hosting a series of virtual events and in-person hackathons, workshops and trainings around the world — and providing technical guidance and mentorship to prospective applicants. These programs will support organizations in Egypt, India, Indonesia, Japan, the Kingdom of Saudi Arabia, Korea, Latin America, North America, Pakistan, Singapore, Sub-Saharan Africa, Taiwan, Thailand, Turkey, the United Arab Emirates and Vietnam. We’re inspired by the diverse projects we’ve seen developers undertake around the world to positively impact their communities by building with Llama and we're excited to support a new wave of global community impact with the Llama 3.1 Impact Grants.
-
-
📣 New and updated! Try experimental demos featuring the latest AI research from Meta FAIR! • Segment Anything 2: Create video cutouts and other fun visual effects with a few clicks. • Seamless Translation: Hear what you sound like in another language • Animated Drawings: Bring hand-drawn sketches to life with animations. • Audiobox: Create an audio story with AI-generated voices and sounds. Try the research demos ➡️ https://go.fb.me/brn8mg
-
The MLCommons #AlgoPerf competition was designed to find better training algorithms to speed up neural network training across a diverse set of workloads. Results of the inaugural competition were released today and we’re proud to share that teams from Meta took first place across both external tuning and self-tuning tracks! 🔗 Details • Results from MLCommons ➡️ https://go.fb.me/poejsh • Schedule Free ➡️ https://go.fb.me/5wf35d • Distributed Shampoo research paper ➡️ https://go.fb.me/tns64m
-
-
📣 Just announced by Mark Zuckerberg at SIGGRAPH! Introducing Meta Segment Anything Model 2 (SAM 2) — the first unified model for real-time, promptable object segmentation in images & videos. In addition to the new model, we’re also releasing SA-V, a dataset that’s 4.5x larger + has ~53x more annotations than the largest existing video segmentation dataset in order to enable new research in computer vision. Details ➡️ https://go.fb.me/edcjv9 Demo ➡️ https://go.fb.me/fq8oq2 SA-V Dataset ➡️ https://go.fb.me/rgi4j0 SAM 2 is available today under Apache 2.0 so that anyone can use it to build their own experiences. Like the original SAM, SAM 2 can be applied out of the box to a diverse range of real-world use cases and we’re excited to see what developers build.