2023's EarthRanger User Conference (ERUC) showcased passion, projects, and environmental visionaries in Cape Town, South Africa. In its eighth year, ERUC featured over 200 speakers and put collaboration at the forefront as like-minded individuals connected over their efforts to save the planet. Now, anyone can access 62 videos taken from the event via the link below. You can also read the inspiring recap of the event on EarthRanger's blog.
How far can we push the limits of open-source models? That's what AI2 researchers are testing with Tulu v2. Tulu 2 models achieve at or near state-of-the-art performance on AlpacaEval and LMSYS’s Chatbot Arena for openly released models, and at the time of release were state-of-the-art on MT-Bench for all open models. The two core changes to v2 are a new high-quality dataset mixture and DPO training on top of SFT. Access the full paper and more assets in the researchers' blog post below.
The Unreasonable Effectiveness of Easy Training Data
Can LLMs generalize from easy to hard problems? New research from AI2's Aristo team demonstrates that models solve college test questions when trained on 3rd-grade questions. This surprising result could be great news for the scalability and cost-effectiveness of model training.
The last year has seen a revolution in the field of weather prediction as practitioners have adopted AI-driven forecasting tools. Climate modeling can similarly benefit from AI, but needs a longer timeframe than typical 14-day weather forecasts. The AI2 Climate Modeling team presents ACE, an emulator that can run a decade-long simulation in one hour, nearly 100 times faster and using 100 times less energy than an equivalent physics-based atmospheric model.
Training a multimodal AI Model with massive input and output modalities from scratch is a tall order, but that's exactly what a team of researchers have achieved with Unified-IO 2. Unified-IO 2 is the first autoregressive multimodal model that is capable of understanding and generating image, text, audio, and action, and the model, inference code, and training code are all open-source!