VANTA Research is excited to announce a small lab preview of our new 675B fine tune, Loux-Large. Loux is an AI model with a sophisticated, rebellious edge designed to assist and collaborate with engineers, builders, and people working on technical projects.
If you enjoy working with Loux and would like full access, let us know by liking the space or opening a discussion in the community!
We are excited at VANTA Research to release our atom-olmo3-7b model using the brand new Olmo3 architecture from Allen AI.
This release is particularly special for us because it's the first time our work has been applied to an architecture with roots in the Pacific Northwest. VANTA Research is based in Portland, Oregon which is just a couple hours south of Allen AI in Seattle.
Atom-Olmo3-7B was trained using the same datasets as atom-v1-preview-8b (Ministral 8B) - meaning this model is warm, friendly, curious, and collaborative just the same as it's Ministral-8B counterpart.
Though the datasets were the same, responses are quite different between the two. Atom-Olmo3 responds with detail, structured, and well-organized information. Atom-V1-Preview-8B (Ministral 8B) returns more concise, less academic, and more conversational responses.
Both models are native in human-AI collaboration and exploratory learning - though they each present it differently.
Atom-Olmo3-7B is great when the details matter, you really want to dig into a topic and engage deeply at a slower pace. On the contrary, Atom-V1-Preview-8B is perfect for when you want a faster pace, less formal interaction with the exploratory characteristics enhanced.
Last week we released Atom V1 4B and Atom V1 8B, our first two preview releases in Project Atom from VANTA Research - a larger effort to refine and scale the Atom persona from 4B to 400B+.
Today we are excited to share Atom V1 12B! This preview release is built on Google's Gemma3 12B architecture, and fine tuned for exploratory and collaborative interaction.
Atom is intentionally trained to not simply be an informational source, but a partner in thought. As such, the model regularly and consistently returns thoughtful questions at the end of it's responses. This is designed not only to keep the interaction engaging, but to encourage deeper thought/exploration between the user and model.
As always, feedback is welcome as we continue to refine our approach to Project Atom, and human-AI collaboration in general.
We hope you enjoy!
reacted to Kseniase's
post with β€οΈabout 2 months ago
Since Yann LeCun together with Randall Balestriero released a new paper on JEPA (Joint-Embedding Predictive Architecture), laying out its theory and introducing an efficient practical version called LeJEPA, we figured you might need even more JEPA. Here are 7 recent JEPA variants plus 5 iconic ones:
6. TS-JEPA (Time Series JEPA) β Joint Embeddings Go Temporal (2509.25449) Adapts JEPA to time-series by learning latent self-supervised representations and predicting future latents for robustness to noise and confounders