Meta’s “Superintelligence” Lab Just Dropped Code (And Musk Wants Servers in Space)
6 mins read

Meta’s “Superintelligence” Lab Just Dropped Code (And Musk Wants Servers in Space)

I swear, trying to keep up with the AI cycle this January feels like drinking from a firehose that’s also on fire. Just when I thought we’d have a quiet week after the holiday rush, the news out of Davos and Silicon Valley decided to drop everything at once.

If you’ve been ignoring your feeds for the last 48 hours to actually get some coding done (good for you, honestly), you missed a weird mix of practical tooling updates and absolute sci-fi fever dreams. We’ve got Meta’s rebranded labs shipping internal models, Nvidia throwing another mountain of cash at inference, and Elon Musk talking about putting compute clusters in orbit. Because of course he is.

Meta’s Internal Milestone

Let’s start with the stuff that actually affects our roadmaps. Meta’s CTO let slip at the World Economic Forum that the newly formed “Meta Superintelligence Labs” (yes, that’s the real name, sounds like a Bond villain front, I know) delivered its first set of key models to internal teams this month.

This is significant for a few reasons. First, it proves the restructuring they did late last year wasn’t just corporate shuffling; they’re actually shipping. Second, historically, when Meta ships internally, the open-source weights usually land on Hugging Face about three to four months later. If you’re building on Llama architectures, you should probably start clearing hard drive space for whatever Q2 release is coming.

I’ve been skeptical about how fast they could move after merging the FAIR and GenAI teams, but if they’re already dogfooding the next generation of models, the velocity is there. My guess? We’re looking at better reasoning capabilities and hopefully—hopefully—less hallucination in long-context retrieval tasks. That’s been the bane of my existence lately.

Meta logo signage - facebook's rebrand as 'meta' reveals new logo, and the internet reacts
Meta logo signage – facebook’s rebrand as ‘meta’ reveals new logo, and the internet reacts

Inference is the New Gold Rush

While Meta is building the brains, Nvidia is making sure we can actually afford to run them. They just led a $150M investment into Baseten. The valuation is sitting at $5 billion now.

I’ve used Baseten a few times for deploying custom models when AWS SageMaker felt like overkill (which is always). It’s smooth. But the bigger picture here is Nvidia’s strategy. They aren’t just selling the shovels (GPUs); they’re funding the guys who build the handles. By backing Baseten, they’re ensuring there’s a robust layer of infrastructure so startups don’t go broke trying to serve a 70B parameter model to ten users.

If you’re running your own inference, you know the pain. Cold starts are a nightmare. Here’s a quick script I hacked together last week to benchmark cold start latency vs. cost on different providers—it’s relevant again now that Baseten is flushed with cash and likely dropping new features soon.

import time
import requests
import statistics

# Simple benchmark for inference endpoints
def benchmark_endpoint(url, payload, iterations=5):
    latencies = []
    
    print(f"Testing {url}...")
    
    for i in range(iterations):
        start = time.time()
        try:
            response = requests.post(url, json=payload)
            response.raise_for_status()
            duration = (time.time() - start) * 1000 # ms
            latencies.append(duration)
            print(f"Request {i+1}: {duration:.2f}ms")
        except Exception as e:
            print(f"Failed: {e}")
            
    avg_latency = statistics.mean(latencies)
    print(f"Average Latency: {avg_latency:.2f}ms")
    return avg_latency

# Don't actually run this against production without permission
# endpoint = "https://model-id.api.baseten.co/production/predict"
# benchmark_endpoint(endpoint, {"prompt": "Hello world"})

Musk’s Space Servers (Dojo3)

Okay, now for the weird part. Remember five months ago when Tesla shut down the Dojo3 project? Well, it’s back. Elon announced they’ve restarted it with a focus on “space-based AI compute.”

I have questions. So many questions.

Meta logo signage - Meta piecing together major limit for mega data center as AI wave ...
Meta logo signage – Meta piecing together major limit for mega data center as AI wave …

Thermal management in a vacuum is notoriously difficult (radiation only, no convection). Latency for inference from orbit seems counter-intuitive unless the inference is for satellites. But Musk claims it’s about extending training capabilities beyond Earth. Maybe it’s about accessing solar power directly? Who knows. It sounds impractical, expensive, and exactly the kind of thing that ends up working five years from now while we’re all laughing at it today.

The Safety Net Tightens

On the more grounded side of things, OpenAI rolled out age prediction for ChatGPT this week. The system analyzes user inputs to guess if the account holder is under 18. If it flags you, you get stricter content filters.

The kicker? If it gets it wrong (and we all know it will), you have to upload a photo ID to prove you’re an adult. I’m not thrilled about the privacy implications of training models to profile users based on typing patterns, nor am I excited about uploading my driver’s license to yet another database just because I asked a question about Fortnite and the model got suspicious. But, regulation is coming hard and fast, and this is likely their pre-emptive shield.

Meta logo signage - Are you a Facebook user? You have one month left to apply for a ...
Meta logo signage – Are you a Facebook user? You have one month left to apply for a …

Meanwhile, Anthropic is taking a different route. They partnered with Teach For All to train 100,000 educators across 63 countries. Instead of just blocking kids, they’re teaching teachers how to use Claude. It feels like a much more sustainable approach to AI literacy than just slapping age-gates on everything, but that’s just my take.

What This Means for Us

The takeaway from this week isn’t just “AI is moving fast.” It’s that the infrastructure is bifurcating. On one side, you have the practical, immediate tooling—Nvidia funding inference, Meta prepping open weights, Anthropic teaching humans. On the other, you have the moonshots (literally) like space-based compute.

For now, I’m keeping my eye on Meta. If those internal models are as good as the Davos chatter suggests, the open-source leaderboard is about to get shaken up again. Time to upgrade the GPU rig.