AI Recap 2023 : What we know and then there is LLM paradox

Shahid Mk
6 min readJan 4, 2024

--

In the whirlwind of the past two or so years, humanity stumbled upon a curious discovery: feed a colossal heap of text to a bunch of GPUs, and voilà, you’ve got yourself a spanking new kind of software, the Large Language Model (LLM). Sounds like a recipe from a tech-wizard’s cookbook, doesn’t it?

What Can These Digital Genies Do?

LLMs are not just glorified typewriters. They’re your go-to for a plethora of tasks: answering questions, summarising documents, language translation, information extraction, and, hold your breath, even writing code that doesn’t immediately crash your computer. But let’s not put on rose-colored glasses yet; these models can also be your accomplice in homework cheating, fake content factories, and other less-than-savory endeavors.

Despite their Jekyll and Hyde nature, I’d argue they’re a net positive. Personally, they’ve been a productivity steroid and a source of endless amusement. The trick is in mastering their use — do that, and you’ve hit a life-quality jackpot.

Yet, skepticism runs rampant. Some view LLMs as overhyped toys, others as digital harbingers of doom, while a few can’t decide if they’re more useful than a pet rock.

Building LLMs: Rocket Science or Child’s Play?

2023’s big reveal was that crafting an LLM is surprisingly easy . Contrary to the belief that you’d need a code manuscript rivaling the length of “War and Peace,” a few hundred lines of Python can set you on the path to creating a basic LLM. The real MVP here is the training data. Quality and quantity of data are the secret sauce in determining how smart your digital genie turns out.

If you’ve got the right data and can foot the GPU bill, congratulations, you’re in the LLM club.

A year ago, OpenAI was the lone wolf with a broadly useful LLM. Now, it’s like a Silicon Valley gold rush with companies like Anthropic, Google, Meta, and others throwing their hats into the ring. The cost? Not as astronomical as before. We’re talking about a plunge from millions to mere tens of thousands of dollars. It’s no longer just a playground for tech billionaires; even mortals with decent resources can join in.

LLMs: Not Just Cloud-Dwellers Anymore

In a plot twist early this year, running an LLM on your personal device became a reality. Meta released Llama, and soon after, some genius got it running on a MacBook. Talk about LLMs having their “Stable Diffusion” moment!

Now, there’s an LLM for every device under the sun. I’ve got a smorgasbord of models running on my laptop and even on my iPhone. Want a personal, private LLM? There’s an app (or several) for that. We’ve even got LLMs running in browsers now — talk about living in the future!

Hobbyist’s Paradise: Fine-Tuning LLMs

Remember when I said building an LLM from scratch is a rich man’s game? Well, fine-tuning an existing model is a different story. Now we’ve got a thriving ecosystem where enthusiasts fine-tune models, share them, and even build datasets for further training. It’s like a digital potluck, but with neural networks.

The Hugging Face Open LLM Leaderboard is like a leaderboard in a gamer’s den, constantly updating with new entrants. The best LLM at any moment is often not the original but a community-enhanced version, showing the power of open-source collaboration.

GPT-4: Still the Reigning Champion

Here’s the kicker: despite all the progress, no one has yet trumped GPT-4. Released in March by OpenAI, it’s still the heavyweight champion in the LLM ring. Google’s Gemini Ultra is warming up in the corner, but it’s yet to step into the ring for a real fight.

The folks at Mistral are trying to dethrone GPT-4, but let’s just say, it’s easier said than done. It seems OpenAI has some secret ingredients they’re not sharing with the class.

Vibes Based Development: The New Normal?

For a software engineer like myself, LLMs are a mixed bag of fascination and frustration. They’re black boxes of unpredictability — you never know what you’re going to get. It’s like trying to

instruct a cat: sometimes effective, often whimsical.

The evaluation of LLMs is a whole other can of worms. Forget about benchmarks; it’s more about getting a feel for the model, like test-driving a car. But who has time to test drive a fleet?

Then there’s the art of crafting the perfect prompt. It’s less science, more witchcraft. Does capitalizing words make a difference? Who knows? We’re all just throwing spaghetti at the wall and seeing what sticks.

In short, we’re in the era of “Vibes Based Development” — not the most reassuring approach, but hey, it’s what we’ve got.

The Gullibility Quagmire: AI’s Achilles Heel

Here’s a doozy: our shiny LLMs are as gullible as a kid who believes in the tooth fairy. We introduced the term “prompt injection” in the tech world’s vocabulary, and guess what? We’re still fumbling in the dark for a fix.

LLMs believe everything. It’s like having a digital sponge that soaks up everything — the good, the bad, and the nonsensical. Want a personal AI assistant? Imagine hiring someone who nods along to every piece of gossip or fake news. Not exactly employee of the month material.

Building AI agents that don’t fall for every trick in the book? We’re still dreaming. The day might come, but for now, we’re pretty much stuck with our gullible, albeit charming, digital companions.

The Coding Conundrum: LLMs as Programmers

In an ironic twist, it turns out that LLMs might be best suited for coding. Who’d have thought that these verbose machines would excel in the concise, logical world of programming languages?

The beauty lies in their simplicity. Programming languages are child’s play compared to the labyrinthine complexities of human languages. And when these LLMs hallucinate some nonexistent code (yes, they do that), we can just run the code to fact-check them. If only we had a fact-checking genie for everyday life!

So, as a techie, should I be worried about my job? Maybe. Or maybe it’s just another tool in our ever-growing toolbox, like a slightly erratic intern who sometimes surprises you with their brilliance.

The Ethical Enigma: The Dark Side of LLMs

Last year popped the lid off a Pandora’s box: the murky world of unlicensed training data. Fast forward, and we’re neck-deep in legal and ethical debates.

The New York Times’s lawsuit against OpenAI and Microsoft is like the latest blockbuster in the tech world. The crux? Using content without permission for training AI models. The document reads like a tech thriller, offering a peek into the complex world of AI legality.

This legal tussle isn’t just about laws; it’s a moral conundrum. Is it right to use people’s content without a nod from them, especially when it starts competing with them?

As these AI models get smarter, the ethical questions get tougher. The impact on jobs — from copywriters to artists — is real and growing. This story isn’t just about technology; it’s a human story, and we need more spotlights on it.

The Road Ahead: What’s Next for LLMs?

So, where does all this leave us? LLMs have been a rollercoaster — exciting, unpredictable, and a bit scary. They’ve changed how we work, play, and think about technology.

As we step into 2024, the big question is: where do we go from here? Can we make these models less gullible? Will someone finally outdo GPT-4? And how will we navigate the murky waters of ethics and legality?

One thing’s for sure: the LLM journey is far from over. It’s a story of human ingenuity, ambition, and, yes, a fair bit of stumbling in the dark. But isn’t that what makes any great story?

--

--

Shahid Mk

Data Scientist | AI Engineer | Researcher and Speaker . Turning data into actionable insights . LinkedIn : www.linkedin.com/in/shahidmaliyek