How This Started Over a House Rental

Two guys talking about a place to live ended up building the data layer behind legal AI.

The Actual Story

It started mundane. Fawzi was looking for a place to live. Jacob had a house. They got to talking—about projects, about interests, about problems they’d been thinking about.

Jacob had been wrestling with a specific research question for months: Why do legal language models confidently generate completely wrong answers? Not big structural failures. Subtle mistakes. Fictional case citations. Minor details that don’t exist. Logical reasoning that drifts into hallucination territory without the model ever noticing.

Fawzi brought a different perspective. He’s deep in the legal world—finishing his Global Law degree at Tilburg University, connected to practicing attorneys across the US, Middle East, and Europe through his family’s international law firm network. He understood what lawyers actually need and what real legal practice looks like.

Playing around with the problem together, something clicked: The issue wasn’t the models. It was the data.

Legal AI was being trained on static legal archives—case databases, statutory records, research repositories. Useful for legal research, but completely wrong for training systems to have conversations about law. And critically, the training data had almost no examples of common mistakes, subtle errors, or the kinds of hallucinations legal professionals deal with regularly.

It was like training a doctor on textbook anatomy but never showing them what an actual human body looks like, or what a misdiagnosis looks like.

They realized you could fix this by:

  • Working directly with practicing lawyers to understand real legal conversations
  • Building datasets specifically for conversational AI (not just research)
  • Deliberately including wrong answers to teach models what to avoid
  • Validating everything with legal professionals who understood the stakes

So in 2025, they started Entropy Partners.

Jacob brought the AI expertise. He’d just graduated from Cognitive Science & Artificial Intelligence at Tilburg and is now doing a Master’s at Utrecht. He understands how to structure data so AI systems can actually learn the right things from it.

Fawzi brought the legal network and domain understanding. Global Law student in his final year at Tilburg, with deep connections to practicing attorneys across multiple jurisdictions. He knew who to talk to and what questions to ask.

Together, they decided to build a new category: legal training datasets engineered specifically for conversational AI, created in collaboration with real attorneys, and deliberately designed to reduce hallucination through strategic error integration.

What We Believe

We’re not building this just to build a business. We believe something specific about the future of legal AI.

In five years, legal AI will be everywhere. It will power contract review tools, legal research platforms, compliance systems, and things we haven’t imagined yet. Hundreds of millions of people will interact with legal AI without knowing they are.

The quality of that AI will be determined almost entirely by the quality of its training data. Architecture matters, but far less than people think. Data matters exponentially more.

If legal AI is trained on the wrong data, it will cause real harm. It will give confidently wrong legal guidance, create liability for the companies using it, and damage the credibility of AI in law for years.

We want to be the foundation layer that makes trustworthy legal AI possible. Not the AI companies themselves. Just the data layer. The boring-but-critical foundation everything else depends on.

In one line

We exist so legal professionals and their clients can actually trust the AI they use.

Who’s Actually Building This

Fawzi Iyad Barakat

CEO & Co-Founder

Fawzi is finishing his Global Law degree at Tilburg University and leads our legal strategy and attorney network.

Law has been part of his world from the beginning. His father is a named partner at a major international firm, which gave Fawzi early exposure to how large-scale legal operations actually work and the real problems they face.

He’s lived internationally, including an extended period in Jordan, giving him deep market knowledge in the Middle East—where most legal AI infrastructure simply doesn’t exist yet.

Over time he’s built real relationships with 50+ practicing attorneys across 12+ jurisdictions in the US, Middle East, and Europe. Not LinkedIn connections. Actual working relationships.

His job at Entropy is simple to describe and hard to do: make sure every dataset we build reflects real legal practice and is created in partnership with the right attorneys in the right markets.

Jacob Bae

CTO & Co-Founder

Jacob graduated from Cognitive Science & Artificial Intelligence at Tilburg University and is now doing a Master’s at Utrecht. He’s the AI researcher who went deep on a specific question: why does legal AI confidently generate completely wrong answers?

His focus has never been “how do we build a bigger model?” but “how do we stop models from hallucinating subtle, high-stakes errors in specialized domains like law?”

Combining cognitive science and machine learning, he designs the pipelines that turn lawyer insight into structured training data—including strategically incorrect examples that teach models what to avoid.

His job at Entropy is to ensure that our data doesn’t just look good in a spreadsheet, but actually moves the needle on hallucination rates and real-world performance.

What We Commit To

We’re playing a long game. These are the lines we don’t cross.

Accuracy Is Non-Negotiable

We will never ship a dataset that doesn’t meet professional legal standards. If we’re not confident attorneys would sign their name under it, it doesn’t go out.

Radical Transparency

You know how your data was created, who validated it, and what it covers. We document methodology and provenance instead of hiding behind buzzwords.

Building for Real Use

We’re not chasing hype cycles. We’re methodically building the data layer for legal AI, starting with foundations we can defend in front of real practitioners.

Serious About Affordability

We price 40–70% below legacy providers by using modern infrastructure and direct attorney networks—not 1990s-era systems. Quality stays high; overhead doesn’t.

Expertise-Based Expansion

We expand into jurisdictions where we have real relationships and expertise, not just ambition. That’s why we started in the Middle East and are growing outwards, not the other way around.

Our 2030 Vision

We’re not aiming to be the loudest AI company. We’re aiming to be the quiet default.

By 2030, we want Entropy Partners to be the default data layer for legal AI. Not just “a vendor you can use”—the foundation most serious systems are built on.

  • 25+ exclusive licensing deals with top-tier legal AI companies.
  • 5,000+ datasets across major practice areas and 10+ key jurisdictions.
  • $10M+ in recurring revenue built on a pipeline we’ve refined for years.
  • Local practitioner networks in English, Arabic, Spanish, French, German, and Mandarin.
  • Recognition as the company that made trustworthy legal AI possible by getting the data right.

We’re not trying to replace lawyers. We’re building the infrastructure that lets AI systems actually respect what lawyers know.