Launched this week

Lenz
Fact-check any statement with source-backed, multi-model AI
28 followers
Fact-check any statement with source-backed, multi-model AI
28 followers
Lenz is an AI fact-checker. Paste any claim and Lenz searches for independent sources, has multiple AI models argue both sides and a separate AI panel review the evidence before reaching a scored verdict — with every source and argument and step of the process visible to you. Most AI tools give you one model's best guess from memory. Lenz orchestrates a multi-model research system based on real sources so no single model's blind spots drive the conclusion. Try it free at lenz.io/ph














I’ve used Lenz to fact check scholarly materials in philosophy. More specifically, contentious topics like Heidegger, where personal opinions can bias the way he’s interpreted. I’ve found that Lenz has a solid way of grading truth claims on the 1 to 10 scale which really helps separate where on the spectrum a claim lies.
An interesting use case @ivan_spasov thanks for sharing!
Philosophy is exactly where things get tricky as it’s less about clear facts and more about interpretation (and bias:)
Really glad the truthfulness score was useful - sometimes claims sit on a spectrum .
Were there were any results that surprised you?
@vicky_dodeva something that surprised me was that results filter opinions from facts very well. To give an example with one of my use cases - an important part of Heidegger’s work is his critique on metaphysics and if philosophy has answered the question of Being as such through the study of metaphysics. He is very harsh in his commentary that historically metaphysics doesn’t actually answer this question. But here is where it gets interesting - does he himself actually answer it? Explicitly, based on primary texts - no, at least not in a direct way. Scholarly texts might say that he actually does though. But this is the real distinction here - what has Heidegger explicitly said and what is an interpretation.
I found that Lenz is actually very good at making that distinction. What is actually true by virtue of the actual text and what is a scholarly interpretation.
@ivan_spasov wow, this is a really thorough example, Ivan, thanks for sharing it. That distinction between primary text and interpretation is exactly what we hope Lenz helps surface
@vicky_dodeva That distinction is extremely valuable in the humanities in general, so things like this could really help verify work in academic discourse.
Happy launching. I have a question. How do you quantify and mitigate correlated bias across your ‘independent’ AI models and sources especially when those models are trained on overlapping datasets and the sources themselves may share upstream information pipelines so that your multi-model consensus doesn’t simply reinforce systemic errors rather than reduce them?
@bogomep Deep question. Overall concept is to pick the highest-quality available sources/evidence and work with that data (plus the LLM background knowledge which is explicitly listed as a separate source in the output). From that perspective, we don't eliminate correlated bias in the sources, but we try to identify diverse sources and have rules such as looking for sources that support both sides of the argument, and looking for local/regional sources -- that's why if you check, for example, for China, you will get Chinese sources alongside the Western ones. Each source is graded for authority, recency, etc. Retracted research papers are excluded, etc.
With respect to the models, there are two key intermediate steps: (1) adversarial debate between two different models that's aimed to crystalize the positions and arguments and is also valuable for the user to see the strongest opposing arguments, and (2) panel of three different AI models reviews the evidence and the debates in parallel and each model evaluates different axis to detect logical fallacies, biases, context issues, etc. Each of the three panel model is instructed to analyze the data from a different perspective. Then a voting makes the final call, and the executive summary is written. This does not solve correlated bias from shared training data entirely, but helps. And this orchestrated multi-model multi-axis AI pipeline of frontier models has a higher likelihood of producing higher quality output than a single frontier model alone.
All sources, citations, debates, panel reviews, and voting scores are available to the user to inspect.
I tried lenz with several claims and I liked the way it actually felt like the tool was looking for the most reliable sources out there to confirm with the biggest possible authority if my claims were true or not.
In both situations, the comeback from the tool was great, however I dont know if i loved the ui, here are my points.
The final delivery makes you feel like you are not using the whole screen, there is soo much blank space, it becomes an infinite scrolling of truth, while you could take the full screen and make the information provided more digestible.
Puttingin one side of the screen an index to know what to expect from the full page would be useful, as well as grouping the expert reviews and arguments so you dont have to see them all before passing to the next phase.
Overall i think the tool is great, im excited to know how it will evolve.
PS. I have a product series on my linkedin called "Products that just make sense" I would love to feature Lenz!
@carolinahunts Wow, thank you for taking the time to try Lenz so thoroughly and glad to hear that the core experience of the product resonated.
On the UI comments, we hear you and your feedback is super valuable. Yes, we’ve heavily leaned into transparency - showing the full reasoning, etc, which led to what you called “the infinite scroll of truth” (spot-on!) We are still working on making that process easier to navigate and digest. We really appreciate the deep dive!
Would love to stay in touch as we evolve this. What’s the best way to connect? I couldn’t find your linkedin on your PH profile.
Thanks again
@vicky_dodeva Sure, my linkedin is https://www.linkedin.com/in/carolinatrujillom/
happy to help!
If multiple models are arguing both sides, how do you ensure the final verdict isn’t just "averaging opinions" but actually getting closer to truth?
@lak7 After the debates, we run another step at which 3 other models in parallel review on three different axis the sources/evidence and the debates, and make the final decision. They flag things like biases, logical fallacies, etc. in the arguments (also shown in the final report), and are pretty good at detecting the weak ones.
Amazing! When do we get SKILL + CLI or an MCP so our agents can do fact checking for us? Or a chrome extension
@milko_slavov There's an API available -- https://lenz.io/api/v1/docs/ API keys are managed from the Account menu. Drop me a note if you need any support. There's also a bookmarklet (https://lenz.io/install when accessed from desktop) as a first step before the chrome extension.
@kostaj Great! Here is the skill + cli: https://github.com/mslavov/lenz-cli
@milko_slavov Wow! Amazing! Welcome to the Lenz team :) I will play with the skill & cli later today.
Congratulations!
Does it rely on a single or multi- model approach to perform the fact checking? Any plans for specialized / domain-focused checks - for example solving historic mysteries?
@stefan_lilov Lenz relies on a pipeline that includes several steps. At the key steps (debates and panel reviews), several different forefront models from different providers are run in parallel. We push the models to not rely on their internal factual knowledge, but to focus on evaluating the selected high-quality sources/evidence instead. From that perspective, we don't use domain-trained/focused models in the pipeline, but it's a strong idea to introduce the knowledge of such models at the source level. We already include LLMs internal knowledge as one of the sources (and it's highlighted for the user), but it's not currently coming from a domain-focused model.
When a user start chatting with Lenz (the follow-up questions), then we use a domain-trained model, that also has full access to the source/evidence materials and is restricted to not discuss topics outside of it's narrow domain.