Tech

Claims in this domain span AI hype and skepticism — from whether AGI is imminent to whether your phone is secretly listening. Disputes around AI coding tools, deepfakes, and bot counts on X are among the most contested.

64 Tech claim verifications avg. score 5.1/10 26 rated true or mostly true 38 rated false or misleading

“The key hardware components of a delivery drone include motors, electronic speed controllers (ESCs), flight controllers, GPS modules, and payload release mechanisms.”

Mostly True

All five components listed — motors, ESCs, flight controllers, GPS modules, and payload release mechanisms — are well-documented as standard hardware in delivery drones across multiple authoritative technical sources. The word "include" signals a non-exhaustive list, so the claim does not purport to be complete. However, other equally essential components such as propellers, batteries, airframes, and communication systems are omitted, which could leave readers with an incomplete picture of delivery drone hardware.

“In April 2026, Turkish authorities dismantled an organized cybercrime network that illegally accessed and sold Turkish citizens' personal data obtained from government systems, operating through a dealership-based distribution model.”

False

No credible evidence confirms that Turkish authorities dismantled a cybercrime network matching this description in April 2026. The closest documented operation occurred on March 26, 2026, involved data from both public institutions and non-government sources like Facebook, and used "query panels" — not a "dealership-based distribution model." The only official Turkish police communication from April 2026 makes no mention of such an operation, and no independent news outlet has reported one.

“Smart stickers that detect ammonia can be used as a non-invasive method to monitor food freshness or spoilage.”

True

Multiple peer-reviewed studies and industry sources confirm that ammonia-detecting smart stickers have been successfully demonstrated as non-invasive food freshness monitors, particularly for protein-rich foods like meat and fish. The claim's "can be used" framing is a capability statement that the evidence clearly supports across several sensor types (colorimetric, graphene-based, NFC-enabled). Most implementations remain at prototype or early commercialization stages rather than widespread consumer deployment, and real-world performance can be affected by humidity and cross-gas interference.

“Graphene-based supercapacitors and batteries offer higher energy density and faster charge cycles than conventional lithium-ion technologies as of April 16, 2026.”

Misleading

The claim bundles a genuine advantage with an unsupported one. Graphene-based technologies do charge significantly faster than conventional lithium-ion — multiple sources confirm this. However, the assertion of "higher energy density" is contradicted by the best available evidence: the leading graphene aluminium-ion battery (GMG) achieves only 26–101 Wh/kg depending on charge rate, well below lithium-ion's commercial 150–250 Wh/kg range. Even the manufacturer's own disclosures acknowledge this gap. The energy density claim relies on theoretical projections and marketing materials, not demonstrated commercial performance.

“There exist multiple programmatic methods and technical architectures for editing industrial mesh-based 3D models with a focus on preserving precision, geometry, and engineering features, including direct mesh manipulation, mesh-to-CAD reconstruction, voxel/SDF workflows, primitive recognition, and hybrid pipelines, each with distinct trade-offs in dimensional accuracy, feature retention, and industrial suitability.”

Mostly True

The claim accurately identifies a well-documented ecosystem of distinct programmatic approaches for editing industrial mesh-based 3D models, supported by peer-reviewed research and industrial documentation from institutions including CNRS, MIT, IEEE/CVPR, and Altair. Each named category — direct mesh manipulation, mesh-to-CAD reconstruction, voxel/SDF workflows, primitive recognition, and hybrid pipelines — has credible evidence behind it. The one material caveat is that several cited "feature-preserving" mesh methods preserve geometric shape fidelity rather than enabling parametric, tolerance-driven engineering edits, a distinction the claim's "trade-offs" language gestures at but does not make explicit.

“As of 2026, factory reset does not reliably erase all personal data from electronic devices, and significant amounts of recoverable personal information remain on many second-hand devices sold or recycled worldwide.”

Mostly True

The core assertion holds: factory resets perform logical deletion rather than physical data destruction, and authoritative technical standards (NIST SP 800-88) classify them as insufficient for assured non-recoverability. Real-world audits of second-hand devices have consistently found recoverable personal data on substantial fractions of resold units. However, the claim understates the protection offered by modern encrypted smartphones, where factory reset destroys encryption keys, rendering residual data practically inaccessible. Some frequently cited prevalence statistics also predate 2026 by nearly a decade.

“The Poincaré embedding model, introduced by Maximilian Nickel and Douwe Kiela in 2017, demonstrated that hierarchical structures can be embedded with low distortion in hyperbolic space.”

Mostly True

The claim accurately identifies the authors, year, and core contribution of the Poincaré embeddings paper, and the broader research community consistently describes the work as demonstrating low-distortion hierarchical embedding in hyperbolic space. The original 2017 paper empirically showed that Poincaré ball embeddings significantly outperform Euclidean baselines on hierarchical datasets like WordNet. However, the paper provides empirical benchmarks rather than formal distortion guarantees, and later research shows distortion can increase for wider hierarchies.

“AI language models can be reliably cited as primary sources in academic papers.”

False

Academic institutions, style guides, and peer-reviewed research uniformly reject the notion that AI language models serve as reliable primary sources. While citation formats exist for disclosing LLM use, these frameworks address transparency and attribution—not epistemic reliability. Documented problems including hallucinated references, citation bias, and factual inaccuracies mean LLM outputs require human verification and cannot substitute for peer-reviewed primary literature in academic work.

“In 2020, Citizen Watch Co. launched Accutron as a standalone brand featuring a new electrostatic energy mechanism that had not previously been used in wristwatches.”

Mostly True

The core facts of this claim are well-supported across numerous credible sources. Accutron was indeed relaunched as a standalone brand in 2020, and its electrostatic-induction power system is widely confirmed as a world first in wristwatches. Two minor imprecisions prevent a fully clean rating: the launch was executed through Citizen's subsidiary Bulova rather than directly by "Citizen Watch Co.," and while the specific electrostatic system is novel, some underlying regulation concepts are not entirely new.

“A Python program can be written to replace all occurrences of the first character in a string with '$', except for the first character itself.”

True

The described task is straightforwardly achievable in Python. Multiple independent sources provide working code — typically combining string slicing with `str.replace()` — that replaces all occurrences of the first character with '$' while preserving the first character itself. The claim says only that such a program "can be written," and the evidence unanimously confirms this. The sole nuance is that Python strings are immutable, so the result is a new string rather than an in-place modification.

“Git is a version control system that operates locally, while GitHub is a cloud-based platform for hosting and collaborating on Git repositories.”

True

This widely accepted distinction between Git and GitHub is directly confirmed by official documentation from both projects. Git's own docs state that "most operations in Git need only local files and resources," and GitHub's docs describe it as "a cloud-based platform where you can store, share, and work together with others to write code." The claim is a standard, accurate characterization with only minor simplifications that do not distort the core meaning.

“Artificial intelligence will cause widespread job loss among software engineers.”

False
· 100+ views

The available evidence does not support the prediction that AI will cause widespread job loss among software engineers. High-authority sources from Morgan Stanley, MIT Sloan, arXiv, and Snowflake consistently point toward augmentation, productivity gains, and net job growth rather than broad displacement. The evidence cited in favor of the claim — worse outcomes for recent graduates in AI-exposed fields, economy-wide self-reports — does not isolate software engineers, does not establish AI as the causal driver, and conflates hiring difficulty with job destruction.

“Oxford University has predicted that the percentage of jobless people will decline as artificial intelligence advances.”

False

No Oxford University source has made the specific prediction attributed to it. Oxford-affiliated research discusses AI's complex labor market effects — noting that mass displacement fears may be overstated and that AI could create new roles — but none of these findings constitute a forecast that the percentage of jobless people will decline as AI advances. The claim conflates cautious, nuanced commentary with a definitive institutional prediction that does not exist in the evidence.

“The World Economic Forum's Future of Jobs Report 2025 states that 60% of employers expect expanding digital access to transform their business operations by 2030.”

Mostly True

The 60% statistic is well-supported by the WEF Future of Jobs Report 2025, as confirmed by the primary EY-hosted document and multiple secondary sources. The claim's wording differs slightly from the report's original language — the report says "broadening digital access" and "transform their business," while the claim says "expanding digital access" and "business operations." These are minor paraphrasing differences that preserve the substantive meaning without creating a false impression.

“Robots will not replace human teachers in schools in the near future.”

Misleading

The broad expert consensus supports the idea that AI will primarily augment teachers rather than fully replace them at scale — but the claim's categorical "will not replace" framing overstates what the evidence shows. At least one real-world school network already uses AI-delivered lessons with non-credentialed supervisors instead of traditional teachers, and mainstream analysis from Brookings acknowledges technology may reduce the number of teachers needed. The claim is directionally sound but misleadingly absolute.

“Artificial intelligence will displace more jobs than it creates on a net basis.”

Misleading

The claim that AI will displace more jobs than it creates on a net basis overstates the available evidence. While documented displacement exists in specific sectors (e.g., computer systems design, entry-level roles, AI-vulnerable occupations), the most authoritative aggregate assessments — from the Federal Reserve, World Economic Forum, PwC, and Goldman Sachs — show near-zero net headcount effects or project net job creation. The claim treats localized displacement as proof of an economy-wide net loss, which current evidence does not support.

“In the first quarter of 2026, approximately 27% of production code merged into main branches was authored or substantially shaped by artificial intelligence systems.”

Misleading

The ~27% figure is directionally plausible but overstates the certainty and universality of the underlying evidence. It appears to derive from a single self-reported developer survey (DX Newsletter, Q1 2026) across 500+ organizations, with no disclosed methodology for how "authored or substantially shaped" was defined or measured. Other available data points use incompatible definitions — "code written," single-company disclosures, or broader global estimates — and range from 25% to over 50%, making any single number highly sensitive to measurement choices.

“By early 2026, the largest empirical study available, covering 4.2 million developers, found that AI-authored code accounted for 26.9% of production code.”

False

No publicly documented study covering 4.2 million developers and reporting 26.9% AI-authored production code exists as of early 2026. The closest real study — published in Science and covering ~160,000 GitHub developers — found 29% AI-written Python code in the US by late 2025, a fundamentally different sample size, metric, and scope. The claim's specific figures appear fabricated or conflated from incompatible sources, making the overall assertion unsupported.

“Large language model hallucinations are produced by the same underlying mechanism that generates correct outputs.”

Mostly True

Both hallucinations and correct outputs do emerge from the same autoregressive next-token prediction process — no separate "hallucination engine" exists within large language models. Multiple peer-reviewed sources confirm this shared generative pipeline. However, the claim omits critical nuance: hallucinations have distinct causal drivers — such as training procedures that reward guessing over expressing uncertainty, data distribution gaps, and prompting effects — that do not equally govern correct outputs. The generation channel is shared, but the upstream conditions that produce errors are separable and require distinct mitigation strategies.

“Five major tech companies, including Anthropic, OpenAI, and Microsoft, have launched AI chatbots specifically for consumer health support in 2026.”

False

The specific claim that five major tech companies launched consumer health chatbots in 2026 is not supported by the evidence. Multiple credible sources confirm dedicated health AI products from only three companies: Anthropic (Claude for Healthcare), OpenAI (ChatGPT Health), and Microsoft (Copilot Health). A possible fourth (Amazon) is weakly documented by a single source describing a different type of tool, and no fifth company launch is substantiated. The numerical assertion — the claim's defining element — is unverified.