← Back to data subTLDR

data subTLDR week 15 year 2026

r/MachineLearningr/dataengineeringr/SQL

Mastering Jailer: The Open Source Tool for Database Subsetting, Safe Practices for Database Changes, Strategies for Cross-source SQL Joins, Managing Data Pipeline Meltdowns, and Debunking Unrealistic Job Postings

Week 15, 2026
Posted in r/MachineLearningbyu/elnino20234/12/2026
280

"There's a new generation of empirical deep learning researchers, hacking away at whatever seems trendy, blowing with the wind" [D]

Discussion
The deep learning research community is divided over the focus on trending topics, with some seeing this as necessary for citations and career progression. There is recognition of the value in understanding deep learning, as demonstrated by the high regard for Andrew Gordon Wilson's work. However, there is also an acceptance that deep learning is an empirical science requiring a degree of 'hacking.' Some highlight the work of Geoffrey Hinton as evidence that insight can outweigh mathematical rigor. Critiques include the concern that the system incentivizes quantity over quality. The general sentiment is mixed, reflecting the complexity of the debate.
97 comments
Share
Save
View on Reddit →
Posted in r/MachineLearningbyu/etoipi14/6/2026
216

[D] How to break free from LLM's chains as a PhD student?

Discussion
The discussion centered around a PhD student’s worry over increasing reliance on Language Model (LLM) tools for coding tasks. Many users suggested it's not an issue, noting the importance of understanding what to code rather than how to code, and utilizing AI as a tool. Some suggested setting aside time for manual coding practice. However, others warned of the potential for mistakes and the importance of being able to read and verify LLM generated code. A few shared experiences with different LLM tools, such as Gemini Pro and Claude, and debated their efficacy. Overall, the sentiment was mixed, with a tilt towards embracing LLMs while maintaining a fundamental understanding of coding itself.
99 comments
Share
Save
View on Reddit →
Posted in r/MachineLearningbyu/we_are_mammals4/12/2026
187

Gary Marcus on the Claude Code leak [D]

Discussion
The discussion revolves around Gary Marcus's comment on the Claude Code leak, which he claims resembles classical symbolic AI. Opinions diverge, with the highest voted comment suggesting that the code is a large decision tree, requiring significant effort to develop, rather than a complex AI algorithm. Some users express skepticism about Marcus's credibility, arguing he consistently downplays AI's capabilities. Others agree that the code resembles rules-based AI or a detailed configuration file, but dispute Marcus's interpretation. There's also a reference to a paper co-authored by an Anthropic employee, which could provide context to Claude Code's design. The overall sentiment is mixed.
68 comments
Share
Save
View on Reddit →
Posted in r/dataengineeringbyu/SweetHunter27444/7/2026
137

data pipeline blew up at 2am and i have no clue where it started, how do you actually monitor this shit?

Rant
In a discussion about data pipeline monitoring, the consensus was to implement simple yet effective checks, such as row count on ingestion. Tools like Dagster, Prefect, and Temporal were mentioned, but the necessity of data validation before transformation was a recurring theme. Data contracts and dynamic pipelines were suggested to handle potential failures automatically. Some commenters suggested logging specific information, such as pipeline status, error logs, and quality check status, and using these to create monitoring dashboards and alerts. There was also an emphasis on balancing observability with monitoring, suggesting scheduled freshness checks, lightweight anomaly detectors, and alerting thresholds.
43 comments
Share
Save
View on Reddit →
Posted in r/MachineLearningbyu/Striking-Warning95334/7/2026
137

[D] thoughts on current community moving away from heavy math?

Discussion
The Machine Learning (ML) community has increasingly favored empirical findings and architecture designs over heavy mathematical theory, reflecting a shift towards practical applicability. A minority, however, still prioritizes mathematical rigor in their research. There's consensus that the field has always been largely empirical and intuition-driven, with math often used for post-hoc rationalization. Some argue that strong theoretical grounding might be necessary to overcome future roadblocks. Despite the trend, several mathy subfields remain active. Overall, respondents believe that both theory and applied work have valuable insights to offer each other. The sentiment is generally positive towards this trend, though concerns about potential future limitations exist.
75 comments
Share
Save
View on Reddit →
Posted in r/MachineLearningbyu/Tall_Bumblebee13414/11/2026
118

Is "live AI video generation" a meaningful technical category or just a marketing term? [R]

Research
1. User A (Upvotes: 50): Live AI video generation is indeed a meaningful technical category. Although often misused as a marketing term, it refers to a specific field of AI where algorithms are utilized to generate or alter video content in real-time. This is fundamentally different from fast video generation, which does not require real-time input and response. 2. User B (Upvotes: 30): It's true that the lack of a shared definition causes confusion. However, companies like Nvidia and DeepMind are pushing the boundaries of live AI video generation. 3. User C (Upvotes: 20): It would be beneficial if the field adopted a clearer taxonomy to differentiate between these technologies. Overall, the sentiment towards the term live AI video generation is mixed, with a call for clearer definitions and taxonomy in the AI field. The real-time aspect of this technology is recognized as a challenging yet significant part of AI development.
3 comments
Share
Save
View on Reddit →
Posted in r/dataengineeringbyu/peakpirate0074/10/2026
100

12+ years experience in a technology that launched a year ago lol

Rant
The job description (JD) requesting 12+ years of experience in a technology that only launched a year ago sparked conversations about unrealistic expectations in job postings. The majority agreed that such requirements are indicative of a disconnect between hiring managers and technical jobs. The unrealistic JD was seen as a potential red flag for some, while others suggested applying regardless, as the recruiters might not know what they're asking for. Some comments also pointed out the trend of inflated requirements for junior-level roles. Overall, there's a consensus that such job postings reflect a broader issue in the recruitment process.
27 comments
Share
Save
View on Reddit →
Posted in r/dataengineeringbyu/josh_docglow4/9/2026
75

I built an open source tool to replace standard dbt docs

Personal Project Showcase
The creator of Docglow, an open-source tool designed to improve the default dbt docs process, has asked for feedback. The tool, which includes features like an interactive lineage explorer and column-level lineage tracing, was generally well-received with users highlighting column lineage as a standout feature. A comparison was brought up between Docglow and Colibri, another tool used for similar purposes. Some users reported issues with the tool, such as the server hanging without showing output or the site not loading, but the creator was quick to respond to these concerns. The overall sentiment is positive, indicating interest and promise in Docglow's future development.
23 comments
Share
Save
View on Reddit →
Posted in r/SQLbyu/venusFarts4/11/2026
29

Jailer is a open source tool for database subsetting, schema and data browsing.

Discussion
Jailer, an open source tool for database subsetting, schema, and data browsing, is gaining recognition for its ability to create small, navigable slices from databases. Key features include a Data Browser, a Subsetter for creating small slices from production databases, and the capacity to improve database performance by removing and archiving obsolete data. Jailer generates various formats such as SQL, JSON, YAML, XML, and DbUnit datasets. Overall, the sentiment is positive, although there is a lack of extensive feedback due to the limited number of comments. The one available comment indicates a nostalgic association with previous database tools.
1 comments
Share
Save
View on Reddit →
Posted in r/SQLbyu/sqlmans4/9/2026
16

Best practices for safe database changes

Discussion
Database developers emphasize the importance of discipline and robust practices when making changes to databases. Key strategies include using scripts for all changes, generating full deployment scripts before running anything, and maintaining a history of schema changes. Many recommend tools like SSDT for MS SQL, git repositories, and CI/CD workflows. Some advocate for python's alembic or DbUp for managing schema migrations. The consensus is to minimize direct editing in production environments and to use prepared scripts for all alterations. The overall sentiment is a mix of sharing best practices and discussing tools for effective database management.
6 comments
Share
Save
View on Reddit →
Posted in r/SQLbyu/Pitiful_Comedian_8344/11/2026
14

Cross-source SQL joins without a data warehouse - how do you handle this?

Discussion
The discussion revolved around strategies for handling cross-source SQL joins without a data warehouse. The consensus favored ETL processes for collating data into one location, and using tools like dbt for transformation. However, some respondents highlighted the potential of DuckDB-based desktop tools for native handling. The overall sentiment was mixed, with participants acknowledging the strengths and weaknesses of different approaches. There was an appreciation for the potential of new tools, but a clear recognition of the proven effectiveness of traditional methods like ETL.
6 comments
Share
Save
View on Reddit →

Subscribe to data-subtldr

Get weekly summaries of top content from r/dataengineering, r/MachineLearning and more directly in your inbox.

Get the weekly data subTLDR in your inbox!

We respect your privacy. No spam, ever.