Sung Kim's Avatar

Sung Kim

@sungkim

A business analyst at heart who enjoys delving into AI, ML, data engineering, data science, data analytics, and modeling. My views are my own. You can also find me at threads: @sung.kim.mw

7,611
Followers
1,153
Following
6,608
Posts
22.01.2024
Joined
Posts Following

Latest posts by Sung Kim @sungkim

Use AI to turn that experience into action. Build the things you’ve always wanted to build.

07.03.2026 05:10 πŸ‘ 12 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

For those of you who have spent decades in the tech industry: AI has leveled the playing field with the young engineers grinding LeetCode.

You bring something they don’tβ€”decades of experience in design patterns, system architecture, and infrastructure.
.

07.03.2026 05:10 πŸ‘ 45 πŸ” 1 πŸ’¬ 4 πŸ“Œ 1
Post image

What arxiv.org/abs/2512.24873

07.03.2026 02:53 πŸ‘ 23 πŸ” 2 πŸ’¬ 4 πŸ“Œ 1

LOL. It's always crypto.

07.03.2026 03:14 πŸ‘ 4 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Otherwise is known as TACO countdown has started.

07.03.2026 02:44 πŸ‘ 7 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Preview
Cursor Goes To War For AI Coding Dominance After becoming the hottest, fastest growing AI coding company, Cursor is confronting a new reality: developers may no longer need a code editor at all.

Source: www.forbes.com/sites/annato...

07.03.2026 02:17 πŸ‘ 7 πŸ” 0 πŸ’¬ 0 πŸ“Œ 1

They also estimate that Claude Code’s $200 monthly plan, which previously may cost up to $2,000 to support, is now may costing Anthropic up to $5,000.

Cursor needs its own model now, just to compete!

07.03.2026 02:14 πŸ‘ 14 πŸ” 2 πŸ’¬ 4 πŸ“Œ 1
Post image

Being an AI wrapper company is brutal.

Cursor, last valued at $29.3 billion (Series D) in November 2025, is on red alert. Despite surpassing $2 billion in ARR and doubling its revenue in the three months since its last round, the pressure is mounting.

07.03.2026 02:14 πŸ‘ 49 πŸ” 8 πŸ’¬ 3 πŸ“Œ 4

Does this mean I have to buy a Mac mini and install OpenClaw to stay up to date with Chinese Grandpas? I really don't want to. Ugh.

06.03.2026 22:15 πŸ‘ 13 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0

UPDATE: It's real, per Tencent.

06.03.2026 22:09 πŸ‘ 36 πŸ” 0 πŸ’¬ 1 πŸ“Œ 2
Post image

Actually, $5,000. Probably, not USD.

06.03.2026 22:02 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image
06.03.2026 21:59 πŸ‘ 6 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
Video thumbnail
06.03.2026 21:59 πŸ‘ 5 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image Post image Post image Post image

Is this real? There's a large turnout for an OpenClaw installation offsite in Shenzhen.

06.03.2026 21:54 πŸ‘ 21 πŸ” 2 πŸ’¬ 2 πŸ“Œ 2
Post image

A better FlashAttention V3?

vLLM Triton Attention has ~800 lines of Triton, same source code across NVIDIA, AMD, and Intel GPUs. On H100, it matches state-of-the-art attention performance. On MI300, ~5.8x faster than earlier implementations.

blog.vllm.ai/2026/03/04/v...

06.03.2026 17:54 πŸ‘ 11 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Post image

Did the author solved the Riemann Hypothesis or not? You tell me.

"Analysis of the Riemann Zeta Function via Recursive Taylor Expansions"

arxiv.org/abs/2603.05122

06.03.2026 17:49 πŸ‘ 5 πŸ” 0 πŸ’¬ 3 πŸ“Œ 0
Preview
BM25 There is a particular kind of respect reserved in engineering for the algorithm that outlives its era. BM25 is one of them. BM25 was born out of information retrieval research in the 1970s and 1980s, ...

When you need to tune for your domain, the parameters give you meaningful handles to turn. The interpretability is genuinely valuable."

arpitbhayani.me/blogs/bm25

06.03.2026 07:11 πŸ‘ 6 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0

BM25 by Arpit Bhayani

"What makes BM25 worth understanding is not just that it works. It is that it works for knowable reasons. Every part of the formula has a clear interpretation. When a result is surprising, you can trace why.

06.03.2026 07:11 πŸ‘ 15 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Video thumbnail

A word of wisdom to live by - do not let your luxury possession possess you.

06.03.2026 06:16 πŸ‘ 10 πŸ” 2 πŸ’¬ 2 πŸ“Œ 0
Video thumbnail

So true.

05.03.2026 23:34 πŸ‘ 27 πŸ” 1 πŸ’¬ 1 πŸ“Œ 1

My thoughts on gpt-5.4 high on Codex CLI

I have no idea if it is better than gpt-5.3-codex or even gpt-5.2, but it devours tokens like a competitive eater at a Las Vegas buffet.

05.03.2026 21:31 πŸ‘ 26 πŸ” 0 πŸ’¬ 4 πŸ“Œ 0
Post image

Intel Panther Lake Die Shot

Why does it look like Impressionist painting? BSPDN.

05.03.2026 20:32 πŸ‘ 37 πŸ” 1 πŸ’¬ 2 πŸ“Œ 0

FYI

05.03.2026 19:31 πŸ‘ 8 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Video thumbnail

Speculative Speculative Decoding (SSD)

It's up to 2x faster than the strongest inference engines in the world, but you need H100 or better GPUs.

Paper: arxiv.org/abs/2603.03251
Repo: github.com/tanishqkumar...

05.03.2026 19:16 πŸ‘ 16 πŸ” 2 πŸ’¬ 1 πŸ“Œ 0

PyTorch's FlexAttention also supports FlashAttention-4 backend.

PyTorch now auto-generates CuTeDSL score/mask modifications and JIT-instantiates FlashAttention-4 for your custom attention variant

The result: 1.2Γ— to 3.2Γ— speedups over Triton on compute-bound workloads.

pytorch.org/blog/flexatt...

05.03.2026 18:52 πŸ‘ 10 πŸ” 0 πŸ’¬ 0 πŸ“Œ 1

- Paper: github.com/Dao-AILab/fl...
- Code: github.com/Dao-AILab/fl...

- Blogposts:
together.ai/blog/flashat...
tridao.me/blog/2026/fl...
research.colfax-intl.com/flashattenti...

05.03.2026 18:47 πŸ‘ 3 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
Post image

FlashAttention-4

I hope it is not pain to work with. It changes the algorithm & pipeline so that softmax & SMEM bandwidth no longer dictate speed. Attn reaches ~1600 TFLOPs, pretty much at matmul speed!

05.03.2026 18:47 πŸ‘ 27 πŸ” 4 πŸ’¬ 2 πŸ“Œ 3

You can always go to other platforms and browse through 100s or 1,000s of postings and view posts by the original authors.

05.03.2026 16:22 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Video thumbnail

OpenAI's Symphony

A Linear Board for agents.

github.com/openai/symph...

05.03.2026 06:35 πŸ‘ 15 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Preview
Teaching LLMs to reason like Bayesians Google researchers demonstrate how Bayesian teaching through supervised fine-tuning enables LLMs to approximate optimal probabilistic reasoning and generalize to new domains.

Teaching LLMs to reason like Bayesians

By training models to mimic optimal probabilistic inference, they improved their ability to update their predictions and generalize across new domains.

research.google/blog/teachin...

05.03.2026 06:22 πŸ‘ 50 πŸ” 2 πŸ’¬ 2 πŸ“Œ 0