Main

related bits

0

processing priority

4

site type

3 (personal blog or private political site, e.g. Blogspot, Substack, also small blogs on own domains)

review version

11

html import

20 (imported)

Events

first seen date

2025-07-20 07:16:49

expired found date

-

created at

2025-07-20 07:16:49

updated at

2025-11-16 18:06:41

Domain name statistics

length

21

crc

9168

tld

2211

nm parts

0

nm random digits

0

nm rare letters

0

Connections

is subdomain of id

69893241 (blogspot.com)

previous id

0

replaced with id

0

related id

-

dns primary id

0

dns alternative id

0

lifecycle status

0 (unclassified, or currently active)

Subdomains and pages

deleted subdomains

0

page imported products

0

page imported random

0

page imported parking

0

Error counters

count skipped due to recent timeouts on the same server IP

0

count content received but rejected due to 11-799

0

count dns errors

0

count cert errors

0

count timeouts

0

count http 429

0

count http 404

0

count http 403

0

count http 5xx

0

next operation date

-

Server

server bits

server ip

-

Mainpage statistics

mp import status

20

mp rejected date

-

mp saved date

-

mp size orig

164546

mp size raw text

16850

mp inner links count

168

mp inner links status

10 (links queued, awaiting import)

Open Graph

title

Adhyayan

description

books writing literature nature

image

site name

author

updated

2026-03-04 01:04:41

raw text

Adhyayan Adhyayan Tuesday, July 15, 2025 Missing learning loop for LLMs   Andrej's take on RL  with chatgpt magic.  How humans understand  touches on the need for this elaboration. 1. RL is powerful, but not the full story RL is gaining traction and will continue to generate useful results , particularly because it’s more leveraged than traditional supervised fine-tuning (SFT). But it has limitations —particularly with long-horizon tasks (tasks that take a long time or many steps). The standard RL approach—rewarding or punishing actions based on final scalar feedback —is very lossy , especially when the task is long and complex. “You're really going to do all that work just to learn a single scalar outcome at the very end?” 2. Human learning isn’t like that Humans don't just get a reward at the end; they reflect . After doing something, we think: What went well? What didn’t? What could I do differently? These explicit lessons are stored consciousl...

Text analysis

redirect type

0 (-)

block type

0 (no issues)

detected language

1 (English)

category id

Zastosowania AI (149)

index version

1

spam phrases

0

Text statistics

text nonlatin

0

text cyrillic

0

text characters

12527

text words

2329

text unique words

874

text lines

530

text sentences

110

text paragraphs

22

text words per sentence

21

text matched phrases

0

text matched dictionaries

0

RSS

rss status

32 (unknown)

rss found date

2025-07-23 05:40:15

rss size orig

645573

rss items

25

rss spam phrases

0

rss detected language

1 (English)

inbefore feed id

-

inbefore status

0 (new)

Sitemap

sitemap status

10 (sitemap found, awaiting processing)

sitemap review version

0

sitemap urls count

0

sitemap urls adult

0

sitemap filtered products

0

sitemap filtered videos

0

sitemap found date

2025-07-28 05:05:26

sitemap process date

-

sitemap first import date

-

sitemap last import date

-