Main

related bits

0

processing priority

3

site type

0 (generic, awaiting analysis)

review version

11

html import

20 (imported)

Events

first seen date

2025-05-16 15:04:18

expired found date

-

created at

2025-05-16 15:04:18

updated at

2026-02-17 05:17:29

Domain name statistics

length

20

crc

35937

tld

86

nm parts

0

nm random digits

0

nm rare letters

0

Connections

is subdomain of id

87719371 (github.io)

previous id

0

replaced with id

0

related id

-

dns primary id

0

dns alternative id

0

lifecycle status

0 (unclassified, or currently active)

Subdomains and pages

deleted subdomains

0

page imported products

0

page imported random

0

page imported parking

0

Error counters

count skipped due to recent timeouts on the same server IP

0

count content received but rejected due to 11-799

0

count dns errors

0

count cert errors

0

count timeouts

0

count http 429

0

count http 404

0

count http 403

0

count http 5xx

0

next operation date

-

Server

server bits

server ip

-

Mainpage statistics

mp import status

20

mp rejected date

-

mp saved date

-

mp size orig

36191

mp size raw text

3300

mp inner links count

21

mp inner links status

20 (imported)

Open Graph

title

论文与代码阅读笔记

description

site name

论文与代码阅读笔记

author

GDD

updated

2026-02-15 06:41:32

raw text

论文与代码阅读笔记 文章 29 标签 34 Home Archives Tags 论文与代码阅读笔记 搜索 Home Archives Tags 论文与代码阅读笔记 Resources for Machine Learning System 置顶 | 发表于 2021-03-02 | 更新于 2023-03-03 | 集群调度 • 分布式训练 • 机器学习系统 • 深度学习框架 总结机器学习系统领域的学习资源,以及值得关注和阅读的paper,主要涉及分布式深度学习、集群调度等。 每天一个没用的代码小技巧 置顶 | 发表于 2022-01-10 | 更新于 2023-07-12 一些小技巧、小方法的合集。每天一个小技巧,效率翻倍! DistFlashAtten:面向长上下文大语言模型训练的内存高效的分布式注意力机制 发表于 2024-05-31 | 更新于 2024-06-09 本文介绍了一种名为DistFlashAttn的分布式内存高效注意力机制,它是为了优化长上下文大型语言模型(LLMs)的训练而设计的。文章提出了三种关键技术:token级别的工作负载平衡、交叠的键值通信,以及重计算梯度检查点算法。这些技术使得DistFlashAttn能够在多设备上高效地分发token块,同时保持内存高效注意力的I/O感知优势。 BPipe: 面向大语言模型训练的内存均衡的流水线并行 发表于 2023-09-25 | 更新于 2023-09-27 | 分布式训练 • 机器学习系统 • ICML • 大模型 流水线并行是在GPU集群中训练大型语言模型的关键技术。然而,流水线并行通常会导致内存不平衡问题,某些GPU的内存压力很大,而其他GPU则没有充分利用其内存。这种不平衡会影响训练性能。为了解决这种低效性,BPIPE在流水线并行中实现内存平衡。BPIPE在训练期间在GPU之间传输中间激活值,使所有GPU都能利用相当大小的内存。通过平衡内存利用率,BPIPE消除冗余重新计算或增加微批大小,从而提高GPT-3等大型语言模型的训练效率。 Lucid:一个可扩展、可解释的实用型深度学习作...

Text analysis

redirect type

0 (-)

block type

0 (no issues)

detected language

126 (language undetectable (empty document, too short, or engines disagree))

category id

Zastosowania AI (149)

index version

2025123101

spam phrases

0

Text statistics

text nonlatin

1698

text cyrillic

0

text characters

2465

text words

305

text unique words

199

text lines

214

text sentences

1

text paragraphs

0

text words per sentence

255

text matched phrases

2

text matched dictionaries

2

RSS

rss path

rss status

1 (priority 1 already searched, no matches found)

rss found date

-

rss size orig

0

rss items

0

rss spam phrases

0

rss detected language

0 (awaiting analysis)

inbefore feed id

-

inbefore status

0 (new)

Sitemap

sitemap status

10 (sitemap found, awaiting processing)

sitemap review version

0

sitemap urls count

0

sitemap urls adult

0

sitemap filtered products

0

sitemap filtered videos

0

sitemap found date

2025-05-16 15:04:19

sitemap process date

-

sitemap first import date

-

sitemap last import date

-